/* * Copyright (C) 2018-2022 Intel Corporation. * * SPDX-License-Identifier: BSD-3-Clause */ #include #include #include #include #include #include #include #include #include #include static uint32_t notification_irq = IRQ_INVALID; static uint64_t smp_call_mask = 0UL; /* run in interrupt context */ static void kick_notification(__unused uint32_t irq, __unused void *data) { /* Notification vector is used to kick target cpu out of non-root mode. * And it also serves for smp call. */ uint16_t pcpu_id = get_pcpu_id(); if (bitmap_test(pcpu_id, &smp_call_mask)) { struct smp_call_info_data *smp_call = &per_cpu(smp_call_info, pcpu_id); if (smp_call->func != NULL) { smp_call->func(smp_call->data); } bitmap_clear_lock(pcpu_id, &smp_call_mask); } } void handle_smp_call(void) { kick_notification(0, NULL); } void smp_call_function(uint64_t mask, smp_call_func_t func, void *data) { uint16_t pcpu_id; struct smp_call_info_data *smp_call; /* wait for previous smp call complete, which may run on other cpus */ while (atomic_cmpxchg64(&smp_call_mask, 0UL, mask) != 0UL); pcpu_id = ffs64(mask); while (pcpu_id < MAX_PCPU_NUM) { bitmap_clear_nolock(pcpu_id, &mask); if (pcpu_id == get_pcpu_id()) { func(data); bitmap_clear_nolock(pcpu_id, &smp_call_mask); } else if (is_pcpu_active(pcpu_id)) { smp_call = &per_cpu(smp_call_info, pcpu_id); smp_call->func = func; smp_call->data = data; struct acrn_vcpu *vcpu = get_ever_run_vcpu(pcpu_id); if ((vcpu != NULL) && (is_lapic_pt_enabled(vcpu))) { vcpu_make_request(vcpu, ACRN_REQUEST_SMP_CALL); } else { send_single_ipi(pcpu_id, NOTIFY_VCPU_VECTOR); } } else { /* pcpu is not in active, print error */ pr_err("pcpu_id %d not in active!", pcpu_id); bitmap_clear_nolock(pcpu_id, &smp_call_mask); } pcpu_id = ffs64(mask); } /* wait for current smp call complete */ wait_sync_change(&smp_call_mask, 0UL); } static int32_t request_notification_irq(irq_action_t func, void *data) { int32_t retval; if (notification_irq != IRQ_INVALID) { pr_info("%s, Notification vector already allocated on this CPU", __func__); retval = -EBUSY; } else { /* all cpu register the same notification vector */ retval = request_irq(NOTIFY_VCPU_IRQ, func, data, IRQF_NONE); if (retval < 0) { pr_err("Failed to add notify isr"); retval = -ENODEV; } else { notification_irq = (uint32_t)retval; } } return retval; } /* * @pre be called only by BSP initialization process */ void setup_notification(void) { /* support IPI notification, Service VM will register all CPU */ if (request_notification_irq(kick_notification, NULL) < 0) { pr_err("Failed to setup notification"); } dev_dbg(DBG_LEVEL_PTIRQ, "NOTIFY: irq[%d] setup vector %x", notification_irq, irq_to_vector(notification_irq)); } /* * posted interrupt handler * @pre (irq - POSTED_INTR_IRQ) < CONFIG_MAX_VM_NUM */ static void handle_pi_notification(uint32_t irq, __unused void *data) { uint32_t vcpu_index = irq - POSTED_INTR_IRQ; ASSERT(vcpu_index < CONFIG_MAX_VM_NUM, ""); vcpu_handle_pi_notification(vcpu_index); } /*pre-condition: be called only by BSP initialization proccess*/ void setup_pi_notification(void) { uint32_t i; for (i = 0U; i < CONFIG_MAX_VM_NUM; i++) { if (request_irq(POSTED_INTR_IRQ + i, handle_pi_notification, NULL, IRQF_NONE) < 0) { pr_err("Failed to setup pi notification"); break; } } }