hv: support halt in hv idle
When all vCPU threads on one pCPU are put to sleep (e.g., when all guests execute HLT), hv would schedule to idle thread. Currently the idle thread executes PAUSE which does not enter any c-state and consumes a lot of power. This patch is to support HLT in the idle thread. When we switch to HLT, we have to make sure events that would wake a vCPU must also be able to wake the pCPU. Those events are either generated by local interrupt or issued by other pCPUs followed by an ipi kick. Each of them have an interrupt involved, so they are also able to wake the halted pCPU. Except when the pCPU has just scheduled to idle thread but not yet halted, interrupts could be missed. sleep-------schedule to idle------IRQ ON---HLT--(kick missed) ^ wake---kick| This areas should be protected. This is done by a safe halt mechanism leveraging STI instruction’s delay effect (same as Linux). vCPUs with lapic_pt or hv with CONFIG_KEEP_IRQ_DISABLED=y does not allow interrupts in root mode, so they could never wake from HLT (INIT kick does not wake HLT in root mode either). They should continue using PAUSE in idle. Tracked-On: #8507 Signed-off-by: Wu Zhou <wu.zhou@intel.com> Reviewed-by: Junjie Mao <junjie.mao@intel.com>
This commit is contained in:
parent
863f220a19
commit
064be1e3e6
|
@ -462,7 +462,25 @@ void stop_pcpus(void)
|
|||
|
||||
void cpu_do_idle(void)
|
||||
{
|
||||
#ifdef CONFIG_KEEP_IRQ_DISABLED
|
||||
asm_pause();
|
||||
#else
|
||||
uint16_t pcpu_id = get_pcpu_id();
|
||||
|
||||
if (per_cpu(mode_to_idle, pcpu_id) == IDLE_MODE_HLT) {
|
||||
asm_safe_hlt();
|
||||
} else {
|
||||
struct acrn_vcpu *vcpu = get_ever_run_vcpu(pcpu_id);
|
||||
|
||||
if ((vcpu != NULL) && !is_lapic_pt_enabled(vcpu)) {
|
||||
CPU_IRQ_ENABLE_ON_CONFIG();
|
||||
}
|
||||
asm_pause();
|
||||
if ((vcpu != NULL) && !is_lapic_pt_enabled(vcpu)) {
|
||||
CPU_IRQ_DISABLE_ON_CONFIG();
|
||||
}
|
||||
}
|
||||
#endif
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -528,9 +528,22 @@ int32_t create_vcpu(uint16_t pcpu_id, struct acrn_vm *vm, struct acrn_vcpu **rtn
|
|||
per_cpu(ever_run_vcpu, pcpu_id) = vcpu;
|
||||
|
||||
if (is_lapic_pt_configured(vm) || is_using_init_ipi()) {
|
||||
/* Lapic_pt pCPU does not enable irq in root mode. So it
|
||||
* should be set to PAUSE idle mode.
|
||||
* At this point the pCPU is possibly in HLT idle. And the
|
||||
* kick mode is to be set to INIT kick, which will not be
|
||||
* able to wake root mode HLT. So a kick(if pCPU is in HLT
|
||||
* idle, the kick mode is certainly ipi kick) will change
|
||||
* it to PAUSE idle right away.
|
||||
*/
|
||||
if (per_cpu(mode_to_idle, pcpu_id) == IDLE_MODE_HLT) {
|
||||
per_cpu(mode_to_idle, pcpu_id) = IDLE_MODE_PAUSE;
|
||||
kick_pcpu(pcpu_id);
|
||||
}
|
||||
per_cpu(mode_to_kick_pcpu, pcpu_id) = DEL_MODE_INIT;
|
||||
} else {
|
||||
per_cpu(mode_to_kick_pcpu, pcpu_id) = DEL_MODE_IPI;
|
||||
per_cpu(mode_to_idle, pcpu_id) = IDLE_MODE_HLT;
|
||||
}
|
||||
pr_info("pcpu=%d, kick-mode=%d, use_init_flag=%d", pcpu_id,
|
||||
per_cpu(mode_to_kick_pcpu, pcpu_id), is_using_init_ipi());
|
||||
|
|
|
@ -25,7 +25,13 @@ void reset_event(struct sched_event *event)
|
|||
spinlock_irqrestore_release(&event->lock, rflag);
|
||||
}
|
||||
|
||||
/* support exclusive waiting only */
|
||||
/* support exclusive waiting only
|
||||
*
|
||||
* During wait, the pCPU could be scheduled to run the idle thread when run queue
|
||||
* is empty. Signal_event() can happen when schedule() is in process.
|
||||
* This signal_event is not going to be lost, for the idle thread will always
|
||||
* check need_reschedule() after it is switched to at schedule().
|
||||
*/
|
||||
void wait_event(struct sched_event *event)
|
||||
{
|
||||
uint64_t rflag;
|
||||
|
|
|
@ -86,9 +86,7 @@ void default_idle(__unused struct thread_object *obj)
|
|||
} else if (need_shutdown_vm(pcpu_id)) {
|
||||
shutdown_vm_from_idle(pcpu_id);
|
||||
} else {
|
||||
CPU_IRQ_ENABLE_ON_CONFIG();
|
||||
cpu_do_idle();
|
||||
CPU_IRQ_DISABLE_ON_CONFIG();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -65,6 +65,9 @@ void init_sched(uint16_t pcpu_id)
|
|||
{
|
||||
struct sched_control *ctl = &per_cpu(sched_ctl, pcpu_id);
|
||||
|
||||
per_cpu(mode_to_idle, pcpu_id) = IDLE_MODE_HLT;
|
||||
per_cpu(mode_to_kick_pcpu, pcpu_id) = DEL_MODE_IPI;
|
||||
|
||||
spinlock_init(&ctl->scheduler_lock);
|
||||
ctl->flags = 0UL;
|
||||
ctl->curr_obj = NULL;
|
||||
|
|
|
@ -533,6 +533,15 @@ static inline void asm_hlt(void)
|
|||
asm volatile ("hlt");
|
||||
}
|
||||
|
||||
/* interrupts remain inhibited on the instruction boundary following
|
||||
* an execution of STI. This will make sure pending interrupts will
|
||||
* wake hlt.
|
||||
*/
|
||||
static inline void asm_safe_hlt(void)
|
||||
{
|
||||
asm volatile ("sti; hlt; cli" : : : "cc");
|
||||
}
|
||||
|
||||
/* Disables interrupts on the current CPU */
|
||||
#ifdef CONFIG_KEEP_IRQ_DISABLED
|
||||
#define CPU_IRQ_DISABLE_ON_CONFIG() do { } while (0)
|
||||
|
|
|
@ -55,6 +55,7 @@ struct per_cpu_region {
|
|||
uint32_t lapic_ldr;
|
||||
uint32_t softirq_servicing;
|
||||
uint32_t mode_to_kick_pcpu;
|
||||
uint32_t mode_to_idle;
|
||||
struct smp_call_info_data smp_call_info;
|
||||
struct list_head softirq_dev_entry_list;
|
||||
#ifdef PROFILING_ON
|
||||
|
|
|
@ -15,6 +15,9 @@
|
|||
#define DEL_MODE_INIT (1U)
|
||||
#define DEL_MODE_IPI (2U)
|
||||
|
||||
#define IDLE_MODE_PAUSE (1U)
|
||||
#define IDLE_MODE_HLT (2U)
|
||||
|
||||
#define THREAD_DATA_SIZE (256U)
|
||||
|
||||
enum thread_object_state {
|
||||
|
|
Loading…
Reference in New Issue