Commit 12a8eee5 authored by Sean Christopherson's avatar Sean Christopherson Committed by Paolo Bonzini

KVM: Move x86 VMX's posted interrupt list_head to vcpu_vmx

Move the seemingly generic block_vcpu_list from kvm_vcpu to vcpu_vmx, and
rename the list and all associated variables to clarify that it tracks
the set of vCPU that need to be poked on a posted interrupt to the wakeup
vector.  The list is not used to track _all_ vCPUs that are blocking, and
the term "blocked" can be misleading as it may refer to a blocking
condition in the host or the guest, where as the PI wakeup case is
specifically for the vCPUs that are actively blocking from within the
guest.

No functional change intended.
Signed-off-by: default avatarSean Christopherson <seanjc@google.com>
Reviewed-by: default avatarMaxim Levitsky <mlevitsk@redhat.com>
Message-Id: <20211208015236.1616697-7-seanjc@google.com>
Signed-off-by: default avatarPaolo Bonzini <pbonzini@redhat.com>
parent e6eec09b
...@@ -19,7 +19,7 @@ ...@@ -19,7 +19,7 @@
* wake the target vCPUs. vCPUs are removed from the list and the notification * wake the target vCPUs. vCPUs are removed from the list and the notification
* vector is reset when the vCPU is scheduled in. * vector is reset when the vCPU is scheduled in.
*/ */
static DEFINE_PER_CPU(struct list_head, blocked_vcpu_on_cpu); static DEFINE_PER_CPU(struct list_head, wakeup_vcpus_on_cpu);
/* /*
* Protect the per-CPU list with a per-CPU spinlock to handle task migration. * Protect the per-CPU list with a per-CPU spinlock to handle task migration.
* When a blocking vCPU is awakened _and_ migrated to a different pCPU, the * When a blocking vCPU is awakened _and_ migrated to a different pCPU, the
...@@ -27,7 +27,7 @@ static DEFINE_PER_CPU(struct list_head, blocked_vcpu_on_cpu); ...@@ -27,7 +27,7 @@ static DEFINE_PER_CPU(struct list_head, blocked_vcpu_on_cpu);
* CPU. IRQs must be disabled when taking this lock, otherwise deadlock will * CPU. IRQs must be disabled when taking this lock, otherwise deadlock will
* occur if a wakeup IRQ arrives and attempts to acquire the lock. * occur if a wakeup IRQ arrives and attempts to acquire the lock.
*/ */
static DEFINE_PER_CPU(raw_spinlock_t, blocked_vcpu_on_cpu_lock); static DEFINE_PER_CPU(raw_spinlock_t, wakeup_vcpus_on_cpu_lock);
static inline struct pi_desc *vcpu_to_pi_desc(struct kvm_vcpu *vcpu) static inline struct pi_desc *vcpu_to_pi_desc(struct kvm_vcpu *vcpu)
{ {
...@@ -51,6 +51,7 @@ static int pi_try_set_control(struct pi_desc *pi_desc, u64 old, u64 new) ...@@ -51,6 +51,7 @@ static int pi_try_set_control(struct pi_desc *pi_desc, u64 old, u64 new)
void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu) void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu)
{ {
struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu);
struct vcpu_vmx *vmx = to_vmx(vcpu);
struct pi_desc old, new; struct pi_desc old, new;
unsigned long flags; unsigned long flags;
unsigned int dest; unsigned int dest;
...@@ -86,9 +87,9 @@ void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu) ...@@ -86,9 +87,9 @@ void vmx_vcpu_pi_load(struct kvm_vcpu *vcpu, int cpu)
* current pCPU if the task was migrated. * current pCPU if the task was migrated.
*/ */
if (pi_desc->nv == POSTED_INTR_WAKEUP_VECTOR) { if (pi_desc->nv == POSTED_INTR_WAKEUP_VECTOR) {
raw_spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->cpu)); raw_spin_lock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu));
list_del(&vcpu->blocked_vcpu_list); list_del(&vmx->pi_wakeup_list);
raw_spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->cpu)); raw_spin_unlock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu));
} }
dest = cpu_physical_id(cpu); dest = cpu_physical_id(cpu);
...@@ -142,15 +143,16 @@ static bool vmx_can_use_vtd_pi(struct kvm *kvm) ...@@ -142,15 +143,16 @@ static bool vmx_can_use_vtd_pi(struct kvm *kvm)
static void pi_enable_wakeup_handler(struct kvm_vcpu *vcpu) static void pi_enable_wakeup_handler(struct kvm_vcpu *vcpu)
{ {
struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu);
struct vcpu_vmx *vmx = to_vmx(vcpu);
struct pi_desc old, new; struct pi_desc old, new;
unsigned long flags; unsigned long flags;
local_irq_save(flags); local_irq_save(flags);
raw_spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->cpu)); raw_spin_lock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu));
list_add_tail(&vcpu->blocked_vcpu_list, list_add_tail(&vmx->pi_wakeup_list,
&per_cpu(blocked_vcpu_on_cpu, vcpu->cpu)); &per_cpu(wakeup_vcpus_on_cpu, vcpu->cpu));
raw_spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, vcpu->cpu)); raw_spin_unlock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu));
WARN(pi_desc->sn, "PI descriptor SN field set before blocking"); WARN(pi_desc->sn, "PI descriptor SN field set before blocking");
...@@ -199,24 +201,23 @@ void vmx_vcpu_pi_put(struct kvm_vcpu *vcpu) ...@@ -199,24 +201,23 @@ void vmx_vcpu_pi_put(struct kvm_vcpu *vcpu)
*/ */
void pi_wakeup_handler(void) void pi_wakeup_handler(void)
{ {
struct kvm_vcpu *vcpu;
int cpu = smp_processor_id(); int cpu = smp_processor_id();
struct vcpu_vmx *vmx;
raw_spin_lock(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); raw_spin_lock(&per_cpu(wakeup_vcpus_on_cpu_lock, cpu));
list_for_each_entry(vcpu, &per_cpu(blocked_vcpu_on_cpu, cpu), list_for_each_entry(vmx, &per_cpu(wakeup_vcpus_on_cpu, cpu),
blocked_vcpu_list) { pi_wakeup_list) {
struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu);
if (pi_test_on(pi_desc)) if (pi_test_on(&vmx->pi_desc))
kvm_vcpu_kick(vcpu); kvm_vcpu_kick(&vmx->vcpu);
} }
raw_spin_unlock(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); raw_spin_unlock(&per_cpu(wakeup_vcpus_on_cpu_lock, cpu));
} }
void __init pi_init_cpu(int cpu) void __init pi_init_cpu(int cpu)
{ {
INIT_LIST_HEAD(&per_cpu(blocked_vcpu_on_cpu, cpu)); INIT_LIST_HEAD(&per_cpu(wakeup_vcpus_on_cpu, cpu));
raw_spin_lock_init(&per_cpu(blocked_vcpu_on_cpu_lock, cpu)); raw_spin_lock_init(&per_cpu(wakeup_vcpus_on_cpu_lock, cpu));
} }
bool pi_has_pending_interrupt(struct kvm_vcpu *vcpu) bool pi_has_pending_interrupt(struct kvm_vcpu *vcpu)
......
...@@ -6943,6 +6943,8 @@ static int vmx_create_vcpu(struct kvm_vcpu *vcpu) ...@@ -6943,6 +6943,8 @@ static int vmx_create_vcpu(struct kvm_vcpu *vcpu)
BUILD_BUG_ON(offsetof(struct vcpu_vmx, vcpu) != 0); BUILD_BUG_ON(offsetof(struct vcpu_vmx, vcpu) != 0);
vmx = to_vmx(vcpu); vmx = to_vmx(vcpu);
INIT_LIST_HEAD(&vmx->pi_wakeup_list);
err = -ENOMEM; err = -ENOMEM;
vmx->vpid = allocate_vpid(); vmx->vpid = allocate_vpid();
......
...@@ -317,6 +317,9 @@ struct vcpu_vmx { ...@@ -317,6 +317,9 @@ struct vcpu_vmx {
/* Posted interrupt descriptor */ /* Posted interrupt descriptor */
struct pi_desc pi_desc; struct pi_desc pi_desc;
/* Used if this vCPU is waiting for PI notification wakeup. */
struct list_head pi_wakeup_list;
/* Support for a guest hypervisor (nested VMX) */ /* Support for a guest hypervisor (nested VMX) */
struct nested_vmx nested; struct nested_vmx nested;
......
...@@ -309,8 +309,6 @@ struct kvm_vcpu { ...@@ -309,8 +309,6 @@ struct kvm_vcpu {
u64 requests; u64 requests;
unsigned long guest_debug; unsigned long guest_debug;
struct list_head blocked_vcpu_list;
struct mutex mutex; struct mutex mutex;
struct kvm_run *run; struct kvm_run *run;
......
...@@ -427,8 +427,6 @@ static void kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id) ...@@ -427,8 +427,6 @@ static void kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id)
#endif #endif
kvm_async_pf_vcpu_init(vcpu); kvm_async_pf_vcpu_init(vcpu);
INIT_LIST_HEAD(&vcpu->blocked_vcpu_list);
kvm_vcpu_set_in_spin_loop(vcpu, false); kvm_vcpu_set_in_spin_loop(vcpu, false);
kvm_vcpu_set_dy_eligible(vcpu, false); kvm_vcpu_set_dy_eligible(vcpu, false);
vcpu->preempted = false; vcpu->preempted = false;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment