From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752540AbaLCHwR (ORCPT ); Wed, 3 Dec 2014 02:52:17 -0500 Received: from mga11.intel.com ([192.55.52.93]:15093 "EHLO mga11.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752416AbaLCHuz (ORCPT ); Wed, 3 Dec 2014 02:50:55 -0500 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.07,506,1413270000"; d="scan'208";a="631852505" From: Feng Wu To: tglx@linutronix.de, mingo@redhat.com, hpa@zytor.com, x86@kernel.org, gleb@kernel.org, pbonzini@redhat.com, dwmw2@infradead.org, joro@8bytes.org, alex.williamson@redhat.com, jiang.liu@linux.intel.com Cc: linux-kernel@vger.kernel.org, iommu@lists.linux-foundation.org, kvm@vger.kernel.org, Feng Wu Subject: [v2 21/25] KVM: Update Posted-Interrupts descriptor during vCPU scheduling Date: Wed, 3 Dec 2014 15:39:50 +0800 Message-Id: <1417592394-24343-22-git-send-email-feng.wu@intel.com> X-Mailer: git-send-email 1.7.1 In-Reply-To: <1417592394-24343-1-git-send-email-feng.wu@intel.com> References: <1417592394-24343-1-git-send-email-feng.wu@intel.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Update Posted-Interrupts descriptor according to the following rules: - Before vCPU block, set 'NV' to POSTED_INTR_WAKEUP_VECTOR - After vCPU block, set 'NV' back to POSTED_INTR_VECTOR Signed-off-by: Feng Wu --- arch/x86/include/asm/kvm_host.h | 5 ++ arch/x86/kvm/vmx.c | 83 +++++++++++++++++++++++++++++++++++++++ arch/x86/kvm/x86.c | 16 +++++++ virt/kvm/kvm_main.c | 11 +++++ 4 files changed, 115 insertions(+), 0 deletions(-) diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h index 6878429..2fd85a5 100644 --- a/arch/x86/include/asm/kvm_host.h +++ b/arch/x86/include/asm/kvm_host.h @@ -775,6 +775,8 @@ struct kvm_x86_ops { void (*sched_in)(struct kvm_vcpu *kvm, int cpu); u64 (*get_pi_desc_addr)(struct kvm_vcpu *vcpu); + int (*vcpu_pre_block)(struct kvm_vcpu *vcpu); + void (*vcpu_post_block)(struct kvm_vcpu *vcpu); }; struct kvm_arch_async_pf { @@ -1100,4 +1102,7 @@ void kvm_deliver_pmi(struct kvm_vcpu *vcpu); bool kvm_find_dest_vcpu(struct kvm *kvm, struct kvm_lapic_irq *irq, struct kvm_vcpu **dest_vcpu); +int kvm_arch_vcpu_pre_block(struct kvm_vcpu *vcpu); +void kvm_arch_vcpu_post_block(struct kvm_vcpu *vcpu); + #endif /* _ASM_X86_KVM_HOST_H */ diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c index 81f239b..a1966b9 100644 --- a/arch/x86/kvm/vmx.c +++ b/arch/x86/kvm/vmx.c @@ -9147,6 +9147,86 @@ static void vmx_sched_in(struct kvm_vcpu *vcpu, int cpu) shrink_ple_window(vcpu); } +static int vmx_vcpu_pre_block(struct kvm_vcpu *vcpu) +{ + struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); + struct pi_desc old; + struct pi_desc new; + + if (!irq_remapping_cap(IRQ_POSTING_CAP)) + return 0; + + memset(&old, 0, sizeof(old)); + memset(&new, 0, sizeof(new)); + + do { + old.control = new.control = pi_desc->control; + + /* + * A posted-interrupt happened in the one of the + * following two cases: + * 1. After the latest pir-to-virr sync operation + * in kvm_arch_vcpu_runnable() function + * 2. In this do-while() loop, a posted-interrupt + * occurs. + * + * For either of above cases, we should not block + * the VCPU. + */ + if (pi_test_on(pi_desc) == 1) { + /* + * Need to set this flag, then the inject will + * be synced from PIR to vIRR before VM-ENTRY. + * In fact, for guest IPI case, in function + * vmx_deliver_posted_interrupt(), this flags + * has already been set, but if the interrupt + * is injected by VT-d PI hardware, we need + * to set this. + */ + kvm_make_request(KVM_REQ_EVENT, vcpu); + return 1; + } + + pi_clear_sn(&new); + + /* set 'NV' to 'wakeup vector' */ + new.nv = POSTED_INTR_WAKEUP_VECTOR; + } while (cmpxchg(&pi_desc->control, old.control, new.control) + != old.control); + + return 0; +} + +static void vmx_vcpu_post_block(struct kvm_vcpu *vcpu) +{ + struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); + struct pi_desc old; + struct pi_desc new; + unsigned int dest = 0; + + if (!irq_remapping_cap(IRQ_POSTING_CAP)) + return; + + pi_set_sn(pi_desc); + + do { + old.control = new.control = pi_desc->control; + + dest = cpu_physical_id(vcpu->cpu); + + if (x2apic_enabled()) + new.ndst = dest; + else + new.ndst = (dest << 8) & 0xFF00; + + /* set 'NV' to 'notification vector' */ + new.nv = POSTED_INTR_VECTOR; + } while (cmpxchg(&pi_desc->control, old.control, new.control) + != old.control); + + pi_clear_sn(pi_desc); +} + static struct kvm_x86_ops vmx_x86_ops = { .cpu_has_kvm_support = cpu_has_kvm_support, .disabled_by_bios = vmx_disabled_by_bios, @@ -9256,6 +9336,9 @@ static struct kvm_x86_ops vmx_x86_ops = { .sched_in = vmx_sched_in, .get_pi_desc_addr = vmx_get_pi_desc_addr, + + .vcpu_pre_block = vmx_vcpu_pre_block, + .vcpu_post_block = vmx_vcpu_post_block, }; static int __init vmx_init(void) diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index 0033df3..9706984 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -7731,6 +7731,22 @@ bool kvm_arch_has_noncoherent_dma(struct kvm *kvm) } EXPORT_SYMBOL_GPL(kvm_arch_has_noncoherent_dma); +int kvm_arch_vcpu_pre_block(struct kvm_vcpu *vcpu) +{ + if (kvm_x86_ops->vcpu_pre_block) + return kvm_x86_ops->vcpu_pre_block(vcpu); + + return 0; +} +EXPORT_SYMBOL_GPL(kvm_arch_vcpu_pre_block); + +void kvm_arch_vcpu_post_block(struct kvm_vcpu *vcpu) +{ + if (kvm_x86_ops->vcpu_post_block) + kvm_x86_ops->vcpu_post_block(vcpu); +} +EXPORT_SYMBOL_GPL(kvm_arch_vcpu_post_block); + EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_exit); EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_inj_virq); EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_page_fault); diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 25ffac9..1be1a45 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -1754,7 +1754,18 @@ void kvm_vcpu_block(struct kvm_vcpu *vcpu) if (signal_pending(current)) break; +#ifdef CONFIG_X86 + if (kvm_arch_vcpu_pre_block(vcpu) == 1) { + kvm_make_request(KVM_REQ_UNHALT, vcpu); + break; + } +#endif + schedule(); + +#ifdef CONFIG_X86 + kvm_arch_vcpu_post_block(vcpu); +#endif } finish_wait(&vcpu->wq, &wait); -- 1.7.1 From mboxrd@z Thu Jan 1 00:00:00 1970 From: Feng Wu Subject: [v2 21/25] KVM: Update Posted-Interrupts descriptor during vCPU scheduling Date: Wed, 3 Dec 2014 15:39:50 +0800 Message-ID: <1417592394-24343-22-git-send-email-feng.wu@intel.com> References: <1417592394-24343-1-git-send-email-feng.wu@intel.com> Mime-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Cc: iommu-cunTk1MwBs9QetFLy7KEm3xJsTq8ys+cHZ5vskTnxNA@public.gmane.org, linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org, kvm-u79uwXL29TY76Z2rM5mHXA@public.gmane.org To: tglx-hfZtesqFncYOwBW4kG4KsQ@public.gmane.org, mingo-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org, hpa-YMNOUZJC4hwAvxtiuMwx3w@public.gmane.org, x86-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org, gleb-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org, pbonzini-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org, dwmw2-wEGCiKHe2LqWVfeAwA7xHQ@public.gmane.org, joro-zLv9SwRftAIdnm+yROfE0A@public.gmane.org, alex.williamson-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org, jiang.liu-VuQAYsv1563Yd54FQh9/CA@public.gmane.org Return-path: In-Reply-To: <1417592394-24343-1-git-send-email-feng.wu-ral2JQCrhuEAvxtiuMwx3w@public.gmane.org> List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Sender: iommu-bounces-cunTk1MwBs9QetFLy7KEm3xJsTq8ys+cHZ5vskTnxNA@public.gmane.org Errors-To: iommu-bounces-cunTk1MwBs9QetFLy7KEm3xJsTq8ys+cHZ5vskTnxNA@public.gmane.org List-Id: kvm.vger.kernel.org Update Posted-Interrupts descriptor according to the following rules: - Before vCPU block, set 'NV' to POSTED_INTR_WAKEUP_VECTOR - After vCPU block, set 'NV' back to POSTED_INTR_VECTOR Signed-off-by: Feng Wu --- arch/x86/include/asm/kvm_host.h | 5 ++ arch/x86/kvm/vmx.c | 83 +++++++++++++++++++++++++++++++++++++++ arch/x86/kvm/x86.c | 16 +++++++ virt/kvm/kvm_main.c | 11 +++++ 4 files changed, 115 insertions(+), 0 deletions(-) diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h index 6878429..2fd85a5 100644 --- a/arch/x86/include/asm/kvm_host.h +++ b/arch/x86/include/asm/kvm_host.h @@ -775,6 +775,8 @@ struct kvm_x86_ops { void (*sched_in)(struct kvm_vcpu *kvm, int cpu); u64 (*get_pi_desc_addr)(struct kvm_vcpu *vcpu); + int (*vcpu_pre_block)(struct kvm_vcpu *vcpu); + void (*vcpu_post_block)(struct kvm_vcpu *vcpu); }; struct kvm_arch_async_pf { @@ -1100,4 +1102,7 @@ void kvm_deliver_pmi(struct kvm_vcpu *vcpu); bool kvm_find_dest_vcpu(struct kvm *kvm, struct kvm_lapic_irq *irq, struct kvm_vcpu **dest_vcpu); +int kvm_arch_vcpu_pre_block(struct kvm_vcpu *vcpu); +void kvm_arch_vcpu_post_block(struct kvm_vcpu *vcpu); + #endif /* _ASM_X86_KVM_HOST_H */ diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c index 81f239b..a1966b9 100644 --- a/arch/x86/kvm/vmx.c +++ b/arch/x86/kvm/vmx.c @@ -9147,6 +9147,86 @@ static void vmx_sched_in(struct kvm_vcpu *vcpu, int cpu) shrink_ple_window(vcpu); } +static int vmx_vcpu_pre_block(struct kvm_vcpu *vcpu) +{ + struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); + struct pi_desc old; + struct pi_desc new; + + if (!irq_remapping_cap(IRQ_POSTING_CAP)) + return 0; + + memset(&old, 0, sizeof(old)); + memset(&new, 0, sizeof(new)); + + do { + old.control = new.control = pi_desc->control; + + /* + * A posted-interrupt happened in the one of the + * following two cases: + * 1. After the latest pir-to-virr sync operation + * in kvm_arch_vcpu_runnable() function + * 2. In this do-while() loop, a posted-interrupt + * occurs. + * + * For either of above cases, we should not block + * the VCPU. + */ + if (pi_test_on(pi_desc) == 1) { + /* + * Need to set this flag, then the inject will + * be synced from PIR to vIRR before VM-ENTRY. + * In fact, for guest IPI case, in function + * vmx_deliver_posted_interrupt(), this flags + * has already been set, but if the interrupt + * is injected by VT-d PI hardware, we need + * to set this. + */ + kvm_make_request(KVM_REQ_EVENT, vcpu); + return 1; + } + + pi_clear_sn(&new); + + /* set 'NV' to 'wakeup vector' */ + new.nv = POSTED_INTR_WAKEUP_VECTOR; + } while (cmpxchg(&pi_desc->control, old.control, new.control) + != old.control); + + return 0; +} + +static void vmx_vcpu_post_block(struct kvm_vcpu *vcpu) +{ + struct pi_desc *pi_desc = vcpu_to_pi_desc(vcpu); + struct pi_desc old; + struct pi_desc new; + unsigned int dest = 0; + + if (!irq_remapping_cap(IRQ_POSTING_CAP)) + return; + + pi_set_sn(pi_desc); + + do { + old.control = new.control = pi_desc->control; + + dest = cpu_physical_id(vcpu->cpu); + + if (x2apic_enabled()) + new.ndst = dest; + else + new.ndst = (dest << 8) & 0xFF00; + + /* set 'NV' to 'notification vector' */ + new.nv = POSTED_INTR_VECTOR; + } while (cmpxchg(&pi_desc->control, old.control, new.control) + != old.control); + + pi_clear_sn(pi_desc); +} + static struct kvm_x86_ops vmx_x86_ops = { .cpu_has_kvm_support = cpu_has_kvm_support, .disabled_by_bios = vmx_disabled_by_bios, @@ -9256,6 +9336,9 @@ static struct kvm_x86_ops vmx_x86_ops = { .sched_in = vmx_sched_in, .get_pi_desc_addr = vmx_get_pi_desc_addr, + + .vcpu_pre_block = vmx_vcpu_pre_block, + .vcpu_post_block = vmx_vcpu_post_block, }; static int __init vmx_init(void) diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index 0033df3..9706984 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -7731,6 +7731,22 @@ bool kvm_arch_has_noncoherent_dma(struct kvm *kvm) } EXPORT_SYMBOL_GPL(kvm_arch_has_noncoherent_dma); +int kvm_arch_vcpu_pre_block(struct kvm_vcpu *vcpu) +{ + if (kvm_x86_ops->vcpu_pre_block) + return kvm_x86_ops->vcpu_pre_block(vcpu); + + return 0; +} +EXPORT_SYMBOL_GPL(kvm_arch_vcpu_pre_block); + +void kvm_arch_vcpu_post_block(struct kvm_vcpu *vcpu) +{ + if (kvm_x86_ops->vcpu_post_block) + kvm_x86_ops->vcpu_post_block(vcpu); +} +EXPORT_SYMBOL_GPL(kvm_arch_vcpu_post_block); + EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_exit); EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_inj_virq); EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_page_fault); diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 25ffac9..1be1a45 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -1754,7 +1754,18 @@ void kvm_vcpu_block(struct kvm_vcpu *vcpu) if (signal_pending(current)) break; +#ifdef CONFIG_X86 + if (kvm_arch_vcpu_pre_block(vcpu) == 1) { + kvm_make_request(KVM_REQ_UNHALT, vcpu); + break; + } +#endif + schedule(); + +#ifdef CONFIG_X86 + kvm_arch_vcpu_post_block(vcpu); +#endif } finish_wait(&vcpu->wq, &wait); -- 1.7.1