All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive
@ 2016-09-21  2:14 Wanpeng Li
  2016-09-21  2:38 ` Wanpeng Li
  2016-09-21 16:35 ` Radim Krčmář
  0 siblings, 2 replies; 4+ messages in thread
From: Wanpeng Li @ 2016-09-21  2:14 UTC (permalink / raw)
  To: linux-kernel, kvm
  Cc: Wanpeng Li, Paolo Bonzini, Radim Krčmář,
	Wincy Van, Yang Zhang

From: Wanpeng Li <wanpeng.li@hotmail.com>

I observed that kvmvapic(to optimize flexpriority=N or AMD) is used 
to boost TPR access when testing kvm-unit-test/eventinj.flat tpr case
on my haswell desktop (w/ flexpriority, w/o APICv). Commit (8d14695f9542 
x86, apicv: add virtual x2apic support) disable virtual x2apic mode 
completely if w/o APICv, and the author also told me that windows guest
can't enter into x2apic mode when he developed the APICv feature several 
years ago. However, it is not truth currently, Interrupt Remapping and 
vIOMMU is added to qemu and the developers from Intel test windows 8 can 
work in x2apic mode w/ Interrupt Remapping enabled recently. 

This patch enables TPR shadow for virtual x2apic mode to boost 
windows guest in x2apic mode even if w/o APICv.

Can pass the kvm-unit-test.

Suggested-by: Radim Krčmář <rkrcmar@redhat.com>
Suggested-by: Wincy Van <fanwenyi0529@gmail.com>
Cc: Paolo Bonzini <pbonzini@redhat.com>
Cc: Radim Krčmář <rkrcmar@redhat.com>
Cc: Wincy Van <fanwenyi0529@gmail.com>
Cc: Yang Zhang <yang.zhang.wz@gmail.com>
Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
---
v2 -> v3:
 * introduce a new set of bitmaps and assign them in vmx_set_msr_bitmap()
v1 -> v2: 
 * leverage the cached msr bitmap

 arch/x86/kvm/vmx.c | 133 ++++++++++++++++++++++++++++++++++++++---------------
 1 file changed, 95 insertions(+), 38 deletions(-)

diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
index 5cede40..4f3042c 100644
--- a/arch/x86/kvm/vmx.c
+++ b/arch/x86/kvm/vmx.c
@@ -927,6 +927,8 @@ static unsigned long *vmx_msr_bitmap_legacy;
 static unsigned long *vmx_msr_bitmap_longmode;
 static unsigned long *vmx_msr_bitmap_legacy_x2apic;
 static unsigned long *vmx_msr_bitmap_longmode_x2apic;
+static unsigned long *vmx_msr_bitmap_legacy_x2apic_apicv_inactive;
+static unsigned long *vmx_msr_bitmap_longmode_x2apic_apicv_inactive;
 static unsigned long *vmx_vmread_bitmap;
 static unsigned long *vmx_vmwrite_bitmap;
 
@@ -2518,10 +2520,18 @@ static void vmx_set_msr_bitmap(struct kvm_vcpu *vcpu)
 	else if (cpu_has_secondary_exec_ctrls() &&
 		 (vmcs_read32(SECONDARY_VM_EXEC_CONTROL) &
 		  SECONDARY_EXEC_VIRTUALIZE_X2APIC_MODE)) {
-		if (is_long_mode(vcpu))
-			msr_bitmap = vmx_msr_bitmap_longmode_x2apic;
-		else
-			msr_bitmap = vmx_msr_bitmap_legacy_x2apic;
+		if (enable_apicv && kvm_vcpu_apicv_active(vcpu)) {
+			if (is_long_mode(vcpu))
+				msr_bitmap = vmx_msr_bitmap_longmode_x2apic;
+			else
+				msr_bitmap = vmx_msr_bitmap_legacy_x2apic;
+		} else if ((enable_apicv && !kvm_vcpu_apicv_active(vcpu)) ||
+				!enable_apicv) {
+			if (is_long_mode(vcpu))
+				msr_bitmap = vmx_msr_bitmap_longmode_x2apic_apicv_inactive;
+			else
+				msr_bitmap = vmx_msr_bitmap_legacy_x2apic_apicv_inactive;
+		}
 	} else {
 		if (is_long_mode(vcpu))
 			msr_bitmap = vmx_msr_bitmap_longmode;
@@ -4678,28 +4688,49 @@ static void vmx_disable_intercept_for_msr(u32 msr, bool longmode_only)
 						msr, MSR_TYPE_R | MSR_TYPE_W);
 }
 
-static void vmx_enable_intercept_msr_read_x2apic(u32 msr)
+static void vmx_enable_intercept_msr_read_x2apic(u32 msr, bool apicv_active)
 {
-	__vmx_enable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
-			msr, MSR_TYPE_R);
-	__vmx_enable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
-			msr, MSR_TYPE_R);
+	if (apicv_active) {
+		__vmx_enable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
+				msr, MSR_TYPE_R);
+		__vmx_enable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
+				msr, MSR_TYPE_R);
+	} else {
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
+				msr, MSR_TYPE_R);
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
+				msr, MSR_TYPE_R);
+	}
 }
 
-static void vmx_disable_intercept_msr_read_x2apic(u32 msr)
+static void vmx_disable_intercept_msr_read_x2apic(u32 msr, bool apicv_active)
 {
-	__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
-			msr, MSR_TYPE_R);
-	__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
-			msr, MSR_TYPE_R);
+	if (apicv_active) {
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
+				msr, MSR_TYPE_R);
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
+				msr, MSR_TYPE_R);
+	} else {
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
+				msr, MSR_TYPE_R);
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
+				msr, MSR_TYPE_R);
+	}
 }
 
-static void vmx_disable_intercept_msr_write_x2apic(u32 msr)
+static void vmx_disable_intercept_msr_write_x2apic(u32 msr, bool apicv_active)
 {
-	__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
-			msr, MSR_TYPE_W);
-	__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
-			msr, MSR_TYPE_W);
+	if (apicv_active) {
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
+				msr, MSR_TYPE_W);
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
+				msr, MSR_TYPE_W);
+	} else {
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
+				msr, MSR_TYPE_W);
+		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
+				msr, MSR_TYPE_W);
+	}
 }
 
 static bool vmx_get_enable_apicv(void)
@@ -6360,22 +6391,32 @@ static __init int hardware_setup(void)
 	if (!vmx_msr_bitmap_legacy_x2apic)
 		goto out2;
 
+	vmx_msr_bitmap_legacy_x2apic_apicv_inactive =
+				(unsigned long *)__get_free_page(GFP_KERNEL);
+	if (!vmx_msr_bitmap_legacy_x2apic_apicv_inactive)
+		goto out3;
+
 	vmx_msr_bitmap_longmode = (unsigned long *)__get_free_page(GFP_KERNEL);
 	if (!vmx_msr_bitmap_longmode)
-		goto out3;
+		goto out4;
 
 	vmx_msr_bitmap_longmode_x2apic =
 				(unsigned long *)__get_free_page(GFP_KERNEL);
 	if (!vmx_msr_bitmap_longmode_x2apic)
-		goto out4;
+		goto out5;
+
+	vmx_msr_bitmap_longmode_x2apic_apicv_inactive =
+				(unsigned long *)__get_free_page(GFP_KERNEL);
+	if (!vmx_msr_bitmap_longmode_x2apic_apicv_inactive)
+		goto out6;
 
 	vmx_vmread_bitmap = (unsigned long *)__get_free_page(GFP_KERNEL);
 	if (!vmx_vmread_bitmap)
-		goto out6;
+		goto out7;
 
 	vmx_vmwrite_bitmap = (unsigned long *)__get_free_page(GFP_KERNEL);
 	if (!vmx_vmwrite_bitmap)
-		goto out7;
+		goto out8;
 
 	memset(vmx_vmread_bitmap, 0xff, PAGE_SIZE);
 	memset(vmx_vmwrite_bitmap, 0xff, PAGE_SIZE);
@@ -6394,7 +6435,7 @@ static __init int hardware_setup(void)
 
 	if (setup_vmcs_config(&vmcs_config) < 0) {
 		r = -EIO;
-		goto out8;
+		goto out9;
 	}
 
 	if (boot_cpu_has(X86_FEATURE_NX))
@@ -6461,20 +6502,35 @@ static __init int hardware_setup(void)
 			vmx_msr_bitmap_legacy, PAGE_SIZE);
 	memcpy(vmx_msr_bitmap_longmode_x2apic,
 			vmx_msr_bitmap_longmode, PAGE_SIZE);
+	memcpy(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
+			vmx_msr_bitmap_legacy, PAGE_SIZE);
+	memcpy(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
+			vmx_msr_bitmap_longmode, PAGE_SIZE);
 
 	set_bit(0, vmx_vpid_bitmap); /* 0 is reserved for host */
 
+	/*
+	 * enable_apicv && kvm_vcpu_apicv_active()
+	 */
 	for (msr = 0x800; msr <= 0x8ff; msr++)
-		vmx_disable_intercept_msr_read_x2apic(msr);
+		vmx_disable_intercept_msr_read_x2apic(msr, true);
 
 	/* TMCCT */
-	vmx_enable_intercept_msr_read_x2apic(0x839);
+	vmx_enable_intercept_msr_read_x2apic(0x839, true);
 	/* TPR */
-	vmx_disable_intercept_msr_write_x2apic(0x808);
+	vmx_disable_intercept_msr_write_x2apic(0x808, true);
 	/* EOI */
-	vmx_disable_intercept_msr_write_x2apic(0x80b);
+	vmx_disable_intercept_msr_write_x2apic(0x80b, true);
 	/* SELF-IPI */
-	vmx_disable_intercept_msr_write_x2apic(0x83f);
+	vmx_disable_intercept_msr_write_x2apic(0x83f, true);
+
+	/*
+	 * (enable_apicv && !kvm_vcpu_apicv_active()) ||
+	 * 	!enable_apicv
+	 */
+	/* TPR */
+	vmx_disable_intercept_msr_read_x2apic(0x808, false);
+	vmx_disable_intercept_msr_write_x2apic(0x808, false);
 
 	if (enable_ept) {
 		kvm_mmu_set_mask_ptes(VMX_EPT_READABLE_MASK,
@@ -6521,14 +6577,18 @@ static __init int hardware_setup(void)
 
 	return alloc_kvm_area();
 
-out8:
+out9:
 	free_page((unsigned long)vmx_vmwrite_bitmap);
-out7:
+out8:
 	free_page((unsigned long)vmx_vmread_bitmap);
+out7:
+	free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic_apicv_inactive);
 out6:
 	free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic);
-out4:
+out5:
 	free_page((unsigned long)vmx_msr_bitmap_longmode);
+out4:
+	free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic_apicv_inactive);
 out3:
 	free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic);
 out2:
@@ -6544,7 +6604,9 @@ out:
 static __exit void hardware_unsetup(void)
 {
 	free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic);
+	free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic_apicv_inactive);
 	free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic);
+	free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic_apicv_inactive);
 	free_page((unsigned long)vmx_msr_bitmap_legacy);
 	free_page((unsigned long)vmx_msr_bitmap_longmode);
 	free_page((unsigned long)vmx_io_bitmap_b);
@@ -8435,12 +8497,7 @@ static void vmx_set_virtual_x2apic_mode(struct kvm_vcpu *vcpu, bool set)
 		return;
 	}
 
-	/*
-	 * There is not point to enable virtualize x2apic without enable
-	 * apicv
-	 */
-	if (!cpu_has_vmx_virtualize_x2apic_mode() ||
-				!kvm_vcpu_apicv_active(vcpu))
+	if (!cpu_has_vmx_virtualize_x2apic_mode())
 		return;
 
 	if (!cpu_need_tpr_shadow(vcpu))
-- 
1.9.1

^ permalink raw reply related	[flat|nested] 4+ messages in thread

* Re: [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive
  2016-09-21  2:14 [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive Wanpeng Li
@ 2016-09-21  2:38 ` Wanpeng Li
  2016-09-21 16:35 ` Radim Krčmář
  1 sibling, 0 replies; 4+ messages in thread
From: Wanpeng Li @ 2016-09-21  2:38 UTC (permalink / raw)
  To: linux-kernel, kvm
  Cc: Wanpeng Li, Paolo Bonzini, Radim Krčmář,
	Wincy Van, Yang Zhang

The subject should be "KVM: VMX: Enable MSR-BASED TPR shadow even if
APICv is inactive"
2016-09-21 10:14 GMT+08:00 Wanpeng Li <kernellwp@gmail.com>:
> From: Wanpeng Li <wanpeng.li@hotmail.com>
>
> I observed that kvmvapic(to optimize flexpriority=N or AMD) is used
> to boost TPR access when testing kvm-unit-test/eventinj.flat tpr case
> on my haswell desktop (w/ flexpriority, w/o APICv). Commit (8d14695f9542
> x86, apicv: add virtual x2apic support) disable virtual x2apic mode
> completely if w/o APICv, and the author also told me that windows guest
> can't enter into x2apic mode when he developed the APICv feature several
> years ago. However, it is not truth currently, Interrupt Remapping and
> vIOMMU is added to qemu and the developers from Intel test windows 8 can
> work in x2apic mode w/ Interrupt Remapping enabled recently.
>
> This patch enables TPR shadow for virtual x2apic mode to boost
> windows guest in x2apic mode even if w/o APICv.
>
> Can pass the kvm-unit-test.
>
> Suggested-by: Radim Krčmář <rkrcmar@redhat.com>
> Suggested-by: Wincy Van <fanwenyi0529@gmail.com>
> Cc: Paolo Bonzini <pbonzini@redhat.com>
> Cc: Radim Krčmář <rkrcmar@redhat.com>
> Cc: Wincy Van <fanwenyi0529@gmail.com>
> Cc: Yang Zhang <yang.zhang.wz@gmail.com>
> Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
> ---
> v2 -> v3:
>  * introduce a new set of bitmaps and assign them in vmx_set_msr_bitmap()
> v1 -> v2:
>  * leverage the cached msr bitmap
>
>  arch/x86/kvm/vmx.c | 133 ++++++++++++++++++++++++++++++++++++++---------------
>  1 file changed, 95 insertions(+), 38 deletions(-)
>
> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
> index 5cede40..4f3042c 100644
> --- a/arch/x86/kvm/vmx.c
> +++ b/arch/x86/kvm/vmx.c
> @@ -927,6 +927,8 @@ static unsigned long *vmx_msr_bitmap_legacy;
>  static unsigned long *vmx_msr_bitmap_longmode;
>  static unsigned long *vmx_msr_bitmap_legacy_x2apic;
>  static unsigned long *vmx_msr_bitmap_longmode_x2apic;
> +static unsigned long *vmx_msr_bitmap_legacy_x2apic_apicv_inactive;
> +static unsigned long *vmx_msr_bitmap_longmode_x2apic_apicv_inactive;
>  static unsigned long *vmx_vmread_bitmap;
>  static unsigned long *vmx_vmwrite_bitmap;
>
> @@ -2518,10 +2520,18 @@ static void vmx_set_msr_bitmap(struct kvm_vcpu *vcpu)
>         else if (cpu_has_secondary_exec_ctrls() &&
>                  (vmcs_read32(SECONDARY_VM_EXEC_CONTROL) &
>                   SECONDARY_EXEC_VIRTUALIZE_X2APIC_MODE)) {
> -               if (is_long_mode(vcpu))
> -                       msr_bitmap = vmx_msr_bitmap_longmode_x2apic;
> -               else
> -                       msr_bitmap = vmx_msr_bitmap_legacy_x2apic;
> +               if (enable_apicv && kvm_vcpu_apicv_active(vcpu)) {
> +                       if (is_long_mode(vcpu))
> +                               msr_bitmap = vmx_msr_bitmap_longmode_x2apic;
> +                       else
> +                               msr_bitmap = vmx_msr_bitmap_legacy_x2apic;
> +               } else if ((enable_apicv && !kvm_vcpu_apicv_active(vcpu)) ||
> +                               !enable_apicv) {
> +                       if (is_long_mode(vcpu))
> +                               msr_bitmap = vmx_msr_bitmap_longmode_x2apic_apicv_inactive;
> +                       else
> +                               msr_bitmap = vmx_msr_bitmap_legacy_x2apic_apicv_inactive;
> +               }
>         } else {
>                 if (is_long_mode(vcpu))
>                         msr_bitmap = vmx_msr_bitmap_longmode;
> @@ -4678,28 +4688,49 @@ static void vmx_disable_intercept_for_msr(u32 msr, bool longmode_only)
>                                                 msr, MSR_TYPE_R | MSR_TYPE_W);
>  }
>
> -static void vmx_enable_intercept_msr_read_x2apic(u32 msr)
> +static void vmx_enable_intercept_msr_read_x2apic(u32 msr, bool apicv_active)
>  {
> -       __vmx_enable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> -                       msr, MSR_TYPE_R);
> -       __vmx_enable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> -                       msr, MSR_TYPE_R);
> +       if (apicv_active) {
> +               __vmx_enable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> +                               msr, MSR_TYPE_R);
> +               __vmx_enable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> +                               msr, MSR_TYPE_R);
> +       } else {
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
> +                               msr, MSR_TYPE_R);
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
> +                               msr, MSR_TYPE_R);
> +       }
>  }
>
> -static void vmx_disable_intercept_msr_read_x2apic(u32 msr)
> +static void vmx_disable_intercept_msr_read_x2apic(u32 msr, bool apicv_active)
>  {
> -       __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> -                       msr, MSR_TYPE_R);
> -       __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> -                       msr, MSR_TYPE_R);
> +       if (apicv_active) {
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> +                               msr, MSR_TYPE_R);
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> +                               msr, MSR_TYPE_R);
> +       } else {
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
> +                               msr, MSR_TYPE_R);
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
> +                               msr, MSR_TYPE_R);
> +       }
>  }
>
> -static void vmx_disable_intercept_msr_write_x2apic(u32 msr)
> +static void vmx_disable_intercept_msr_write_x2apic(u32 msr, bool apicv_active)
>  {
> -       __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> -                       msr, MSR_TYPE_W);
> -       __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> -                       msr, MSR_TYPE_W);
> +       if (apicv_active) {
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> +                               msr, MSR_TYPE_W);
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> +                               msr, MSR_TYPE_W);
> +       } else {
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
> +                               msr, MSR_TYPE_W);
> +               __vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
> +                               msr, MSR_TYPE_W);
> +       }
>  }
>
>  static bool vmx_get_enable_apicv(void)
> @@ -6360,22 +6391,32 @@ static __init int hardware_setup(void)
>         if (!vmx_msr_bitmap_legacy_x2apic)
>                 goto out2;
>
> +       vmx_msr_bitmap_legacy_x2apic_apicv_inactive =
> +                               (unsigned long *)__get_free_page(GFP_KERNEL);
> +       if (!vmx_msr_bitmap_legacy_x2apic_apicv_inactive)
> +               goto out3;
> +
>         vmx_msr_bitmap_longmode = (unsigned long *)__get_free_page(GFP_KERNEL);
>         if (!vmx_msr_bitmap_longmode)
> -               goto out3;
> +               goto out4;
>
>         vmx_msr_bitmap_longmode_x2apic =
>                                 (unsigned long *)__get_free_page(GFP_KERNEL);
>         if (!vmx_msr_bitmap_longmode_x2apic)
> -               goto out4;
> +               goto out5;
> +
> +       vmx_msr_bitmap_longmode_x2apic_apicv_inactive =
> +                               (unsigned long *)__get_free_page(GFP_KERNEL);
> +       if (!vmx_msr_bitmap_longmode_x2apic_apicv_inactive)
> +               goto out6;
>
>         vmx_vmread_bitmap = (unsigned long *)__get_free_page(GFP_KERNEL);
>         if (!vmx_vmread_bitmap)
> -               goto out6;
> +               goto out7;
>
>         vmx_vmwrite_bitmap = (unsigned long *)__get_free_page(GFP_KERNEL);
>         if (!vmx_vmwrite_bitmap)
> -               goto out7;
> +               goto out8;
>
>         memset(vmx_vmread_bitmap, 0xff, PAGE_SIZE);
>         memset(vmx_vmwrite_bitmap, 0xff, PAGE_SIZE);
> @@ -6394,7 +6435,7 @@ static __init int hardware_setup(void)
>
>         if (setup_vmcs_config(&vmcs_config) < 0) {
>                 r = -EIO;
> -               goto out8;
> +               goto out9;
>         }
>
>         if (boot_cpu_has(X86_FEATURE_NX))
> @@ -6461,20 +6502,35 @@ static __init int hardware_setup(void)
>                         vmx_msr_bitmap_legacy, PAGE_SIZE);
>         memcpy(vmx_msr_bitmap_longmode_x2apic,
>                         vmx_msr_bitmap_longmode, PAGE_SIZE);
> +       memcpy(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
> +                       vmx_msr_bitmap_legacy, PAGE_SIZE);
> +       memcpy(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
> +                       vmx_msr_bitmap_longmode, PAGE_SIZE);
>
>         set_bit(0, vmx_vpid_bitmap); /* 0 is reserved for host */
>
> +       /*
> +        * enable_apicv && kvm_vcpu_apicv_active()
> +        */
>         for (msr = 0x800; msr <= 0x8ff; msr++)
> -               vmx_disable_intercept_msr_read_x2apic(msr);
> +               vmx_disable_intercept_msr_read_x2apic(msr, true);
>
>         /* TMCCT */
> -       vmx_enable_intercept_msr_read_x2apic(0x839);
> +       vmx_enable_intercept_msr_read_x2apic(0x839, true);
>         /* TPR */
> -       vmx_disable_intercept_msr_write_x2apic(0x808);
> +       vmx_disable_intercept_msr_write_x2apic(0x808, true);
>         /* EOI */
> -       vmx_disable_intercept_msr_write_x2apic(0x80b);
> +       vmx_disable_intercept_msr_write_x2apic(0x80b, true);
>         /* SELF-IPI */
> -       vmx_disable_intercept_msr_write_x2apic(0x83f);
> +       vmx_disable_intercept_msr_write_x2apic(0x83f, true);
> +
> +       /*
> +        * (enable_apicv && !kvm_vcpu_apicv_active()) ||
> +        *      !enable_apicv
> +        */
> +       /* TPR */
> +       vmx_disable_intercept_msr_read_x2apic(0x808, false);
> +       vmx_disable_intercept_msr_write_x2apic(0x808, false);
>
>         if (enable_ept) {
>                 kvm_mmu_set_mask_ptes(VMX_EPT_READABLE_MASK,
> @@ -6521,14 +6577,18 @@ static __init int hardware_setup(void)
>
>         return alloc_kvm_area();
>
> -out8:
> +out9:
>         free_page((unsigned long)vmx_vmwrite_bitmap);
> -out7:
> +out8:
>         free_page((unsigned long)vmx_vmread_bitmap);
> +out7:
> +       free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic_apicv_inactive);
>  out6:
>         free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic);
> -out4:
> +out5:
>         free_page((unsigned long)vmx_msr_bitmap_longmode);
> +out4:
> +       free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic_apicv_inactive);
>  out3:
>         free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic);
>  out2:
> @@ -6544,7 +6604,9 @@ out:
>  static __exit void hardware_unsetup(void)
>  {
>         free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic);
> +       free_page((unsigned long)vmx_msr_bitmap_legacy_x2apic_apicv_inactive);
>         free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic);
> +       free_page((unsigned long)vmx_msr_bitmap_longmode_x2apic_apicv_inactive);
>         free_page((unsigned long)vmx_msr_bitmap_legacy);
>         free_page((unsigned long)vmx_msr_bitmap_longmode);
>         free_page((unsigned long)vmx_io_bitmap_b);
> @@ -8435,12 +8497,7 @@ static void vmx_set_virtual_x2apic_mode(struct kvm_vcpu *vcpu, bool set)
>                 return;
>         }
>
> -       /*
> -        * There is not point to enable virtualize x2apic without enable
> -        * apicv
> -        */
> -       if (!cpu_has_vmx_virtualize_x2apic_mode() ||
> -                               !kvm_vcpu_apicv_active(vcpu))
> +       if (!cpu_has_vmx_virtualize_x2apic_mode())
>                 return;
>
>         if (!cpu_need_tpr_shadow(vcpu))
> --
> 1.9.1
>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive
  2016-09-21  2:14 [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive Wanpeng Li
  2016-09-21  2:38 ` Wanpeng Li
@ 2016-09-21 16:35 ` Radim Krčmář
  2016-09-21 23:17   ` Wanpeng Li
  1 sibling, 1 reply; 4+ messages in thread
From: Radim Krčmář @ 2016-09-21 16:35 UTC (permalink / raw)
  To: Wanpeng Li
  Cc: linux-kernel, kvm, Wanpeng Li, Paolo Bonzini, Wincy Van, Yang Zhang

2016-09-21 10:14+0800, Wanpeng Li:
> From: Wanpeng Li <wanpeng.li@hotmail.com>
> 
> I observed that kvmvapic(to optimize flexpriority=N or AMD) is used 
> to boost TPR access when testing kvm-unit-test/eventinj.flat tpr case
> on my haswell desktop (w/ flexpriority, w/o APICv). Commit (8d14695f9542 
> x86, apicv: add virtual x2apic support) disable virtual x2apic mode 
> completely if w/o APICv, and the author also told me that windows guest
> can't enter into x2apic mode when he developed the APICv feature several 
> years ago. However, it is not truth currently, Interrupt Remapping and 
> vIOMMU is added to qemu and the developers from Intel test windows 8 can 
> work in x2apic mode w/ Interrupt Remapping enabled recently. 
> 
> This patch enables TPR shadow for virtual x2apic mode to boost 
> windows guest in x2apic mode even if w/o APICv.
> 
> Can pass the kvm-unit-test.
> 
> Suggested-by: Radim Krčmář <rkrcmar@redhat.com>
> Suggested-by: Wincy Van <fanwenyi0529@gmail.com>
> Cc: Paolo Bonzini <pbonzini@redhat.com>
> Cc: Radim Krčmář <rkrcmar@redhat.com>
> Cc: Wincy Van <fanwenyi0529@gmail.com>
> Cc: Yang Zhang <yang.zhang.wz@gmail.com>
> Signed-off-by: Wanpeng Li <wanpeng.li@hotmail.com>
> ---
> v2 -> v3:
>  * introduce a new set of bitmaps and assign them in vmx_set_msr_bitmap()
> v1 -> v2: 
>  * leverage the cached msr bitmap
> 
> diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
> @@ -2518,10 +2520,18 @@ static void vmx_set_msr_bitmap(struct kvm_vcpu *vcpu)
>  	else if (cpu_has_secondary_exec_ctrls() &&
>  		 (vmcs_read32(SECONDARY_VM_EXEC_CONTROL) &
>  		  SECONDARY_EXEC_VIRTUALIZE_X2APIC_MODE)) {
> -		if (is_long_mode(vcpu))
> -			msr_bitmap = vmx_msr_bitmap_longmode_x2apic;
> -		else
> -			msr_bitmap = vmx_msr_bitmap_legacy_x2apic;
> +		if (enable_apicv && kvm_vcpu_apicv_active(vcpu)) {
> +			if (is_long_mode(vcpu))
> +				msr_bitmap = vmx_msr_bitmap_longmode_x2apic;
> +			else
> +				msr_bitmap = vmx_msr_bitmap_legacy_x2apic;
> +		} else if ((enable_apicv && !kvm_vcpu_apicv_active(vcpu)) ||
> +				!enable_apicv) {

Could only be "else" without if here.  I'll change that when refactoring
of vmx bitmap handling, posting soon (hopefully).

> +			if (is_long_mode(vcpu))
> +				msr_bitmap = vmx_msr_bitmap_longmode_x2apic_apicv_inactive;
> +			else
> +				msr_bitmap = vmx_msr_bitmap_legacy_x2apic_apicv_inactive;
> +		}
>  	} else {
>  		if (is_long_mode(vcpu))
>  			msr_bitmap = vmx_msr_bitmap_longmode;
> @@ -4678,28 +4688,49 @@ static void vmx_disable_intercept_for_msr(u32 msr, bool longmode_only)
> -static void vmx_enable_intercept_msr_read_x2apic(u32 msr)
> +static void vmx_enable_intercept_msr_read_x2apic(u32 msr, bool apicv_active)
>  {
> -	__vmx_enable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> -			msr, MSR_TYPE_R);
> -	__vmx_enable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> -			msr, MSR_TYPE_R);
> +	if (apicv_active) {
> +		__vmx_enable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic,
> +				msr, MSR_TYPE_R);
> +		__vmx_enable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic,
> +				msr, MSR_TYPE_R);
> +	} else {
> +		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_legacy_x2apic_apicv_inactive,
> +				msr, MSR_TYPE_R);
> +		__vmx_disable_intercept_for_msr(vmx_msr_bitmap_longmode_x2apic_apicv_inactive,
> +				msr, MSR_TYPE_R);

Pasto here (disable vs. enable).  It's not called with !apicv_active and
the refactoring will delete the whole function, so

Reviwed-by: Radim Krčmář <rkrcmar@redhat.com>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive
  2016-09-21 16:35 ` Radim Krčmář
@ 2016-09-21 23:17   ` Wanpeng Li
  0 siblings, 0 replies; 4+ messages in thread
From: Wanpeng Li @ 2016-09-21 23:17 UTC (permalink / raw)
  To: Radim Krčmář
  Cc: linux-kernel, kvm, Wanpeng Li, Paolo Bonzini, Wincy Van, Yang Zhang

2016-09-22 0:35 GMT+08:00 Radim Krčmář <rkrcmar@redhat.com>:
> 2016-09-21 10:14+0800, Wanpeng Li:
[...]
> Could only be "else" without if here.  I'll change that when refactoring
> of vmx bitmap handling, posting soon (hopefully).
>
[...]
>
> Pasto here (disable vs. enable).  It's not called with !apicv_active and
> the refactoring will delete the whole function, so

Your refactoring is based on this patch, right? So I will send out v4
to fix the issue which you pointed out in this patch for further
bisect.

Regards,
Wanpeng Li

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2016-09-21 23:17 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2016-09-21  2:14 [PATCH v3] Enable MSR-BASED TPR shadow even if APICv is inactive Wanpeng Li
2016-09-21  2:38 ` Wanpeng Li
2016-09-21 16:35 ` Radim Krčmář
2016-09-21 23:17   ` Wanpeng Li

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.