* [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable
2020-07-25 3:26 [PATCH 0/5 v2] KVM: x86: Fill in conforming {vmx|svm}_x86_ops and {vmx|svm}_nested_ops via macros Krish Sadhukhan
@ 2020-07-25 3:26 ` Krish Sadhukhan
2020-07-25 18:41 ` kernel test robot
2020-07-27 9:39 ` Vitaly Kuznetsov
2020-07-25 3:26 ` [PATCH 2/5 v2] KVM: SVM: Fill in conforming svm_x86_ops via macro Krish Sadhukhan
` (3 subsequent siblings)
4 siblings, 2 replies; 10+ messages in thread
From: Krish Sadhukhan @ 2020-07-25 3:26 UTC (permalink / raw)
To: kvm; +Cc: jmattson, sean.j.christopherson, pbonzini, vkuznets
Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
---
arch/x86/include/asm/kvm_host.h | 12 ++++++------
arch/x86/kvm/svm/svm.c | 12 ++++++------
arch/x86/kvm/vmx/vmx.c | 8 ++++----
arch/x86/kvm/x86.c | 22 +++++++++++-----------
include/linux/kvm_host.h | 2 +-
virt/kvm/kvm_main.c | 4 ++--
6 files changed, 30 insertions(+), 30 deletions(-)
diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
index be5363b..ccad66d 100644
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -1080,7 +1080,7 @@ static inline u16 kvm_lapic_irq_dest_mode(bool dest_mode_logical)
struct kvm_x86_ops {
int (*hardware_enable)(void);
void (*hardware_disable)(void);
- void (*hardware_unsetup)(void);
+ void (*hardware_teardown)(void);
bool (*cpu_has_accelerated_tpr)(void);
bool (*has_emulated_msr)(u32 index);
void (*cpuid_update)(struct kvm_vcpu *vcpu);
@@ -1141,7 +1141,7 @@ struct kvm_x86_ops {
*/
void (*tlb_flush_guest)(struct kvm_vcpu *vcpu);
- enum exit_fastpath_completion (*run)(struct kvm_vcpu *vcpu);
+ enum exit_fastpath_completion (*vcpu_run)(struct kvm_vcpu *vcpu);
int (*handle_exit)(struct kvm_vcpu *vcpu,
enum exit_fastpath_completion exit_fastpath);
int (*skip_emulated_instruction)(struct kvm_vcpu *vcpu);
@@ -1150,8 +1150,8 @@ struct kvm_x86_ops {
u32 (*get_interrupt_shadow)(struct kvm_vcpu *vcpu);
void (*patch_hypercall)(struct kvm_vcpu *vcpu,
unsigned char *hypercall_addr);
- void (*set_irq)(struct kvm_vcpu *vcpu);
- void (*set_nmi)(struct kvm_vcpu *vcpu);
+ void (*inject_irq)(struct kvm_vcpu *vcpu);
+ void (*inject_nmi)(struct kvm_vcpu *vcpu);
void (*queue_exception)(struct kvm_vcpu *vcpu);
void (*cancel_injection)(struct kvm_vcpu *vcpu);
int (*interrupt_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
@@ -1258,8 +1258,8 @@ struct kvm_x86_ops {
void (*enable_smi_window)(struct kvm_vcpu *vcpu);
int (*mem_enc_op)(struct kvm *kvm, void __user *argp);
- int (*mem_enc_reg_region)(struct kvm *kvm, struct kvm_enc_region *argp);
- int (*mem_enc_unreg_region)(struct kvm *kvm, struct kvm_enc_region *argp);
+ int (*mem_enc_register_region)(struct kvm *kvm, struct kvm_enc_region *argp);
+ int (*mem_enc_unregister_region)(struct kvm *kvm, struct kvm_enc_region *argp);
int (*get_msr_feature)(struct kvm_msr_entry *entry);
diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
index c0da4dd..24755eb 100644
--- a/arch/x86/kvm/svm/svm.c
+++ b/arch/x86/kvm/svm/svm.c
@@ -3969,7 +3969,7 @@ static int svm_vm_init(struct kvm *kvm)
}
static struct kvm_x86_ops svm_x86_ops __initdata = {
- .hardware_unsetup = svm_hardware_teardown,
+ .hardware_teardown = svm_hardware_teardown,
.hardware_enable = svm_hardware_enable,
.hardware_disable = svm_hardware_disable,
.cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr,
@@ -4016,15 +4016,15 @@ static int svm_vm_init(struct kvm *kvm)
.tlb_flush_gva = svm_flush_tlb_gva,
.tlb_flush_guest = svm_flush_tlb,
- .run = svm_vcpu_run,
+ .vcpu_run = svm_vcpu_run,
.handle_exit = handle_exit,
.skip_emulated_instruction = skip_emulated_instruction,
.update_emulated_instruction = NULL,
.set_interrupt_shadow = svm_set_interrupt_shadow,
.get_interrupt_shadow = svm_get_interrupt_shadow,
.patch_hypercall = svm_patch_hypercall,
- .set_irq = svm_set_irq,
- .set_nmi = svm_inject_nmi,
+ .inject_irq = svm_set_irq,
+ .inject_nmi = svm_inject_nmi,
.queue_exception = svm_queue_exception,
.cancel_injection = svm_cancel_injection,
.interrupt_allowed = svm_interrupt_allowed,
@@ -4080,8 +4080,8 @@ static int svm_vm_init(struct kvm *kvm)
.enable_smi_window = enable_smi_window,
.mem_enc_op = svm_mem_enc_op,
- .mem_enc_reg_region = svm_register_enc_region,
- .mem_enc_unreg_region = svm_unregister_enc_region,
+ .mem_enc_register_region = svm_register_enc_region,
+ .mem_enc_unregister_region = svm_unregister_enc_region,
.need_emulation_on_page_fault = svm_need_emulation_on_page_fault,
diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
index cb22f33..90d91524 100644
--- a/arch/x86/kvm/vmx/vmx.c
+++ b/arch/x86/kvm/vmx/vmx.c
@@ -7844,7 +7844,7 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
}
static struct kvm_x86_ops vmx_x86_ops __initdata = {
- .hardware_unsetup = hardware_unsetup,
+ .hardware_teardown = hardware_unsetup,
.hardware_enable = hardware_enable,
.hardware_disable = hardware_disable,
@@ -7889,15 +7889,15 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
.tlb_flush_gva = vmx_flush_tlb_gva,
.tlb_flush_guest = vmx_flush_tlb_guest,
- .run = vmx_vcpu_run,
+ .vcpu_run = vmx_vcpu_run,
.handle_exit = vmx_handle_exit,
.skip_emulated_instruction = vmx_skip_emulated_instruction,
.update_emulated_instruction = vmx_update_emulated_instruction,
.set_interrupt_shadow = vmx_set_interrupt_shadow,
.get_interrupt_shadow = vmx_get_interrupt_shadow,
.patch_hypercall = vmx_patch_hypercall,
- .set_irq = vmx_inject_irq,
- .set_nmi = vmx_inject_nmi,
+ .inject_irq = vmx_inject_irq,
+ .inject_nmi = vmx_inject_nmi,
.queue_exception = vmx_queue_exception,
.cancel_injection = vmx_cancel_injection,
.interrupt_allowed = vmx_interrupt_allowed,
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 3b92db4..e850fb3 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -5270,8 +5270,8 @@ long kvm_arch_vm_ioctl(struct file *filp,
goto out;
r = -ENOTTY;
- if (kvm_x86_ops.mem_enc_reg_region)
- r = kvm_x86_ops.mem_enc_reg_region(kvm, ®ion);
+ if (kvm_x86_ops.mem_enc_register_region)
+ r = kvm_x86_ops.mem_enc_register_region(kvm, ®ion);
break;
}
case KVM_MEMORY_ENCRYPT_UNREG_REGION: {
@@ -5282,8 +5282,8 @@ long kvm_arch_vm_ioctl(struct file *filp,
goto out;
r = -ENOTTY;
- if (kvm_x86_ops.mem_enc_unreg_region)
- r = kvm_x86_ops.mem_enc_unreg_region(kvm, ®ion);
+ if (kvm_x86_ops.mem_enc_unregister_region)
+ r = kvm_x86_ops.mem_enc_unregister_region(kvm, ®ion);
break;
}
case KVM_HYPERV_EVENTFD: {
@@ -7788,10 +7788,10 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
*/
else if (!vcpu->arch.exception.pending) {
if (vcpu->arch.nmi_injected) {
- kvm_x86_ops.set_nmi(vcpu);
+ kvm_x86_ops.inject_nmi(vcpu);
can_inject = false;
} else if (vcpu->arch.interrupt.injected) {
- kvm_x86_ops.set_irq(vcpu);
+ kvm_x86_ops.inject_irq(vcpu);
can_inject = false;
}
}
@@ -7867,7 +7867,7 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
if (r) {
--vcpu->arch.nmi_pending;
vcpu->arch.nmi_injected = true;
- kvm_x86_ops.set_nmi(vcpu);
+ kvm_x86_ops.inject_nmi(vcpu);
can_inject = false;
WARN_ON(kvm_x86_ops.nmi_allowed(vcpu, true) < 0);
}
@@ -7881,7 +7881,7 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
goto busy;
if (r) {
kvm_queue_interrupt(vcpu, kvm_cpu_get_interrupt(vcpu), false);
- kvm_x86_ops.set_irq(vcpu);
+ kvm_x86_ops.inject_irq(vcpu);
WARN_ON(kvm_x86_ops.interrupt_allowed(vcpu, true) < 0);
}
if (kvm_cpu_has_injectable_intr(vcpu))
@@ -8517,7 +8517,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_RELOAD;
}
- exit_fastpath = kvm_x86_ops.run(vcpu);
+ exit_fastpath = kvm_x86_ops.vcpu_run(vcpu);
/*
* Do this here before restoring debug registers on the host. And
@@ -9793,9 +9793,9 @@ int kvm_arch_hardware_setup(void *opaque)
return 0;
}
-void kvm_arch_hardware_unsetup(void)
+void kvm_arch_hardware_teardown(void)
{
- kvm_x86_ops.hardware_unsetup();
+ kvm_x86_ops.hardware_teardown();
}
int kvm_arch_check_processor_compat(void *opaque)
diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
index d564855..b49312c 100644
--- a/include/linux/kvm_host.h
+++ b/include/linux/kvm_host.h
@@ -894,7 +894,7 @@ int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
int kvm_arch_hardware_enable(void);
void kvm_arch_hardware_disable(void);
int kvm_arch_hardware_setup(void *opaque);
-void kvm_arch_hardware_unsetup(void);
+void kvm_arch_hardware_teardown(void);
int kvm_arch_check_processor_compat(void *opaque);
int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu);
bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu);
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
index a852af5..4625f3a 100644
--- a/virt/kvm/kvm_main.c
+++ b/virt/kvm/kvm_main.c
@@ -4786,7 +4786,7 @@ int kvm_init(void *opaque, unsigned vcpu_size, unsigned vcpu_align,
unregister_reboot_notifier(&kvm_reboot_notifier);
cpuhp_remove_state_nocalls(CPUHP_AP_KVM_STARTING);
out_free_2:
- kvm_arch_hardware_unsetup();
+ kvm_arch_hardware_teardown();
out_free_1:
free_cpumask_var(cpus_hardware_enabled);
out_free_0:
@@ -4808,7 +4808,7 @@ void kvm_exit(void)
unregister_reboot_notifier(&kvm_reboot_notifier);
cpuhp_remove_state_nocalls(CPUHP_AP_KVM_STARTING);
on_each_cpu(hardware_disable_nolock, NULL, 1);
- kvm_arch_hardware_unsetup();
+ kvm_arch_hardware_teardown();
kvm_arch_exit();
kvm_irqfd_exit();
free_cpumask_var(cpus_hardware_enabled);
--
1.8.3.1
^ permalink raw reply related [flat|nested] 10+ messages in thread
* Re: [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable
2020-07-25 3:26 ` [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable Krish Sadhukhan
@ 2020-07-25 18:41 ` kernel test robot
2020-07-27 9:39 ` Vitaly Kuznetsov
1 sibling, 0 replies; 10+ messages in thread
From: kernel test robot @ 2020-07-25 18:41 UTC (permalink / raw)
To: Krish Sadhukhan, kvm
Cc: kbuild-all, jmattson, sean.j.christopherson, pbonzini, vkuznets
[-- Attachment #1: Type: text/plain, Size: 2263 bytes --]
Hi Krish,
Thank you for the patch! Perhaps something to improve:
[auto build test WARNING on vhost/linux-next]
[also build test WARNING on v5.8-rc6 next-20200724]
[cannot apply to kvm/linux-next]
[If your patch is applied to the wrong git tree, kindly drop us a note.
And when submitting patch, we suggest to use '--base' as documented in
https://git-scm.com/docs/git-format-patch]
url: https://github.com/0day-ci/linux/commits/Krish-Sadhukhan/KVM-x86-Fill-in-conforming-vmx-svm-_x86_ops-and-vmx-svm-_nested_ops-via-macros/20200725-113105
base: https://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost.git linux-next
config: s390-allyesconfig (attached as .config)
compiler: s390-linux-gcc (GCC) 9.3.0
reproduce (this is a W=1 build):
wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross
chmod +x ~/bin/make.cross
# save the attached .config to linux build tree
COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-9.3.0 make.cross ARCH=s390
If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp@intel.com>
All warnings (new ones prefixed by >>):
>> arch/s390/kvm/kvm-s390.c:315:6: warning: no previous prototype for 'kvm_arch_hardware_unsetup' [-Wmissing-prototypes]
315 | void kvm_arch_hardware_unsetup(void)
| ^~~~~~~~~~~~~~~~~~~~~~~~~
vim +/kvm_arch_hardware_unsetup +315 arch/s390/kvm/kvm-s390.c
b0c632db637d68a Heiko Carstens 2008-03-25 314
b0c632db637d68a Heiko Carstens 2008-03-25 @315 void kvm_arch_hardware_unsetup(void)
b0c632db637d68a Heiko Carstens 2008-03-25 316 {
b2d73b2a0ad1c75 Martin Schwidefsky 2016-03-08 317 gmap_unregister_pte_notifier(&gmap_notifier);
a3508fbe9dc6dd3 David Hildenbrand 2015-07-08 318 gmap_unregister_pte_notifier(&vsie_gmap_notifier);
fdf036507f1fc03 Fan Zhang 2015-05-13 319 atomic_notifier_chain_unregister(&s390_epoch_delta_notifier,
fdf036507f1fc03 Fan Zhang 2015-05-13 320 &kvm_clock_notifier);
b0c632db637d68a Heiko Carstens 2008-03-25 321 }
b0c632db637d68a Heiko Carstens 2008-03-25 322
---
0-DAY CI Kernel Test Service, Intel Corporation
https://lists.01.org/hyperkitty/list/kbuild-all@lists.01.org
[-- Attachment #2: .config.gz --]
[-- Type: application/gzip, Size: 62639 bytes --]
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable
2020-07-25 3:26 ` [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable Krish Sadhukhan
2020-07-25 18:41 ` kernel test robot
@ 2020-07-27 9:39 ` Vitaly Kuznetsov
2020-07-27 17:30 ` Krish Sadhukhan
1 sibling, 1 reply; 10+ messages in thread
From: Vitaly Kuznetsov @ 2020-07-27 9:39 UTC (permalink / raw)
To: Krish Sadhukhan, kvm; +Cc: jmattson, sean.j.christopherson, pbonzini
Krish Sadhukhan <krish.sadhukhan@oracle.com> writes:
> Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
> Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
> Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
> Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
> ---
> arch/x86/include/asm/kvm_host.h | 12 ++++++------
> arch/x86/kvm/svm/svm.c | 12 ++++++------
> arch/x86/kvm/vmx/vmx.c | 8 ++++----
> arch/x86/kvm/x86.c | 22 +++++++++++-----------
> include/linux/kvm_host.h | 2 +-
> virt/kvm/kvm_main.c | 4 ++--
> 6 files changed, 30 insertions(+), 30 deletions(-)
>
> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
> index be5363b..ccad66d 100644
> --- a/arch/x86/include/asm/kvm_host.h
> +++ b/arch/x86/include/asm/kvm_host.h
> @@ -1080,7 +1080,7 @@ static inline u16 kvm_lapic_irq_dest_mode(bool dest_mode_logical)
> struct kvm_x86_ops {
> int (*hardware_enable)(void);
> void (*hardware_disable)(void);
> - void (*hardware_unsetup)(void);
> + void (*hardware_teardown)(void);
I definitely welcome the change but we may want to fix other arches as
well:
git grep hardware_unsetup HEAD
HEAD:arch/arm64/include/asm/kvm_host.h:static inline void kvm_arch_hardware_unsetup(void) {}
HEAD:arch/mips/include/asm/kvm_host.h:static inline void kvm_arch_hardware_unsetup(void) {}
HEAD:arch/powerpc/include/asm/kvm_host.h:static inline void kvm_arch_hardware_unsetup(void) {}
HEAD:arch/s390/kvm/kvm-s390.c:void kvm_arch_hardware_unsetup(void)
> bool (*cpu_has_accelerated_tpr)(void);
> bool (*has_emulated_msr)(u32 index);
> void (*cpuid_update)(struct kvm_vcpu *vcpu);
> @@ -1141,7 +1141,7 @@ struct kvm_x86_ops {
> */
> void (*tlb_flush_guest)(struct kvm_vcpu *vcpu);
>
> - enum exit_fastpath_completion (*run)(struct kvm_vcpu *vcpu);
> + enum exit_fastpath_completion (*vcpu_run)(struct kvm_vcpu *vcpu);
> int (*handle_exit)(struct kvm_vcpu *vcpu,
> enum exit_fastpath_completion exit_fastpath);
> int (*skip_emulated_instruction)(struct kvm_vcpu *vcpu);
> @@ -1150,8 +1150,8 @@ struct kvm_x86_ops {
> u32 (*get_interrupt_shadow)(struct kvm_vcpu *vcpu);
> void (*patch_hypercall)(struct kvm_vcpu *vcpu,
> unsigned char *hypercall_addr);
> - void (*set_irq)(struct kvm_vcpu *vcpu);
> - void (*set_nmi)(struct kvm_vcpu *vcpu);
> + void (*inject_irq)(struct kvm_vcpu *vcpu);
> + void (*inject_nmi)(struct kvm_vcpu *vcpu);
> void (*queue_exception)(struct kvm_vcpu *vcpu);
> void (*cancel_injection)(struct kvm_vcpu *vcpu);
> int (*interrupt_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
> @@ -1258,8 +1258,8 @@ struct kvm_x86_ops {
> void (*enable_smi_window)(struct kvm_vcpu *vcpu);
>
> int (*mem_enc_op)(struct kvm *kvm, void __user *argp);
> - int (*mem_enc_reg_region)(struct kvm *kvm, struct kvm_enc_region *argp);
> - int (*mem_enc_unreg_region)(struct kvm *kvm, struct kvm_enc_region *argp);
> + int (*mem_enc_register_region)(struct kvm *kvm, struct kvm_enc_region *argp);
> + int (*mem_enc_unregister_region)(struct kvm *kvm, struct kvm_enc_region *argp);
These two should probably pair with
KVM_MEMORY_ENCRYPT_REG_REGION
KVM_MEMORY_ENCRYPT_UNREG_REGION
shortening "memory" as "mem" and "encrypt" as "enc" is probably OK
because I can't think of any other meaning but shortening "register" as
"reg" may be percieved as CPU register.
What if we change ioctls too:
KVM_MEM_ENC_REGISTER_REGION
KVM_MEM_ENC_UNREGISTER_REGION
to make this all consistent?
>
> int (*get_msr_feature)(struct kvm_msr_entry *entry);
>
> diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
> index c0da4dd..24755eb 100644
> --- a/arch/x86/kvm/svm/svm.c
> +++ b/arch/x86/kvm/svm/svm.c
> @@ -3969,7 +3969,7 @@ static int svm_vm_init(struct kvm *kvm)
> }
>
> static struct kvm_x86_ops svm_x86_ops __initdata = {
> - .hardware_unsetup = svm_hardware_teardown,
> + .hardware_teardown = svm_hardware_teardown,
> .hardware_enable = svm_hardware_enable,
> .hardware_disable = svm_hardware_disable,
> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr,
> @@ -4016,15 +4016,15 @@ static int svm_vm_init(struct kvm *kvm)
> .tlb_flush_gva = svm_flush_tlb_gva,
> .tlb_flush_guest = svm_flush_tlb,
>
> - .run = svm_vcpu_run,
> + .vcpu_run = svm_vcpu_run,
> .handle_exit = handle_exit,
> .skip_emulated_instruction = skip_emulated_instruction,
> .update_emulated_instruction = NULL,
> .set_interrupt_shadow = svm_set_interrupt_shadow,
> .get_interrupt_shadow = svm_get_interrupt_shadow,
> .patch_hypercall = svm_patch_hypercall,
> - .set_irq = svm_set_irq,
> - .set_nmi = svm_inject_nmi,
> + .inject_irq = svm_set_irq,
> + .inject_nmi = svm_inject_nmi,
> .queue_exception = svm_queue_exception,
> .cancel_injection = svm_cancel_injection,
> .interrupt_allowed = svm_interrupt_allowed,
> @@ -4080,8 +4080,8 @@ static int svm_vm_init(struct kvm *kvm)
> .enable_smi_window = enable_smi_window,
>
> .mem_enc_op = svm_mem_enc_op,
> - .mem_enc_reg_region = svm_register_enc_region,
> - .mem_enc_unreg_region = svm_unregister_enc_region,
> + .mem_enc_register_region = svm_register_enc_region,
> + .mem_enc_unregister_region = svm_unregister_enc_region,
>
> .need_emulation_on_page_fault = svm_need_emulation_on_page_fault,
>
> diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
> index cb22f33..90d91524 100644
> --- a/arch/x86/kvm/vmx/vmx.c
> +++ b/arch/x86/kvm/vmx/vmx.c
> @@ -7844,7 +7844,7 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
> }
>
> static struct kvm_x86_ops vmx_x86_ops __initdata = {
> - .hardware_unsetup = hardware_unsetup,
> + .hardware_teardown = hardware_unsetup,
>
> .hardware_enable = hardware_enable,
> .hardware_disable = hardware_disable,
> @@ -7889,15 +7889,15 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
> .tlb_flush_gva = vmx_flush_tlb_gva,
> .tlb_flush_guest = vmx_flush_tlb_guest,
>
> - .run = vmx_vcpu_run,
> + .vcpu_run = vmx_vcpu_run,
> .handle_exit = vmx_handle_exit,
> .skip_emulated_instruction = vmx_skip_emulated_instruction,
> .update_emulated_instruction = vmx_update_emulated_instruction,
> .set_interrupt_shadow = vmx_set_interrupt_shadow,
> .get_interrupt_shadow = vmx_get_interrupt_shadow,
> .patch_hypercall = vmx_patch_hypercall,
> - .set_irq = vmx_inject_irq,
> - .set_nmi = vmx_inject_nmi,
> + .inject_irq = vmx_inject_irq,
> + .inject_nmi = vmx_inject_nmi,
> .queue_exception = vmx_queue_exception,
> .cancel_injection = vmx_cancel_injection,
> .interrupt_allowed = vmx_interrupt_allowed,
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index 3b92db4..e850fb3 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -5270,8 +5270,8 @@ long kvm_arch_vm_ioctl(struct file *filp,
> goto out;
>
> r = -ENOTTY;
> - if (kvm_x86_ops.mem_enc_reg_region)
> - r = kvm_x86_ops.mem_enc_reg_region(kvm, ®ion);
> + if (kvm_x86_ops.mem_enc_register_region)
> + r = kvm_x86_ops.mem_enc_register_region(kvm, ®ion);
> break;
> }
> case KVM_MEMORY_ENCRYPT_UNREG_REGION: {
> @@ -5282,8 +5282,8 @@ long kvm_arch_vm_ioctl(struct file *filp,
> goto out;
>
> r = -ENOTTY;
> - if (kvm_x86_ops.mem_enc_unreg_region)
> - r = kvm_x86_ops.mem_enc_unreg_region(kvm, ®ion);
> + if (kvm_x86_ops.mem_enc_unregister_region)
> + r = kvm_x86_ops.mem_enc_unregister_region(kvm, ®ion);
> break;
> }
> case KVM_HYPERV_EVENTFD: {
> @@ -7788,10 +7788,10 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
> */
> else if (!vcpu->arch.exception.pending) {
> if (vcpu->arch.nmi_injected) {
> - kvm_x86_ops.set_nmi(vcpu);
> + kvm_x86_ops.inject_nmi(vcpu);
> can_inject = false;
> } else if (vcpu->arch.interrupt.injected) {
> - kvm_x86_ops.set_irq(vcpu);
> + kvm_x86_ops.inject_irq(vcpu);
> can_inject = false;
> }
> }
> @@ -7867,7 +7867,7 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
> if (r) {
> --vcpu->arch.nmi_pending;
> vcpu->arch.nmi_injected = true;
> - kvm_x86_ops.set_nmi(vcpu);
> + kvm_x86_ops.inject_nmi(vcpu);
> can_inject = false;
> WARN_ON(kvm_x86_ops.nmi_allowed(vcpu, true) < 0);
> }
> @@ -7881,7 +7881,7 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
> goto busy;
> if (r) {
> kvm_queue_interrupt(vcpu, kvm_cpu_get_interrupt(vcpu), false);
> - kvm_x86_ops.set_irq(vcpu);
> + kvm_x86_ops.inject_irq(vcpu);
> WARN_ON(kvm_x86_ops.interrupt_allowed(vcpu, true) < 0);
> }
> if (kvm_cpu_has_injectable_intr(vcpu))
> @@ -8517,7 +8517,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
> vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_RELOAD;
> }
>
> - exit_fastpath = kvm_x86_ops.run(vcpu);
> + exit_fastpath = kvm_x86_ops.vcpu_run(vcpu);
>
> /*
> * Do this here before restoring debug registers on the host. And
> @@ -9793,9 +9793,9 @@ int kvm_arch_hardware_setup(void *opaque)
> return 0;
> }
>
> -void kvm_arch_hardware_unsetup(void)
> +void kvm_arch_hardware_teardown(void)
> {
> - kvm_x86_ops.hardware_unsetup();
> + kvm_x86_ops.hardware_teardown();
> }
>
> int kvm_arch_check_processor_compat(void *opaque)
> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
> index d564855..b49312c 100644
> --- a/include/linux/kvm_host.h
> +++ b/include/linux/kvm_host.h
> @@ -894,7 +894,7 @@ int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
> int kvm_arch_hardware_enable(void);
> void kvm_arch_hardware_disable(void);
> int kvm_arch_hardware_setup(void *opaque);
> -void kvm_arch_hardware_unsetup(void);
> +void kvm_arch_hardware_teardown(void);
> int kvm_arch_check_processor_compat(void *opaque);
> int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu);
> bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu);
> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
> index a852af5..4625f3a 100644
> --- a/virt/kvm/kvm_main.c
> +++ b/virt/kvm/kvm_main.c
> @@ -4786,7 +4786,7 @@ int kvm_init(void *opaque, unsigned vcpu_size, unsigned vcpu_align,
> unregister_reboot_notifier(&kvm_reboot_notifier);
> cpuhp_remove_state_nocalls(CPUHP_AP_KVM_STARTING);
> out_free_2:
> - kvm_arch_hardware_unsetup();
> + kvm_arch_hardware_teardown();
> out_free_1:
> free_cpumask_var(cpus_hardware_enabled);
> out_free_0:
> @@ -4808,7 +4808,7 @@ void kvm_exit(void)
> unregister_reboot_notifier(&kvm_reboot_notifier);
> cpuhp_remove_state_nocalls(CPUHP_AP_KVM_STARTING);
> on_each_cpu(hardware_disable_nolock, NULL, 1);
> - kvm_arch_hardware_unsetup();
> + kvm_arch_hardware_teardown();
And this probably means we'll have to take care of non-x86 arches after
all.
> kvm_arch_exit();
> kvm_irqfd_exit();
> free_cpumask_var(cpus_hardware_enabled);
--
Vitaly
^ permalink raw reply [flat|nested] 10+ messages in thread
* Re: [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable
2020-07-27 9:39 ` Vitaly Kuznetsov
@ 2020-07-27 17:30 ` Krish Sadhukhan
0 siblings, 0 replies; 10+ messages in thread
From: Krish Sadhukhan @ 2020-07-27 17:30 UTC (permalink / raw)
To: Vitaly Kuznetsov, kvm; +Cc: jmattson, sean.j.christopherson, pbonzini
On 7/27/20 2:39 AM, Vitaly Kuznetsov wrote:
> Krish Sadhukhan <krish.sadhukhan@oracle.com> writes:
>
>> Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
>> Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
>> Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
>> Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
>> ---
>> arch/x86/include/asm/kvm_host.h | 12 ++++++------
>> arch/x86/kvm/svm/svm.c | 12 ++++++------
>> arch/x86/kvm/vmx/vmx.c | 8 ++++----
>> arch/x86/kvm/x86.c | 22 +++++++++++-----------
>> include/linux/kvm_host.h | 2 +-
>> virt/kvm/kvm_main.c | 4 ++--
>> 6 files changed, 30 insertions(+), 30 deletions(-)
>>
>> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
>> index be5363b..ccad66d 100644
>> --- a/arch/x86/include/asm/kvm_host.h
>> +++ b/arch/x86/include/asm/kvm_host.h
>> @@ -1080,7 +1080,7 @@ static inline u16 kvm_lapic_irq_dest_mode(bool dest_mode_logical)
>> struct kvm_x86_ops {
>> int (*hardware_enable)(void);
>> void (*hardware_disable)(void);
>> - void (*hardware_unsetup)(void);
>> + void (*hardware_teardown)(void);
> I definitely welcome the change but we may want to fix other arches as
> well:
>
> git grep hardware_unsetup HEAD
>
> HEAD:arch/arm64/include/asm/kvm_host.h:static inline void kvm_arch_hardware_unsetup(void) {}
> HEAD:arch/mips/include/asm/kvm_host.h:static inline void kvm_arch_hardware_unsetup(void) {}
> HEAD:arch/powerpc/include/asm/kvm_host.h:static inline void kvm_arch_hardware_unsetup(void) {}
> HEAD:arch/s390/kvm/kvm-s390.c:void kvm_arch_hardware_unsetup(void)
Absolutely. I should have thought about other arches when making that
change !
>
>> bool (*cpu_has_accelerated_tpr)(void);
>> bool (*has_emulated_msr)(u32 index);
>> void (*cpuid_update)(struct kvm_vcpu *vcpu);
>> @@ -1141,7 +1141,7 @@ struct kvm_x86_ops {
>> */
>> void (*tlb_flush_guest)(struct kvm_vcpu *vcpu);
>>
>> - enum exit_fastpath_completion (*run)(struct kvm_vcpu *vcpu);
>> + enum exit_fastpath_completion (*vcpu_run)(struct kvm_vcpu *vcpu);
>> int (*handle_exit)(struct kvm_vcpu *vcpu,
>> enum exit_fastpath_completion exit_fastpath);
>> int (*skip_emulated_instruction)(struct kvm_vcpu *vcpu);
>> @@ -1150,8 +1150,8 @@ struct kvm_x86_ops {
>> u32 (*get_interrupt_shadow)(struct kvm_vcpu *vcpu);
>> void (*patch_hypercall)(struct kvm_vcpu *vcpu,
>> unsigned char *hypercall_addr);
>> - void (*set_irq)(struct kvm_vcpu *vcpu);
>> - void (*set_nmi)(struct kvm_vcpu *vcpu);
>> + void (*inject_irq)(struct kvm_vcpu *vcpu);
>> + void (*inject_nmi)(struct kvm_vcpu *vcpu);
>> void (*queue_exception)(struct kvm_vcpu *vcpu);
>> void (*cancel_injection)(struct kvm_vcpu *vcpu);
>> int (*interrupt_allowed)(struct kvm_vcpu *vcpu, bool for_injection);
>> @@ -1258,8 +1258,8 @@ struct kvm_x86_ops {
>> void (*enable_smi_window)(struct kvm_vcpu *vcpu);
>>
>> int (*mem_enc_op)(struct kvm *kvm, void __user *argp);
>> - int (*mem_enc_reg_region)(struct kvm *kvm, struct kvm_enc_region *argp);
>> - int (*mem_enc_unreg_region)(struct kvm *kvm, struct kvm_enc_region *argp);
>> + int (*mem_enc_register_region)(struct kvm *kvm, struct kvm_enc_region *argp);
>> + int (*mem_enc_unregister_region)(struct kvm *kvm, struct kvm_enc_region *argp);
> These two should probably pair with
>
> KVM_MEMORY_ENCRYPT_REG_REGION
> KVM_MEMORY_ENCRYPT_UNREG_REGION
>
> shortening "memory" as "mem" and "encrypt" as "enc" is probably OK
> because I can't think of any other meaning but shortening "register" as
> "reg" may be percieved as CPU register.
>
> What if we change ioctls too:
> KVM_MEM_ENC_REGISTER_REGION
> KVM_MEM_ENC_UNREGISTER_REGION
>
> to make this all consistent?
Works for me. Will send v3. Thanks !
>
>>
>> int (*get_msr_feature)(struct kvm_msr_entry *entry);
>>
>> diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
>> index c0da4dd..24755eb 100644
>> --- a/arch/x86/kvm/svm/svm.c
>> +++ b/arch/x86/kvm/svm/svm.c
>> @@ -3969,7 +3969,7 @@ static int svm_vm_init(struct kvm *kvm)
>> }
>>
>> static struct kvm_x86_ops svm_x86_ops __initdata = {
>> - .hardware_unsetup = svm_hardware_teardown,
>> + .hardware_teardown = svm_hardware_teardown,
>> .hardware_enable = svm_hardware_enable,
>> .hardware_disable = svm_hardware_disable,
>> .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr,
>> @@ -4016,15 +4016,15 @@ static int svm_vm_init(struct kvm *kvm)
>> .tlb_flush_gva = svm_flush_tlb_gva,
>> .tlb_flush_guest = svm_flush_tlb,
>>
>> - .run = svm_vcpu_run,
>> + .vcpu_run = svm_vcpu_run,
>> .handle_exit = handle_exit,
>> .skip_emulated_instruction = skip_emulated_instruction,
>> .update_emulated_instruction = NULL,
>> .set_interrupt_shadow = svm_set_interrupt_shadow,
>> .get_interrupt_shadow = svm_get_interrupt_shadow,
>> .patch_hypercall = svm_patch_hypercall,
>> - .set_irq = svm_set_irq,
>> - .set_nmi = svm_inject_nmi,
>> + .inject_irq = svm_set_irq,
>> + .inject_nmi = svm_inject_nmi,
>> .queue_exception = svm_queue_exception,
>> .cancel_injection = svm_cancel_injection,
>> .interrupt_allowed = svm_interrupt_allowed,
>> @@ -4080,8 +4080,8 @@ static int svm_vm_init(struct kvm *kvm)
>> .enable_smi_window = enable_smi_window,
>>
>> .mem_enc_op = svm_mem_enc_op,
>> - .mem_enc_reg_region = svm_register_enc_region,
>> - .mem_enc_unreg_region = svm_unregister_enc_region,
>> + .mem_enc_register_region = svm_register_enc_region,
>> + .mem_enc_unregister_region = svm_unregister_enc_region,
>>
>> .need_emulation_on_page_fault = svm_need_emulation_on_page_fault,
>>
>> diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
>> index cb22f33..90d91524 100644
>> --- a/arch/x86/kvm/vmx/vmx.c
>> +++ b/arch/x86/kvm/vmx/vmx.c
>> @@ -7844,7 +7844,7 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
>> }
>>
>> static struct kvm_x86_ops vmx_x86_ops __initdata = {
>> - .hardware_unsetup = hardware_unsetup,
>> + .hardware_teardown = hardware_unsetup,
>>
>> .hardware_enable = hardware_enable,
>> .hardware_disable = hardware_disable,
>> @@ -7889,15 +7889,15 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
>> .tlb_flush_gva = vmx_flush_tlb_gva,
>> .tlb_flush_guest = vmx_flush_tlb_guest,
>>
>> - .run = vmx_vcpu_run,
>> + .vcpu_run = vmx_vcpu_run,
>> .handle_exit = vmx_handle_exit,
>> .skip_emulated_instruction = vmx_skip_emulated_instruction,
>> .update_emulated_instruction = vmx_update_emulated_instruction,
>> .set_interrupt_shadow = vmx_set_interrupt_shadow,
>> .get_interrupt_shadow = vmx_get_interrupt_shadow,
>> .patch_hypercall = vmx_patch_hypercall,
>> - .set_irq = vmx_inject_irq,
>> - .set_nmi = vmx_inject_nmi,
>> + .inject_irq = vmx_inject_irq,
>> + .inject_nmi = vmx_inject_nmi,
>> .queue_exception = vmx_queue_exception,
>> .cancel_injection = vmx_cancel_injection,
>> .interrupt_allowed = vmx_interrupt_allowed,
>> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
>> index 3b92db4..e850fb3 100644
>> --- a/arch/x86/kvm/x86.c
>> +++ b/arch/x86/kvm/x86.c
>> @@ -5270,8 +5270,8 @@ long kvm_arch_vm_ioctl(struct file *filp,
>> goto out;
>>
>> r = -ENOTTY;
>> - if (kvm_x86_ops.mem_enc_reg_region)
>> - r = kvm_x86_ops.mem_enc_reg_region(kvm, ®ion);
>> + if (kvm_x86_ops.mem_enc_register_region)
>> + r = kvm_x86_ops.mem_enc_register_region(kvm, ®ion);
>> break;
>> }
>> case KVM_MEMORY_ENCRYPT_UNREG_REGION: {
>> @@ -5282,8 +5282,8 @@ long kvm_arch_vm_ioctl(struct file *filp,
>> goto out;
>>
>> r = -ENOTTY;
>> - if (kvm_x86_ops.mem_enc_unreg_region)
>> - r = kvm_x86_ops.mem_enc_unreg_region(kvm, ®ion);
>> + if (kvm_x86_ops.mem_enc_unregister_region)
>> + r = kvm_x86_ops.mem_enc_unregister_region(kvm, ®ion);
>> break;
>> }
>> case KVM_HYPERV_EVENTFD: {
>> @@ -7788,10 +7788,10 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
>> */
>> else if (!vcpu->arch.exception.pending) {
>> if (vcpu->arch.nmi_injected) {
>> - kvm_x86_ops.set_nmi(vcpu);
>> + kvm_x86_ops.inject_nmi(vcpu);
>> can_inject = false;
>> } else if (vcpu->arch.interrupt.injected) {
>> - kvm_x86_ops.set_irq(vcpu);
>> + kvm_x86_ops.inject_irq(vcpu);
>> can_inject = false;
>> }
>> }
>> @@ -7867,7 +7867,7 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
>> if (r) {
>> --vcpu->arch.nmi_pending;
>> vcpu->arch.nmi_injected = true;
>> - kvm_x86_ops.set_nmi(vcpu);
>> + kvm_x86_ops.inject_nmi(vcpu);
>> can_inject = false;
>> WARN_ON(kvm_x86_ops.nmi_allowed(vcpu, true) < 0);
>> }
>> @@ -7881,7 +7881,7 @@ static void inject_pending_event(struct kvm_vcpu *vcpu, bool *req_immediate_exit
>> goto busy;
>> if (r) {
>> kvm_queue_interrupt(vcpu, kvm_cpu_get_interrupt(vcpu), false);
>> - kvm_x86_ops.set_irq(vcpu);
>> + kvm_x86_ops.inject_irq(vcpu);
>> WARN_ON(kvm_x86_ops.interrupt_allowed(vcpu, true) < 0);
>> }
>> if (kvm_cpu_has_injectable_intr(vcpu))
>> @@ -8517,7 +8517,7 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)
>> vcpu->arch.switch_db_regs &= ~KVM_DEBUGREG_RELOAD;
>> }
>>
>> - exit_fastpath = kvm_x86_ops.run(vcpu);
>> + exit_fastpath = kvm_x86_ops.vcpu_run(vcpu);
>>
>> /*
>> * Do this here before restoring debug registers on the host. And
>> @@ -9793,9 +9793,9 @@ int kvm_arch_hardware_setup(void *opaque)
>> return 0;
>> }
>>
>> -void kvm_arch_hardware_unsetup(void)
>> +void kvm_arch_hardware_teardown(void)
>> {
>> - kvm_x86_ops.hardware_unsetup();
>> + kvm_x86_ops.hardware_teardown();
>> }
>>
>> int kvm_arch_check_processor_compat(void *opaque)
>> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
>> index d564855..b49312c 100644
>> --- a/include/linux/kvm_host.h
>> +++ b/include/linux/kvm_host.h
>> @@ -894,7 +894,7 @@ int kvm_arch_vcpu_ioctl_set_guest_debug(struct kvm_vcpu *vcpu,
>> int kvm_arch_hardware_enable(void);
>> void kvm_arch_hardware_disable(void);
>> int kvm_arch_hardware_setup(void *opaque);
>> -void kvm_arch_hardware_unsetup(void);
>> +void kvm_arch_hardware_teardown(void);
>> int kvm_arch_check_processor_compat(void *opaque);
>> int kvm_arch_vcpu_runnable(struct kvm_vcpu *vcpu);
>> bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu);
>> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
>> index a852af5..4625f3a 100644
>> --- a/virt/kvm/kvm_main.c
>> +++ b/virt/kvm/kvm_main.c
>> @@ -4786,7 +4786,7 @@ int kvm_init(void *opaque, unsigned vcpu_size, unsigned vcpu_align,
>> unregister_reboot_notifier(&kvm_reboot_notifier);
>> cpuhp_remove_state_nocalls(CPUHP_AP_KVM_STARTING);
>> out_free_2:
>> - kvm_arch_hardware_unsetup();
>> + kvm_arch_hardware_teardown();
>> out_free_1:
>> free_cpumask_var(cpus_hardware_enabled);
>> out_free_0:
>> @@ -4808,7 +4808,7 @@ void kvm_exit(void)
>> unregister_reboot_notifier(&kvm_reboot_notifier);
>> cpuhp_remove_state_nocalls(CPUHP_AP_KVM_STARTING);
>> on_each_cpu(hardware_disable_nolock, NULL, 1);
>> - kvm_arch_hardware_unsetup();
>> + kvm_arch_hardware_teardown();
> And this probably means we'll have to take care of non-x86 arches after
> all.
>
>> kvm_arch_exit();
>> kvm_irqfd_exit();
>> free_cpumask_var(cpus_hardware_enabled);
^ permalink raw reply [flat|nested] 10+ messages in thread
* [PATCH 2/5 v2] KVM: SVM: Fill in conforming svm_x86_ops via macro
2020-07-25 3:26 [PATCH 0/5 v2] KVM: x86: Fill in conforming {vmx|svm}_x86_ops and {vmx|svm}_nested_ops via macros Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable Krish Sadhukhan
@ 2020-07-25 3:26 ` Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 3/5 v2] KVM: nSVM: Fill in conforming svm_nested_ops " Krish Sadhukhan
` (2 subsequent siblings)
4 siblings, 0 replies; 10+ messages in thread
From: Krish Sadhukhan @ 2020-07-25 3:26 UTC (permalink / raw)
To: kvm; +Cc: jmattson, sean.j.christopherson, pbonzini, vkuznets
The names of some of the svm_x86_ops functions do not have a corresponding
'svm_' prefix. Generate the names using a macro so that the names are
conformant. Fixing the naming will help in better readability and
maintenance of the code.
Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
---
arch/x86/kvm/svm/avic.c | 4 +-
arch/x86/kvm/svm/nested.c | 2 +-
arch/x86/kvm/svm/sev.c | 6 +-
arch/x86/kvm/svm/svm.c | 218 +++++++++++++++++++++++-----------------------
arch/x86/kvm/svm/svm.h | 8 +-
5 files changed, 120 insertions(+), 118 deletions(-)
diff --git a/arch/x86/kvm/svm/avic.c b/arch/x86/kvm/svm/avic.c
index e80daa9..619391e 100644
--- a/arch/x86/kvm/svm/avic.c
+++ b/arch/x86/kvm/svm/avic.c
@@ -579,7 +579,7 @@ int avic_init_vcpu(struct vcpu_svm *svm)
return ret;
}
-void avic_post_state_restore(struct kvm_vcpu *vcpu)
+void svm_avic_post_state_restore(struct kvm_vcpu *vcpu)
{
if (avic_handle_apic_id_update(vcpu) != 0)
return;
@@ -660,7 +660,7 @@ void svm_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu)
* we need to check and update the AVIC logical APIC ID table
* accordingly before re-activating.
*/
- avic_post_state_restore(vcpu);
+ svm_avic_post_state_restore(vcpu);
vmcb->control.int_ctl |= AVIC_ENABLE_MASK;
} else {
vmcb->control.int_ctl &= ~AVIC_ENABLE_MASK;
diff --git a/arch/x86/kvm/svm/nested.c b/arch/x86/kvm/svm/nested.c
index 6bceafb..3be6256 100644
--- a/arch/x86/kvm/svm/nested.c
+++ b/arch/x86/kvm/svm/nested.c
@@ -348,7 +348,7 @@ static void nested_prepare_vmcb_control(struct vcpu_svm *svm)
/* Guest paging mode is active - reset mmu */
kvm_mmu_reset_context(&svm->vcpu);
- svm_flush_tlb(&svm->vcpu);
+ svm_tlb_flush(&svm->vcpu);
svm->vmcb->control.tsc_offset = svm->vcpu.arch.tsc_offset =
svm->vcpu.arch.l1_tsc_offset + svm->nested.ctl.tsc_offset;
diff --git a/arch/x86/kvm/svm/sev.c b/arch/x86/kvm/svm/sev.c
index 5573a97..1ca9f60 100644
--- a/arch/x86/kvm/svm/sev.c
+++ b/arch/x86/kvm/svm/sev.c
@@ -969,7 +969,7 @@ int svm_mem_enc_op(struct kvm *kvm, void __user *argp)
return r;
}
-int svm_register_enc_region(struct kvm *kvm,
+int svm_mem_enc_register_region(struct kvm *kvm,
struct kvm_enc_region *range)
{
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
@@ -1038,8 +1038,8 @@ static void __unregister_enc_region_locked(struct kvm *kvm,
kfree(region);
}
-int svm_unregister_enc_region(struct kvm *kvm,
- struct kvm_enc_region *range)
+int svm_mem_enc_unregister_region(struct kvm *kvm,
+ struct kvm_enc_region *range)
{
struct enc_region *region;
int ret;
diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
index 24755eb..d63181e 100644
--- a/arch/x86/kvm/svm/svm.c
+++ b/arch/x86/kvm/svm/svm.c
@@ -254,7 +254,7 @@ static inline void invlpga(unsigned long addr, u32 asid)
asm volatile (__ex("invlpga %1, %0") : : "c"(asid), "a"(addr));
}
-static int get_npt_level(struct kvm_vcpu *vcpu)
+static int svm_get_tdp_level(struct kvm_vcpu *vcpu)
{
#ifdef CONFIG_X86_64
return PT64_ROOT_4LEVEL;
@@ -312,7 +312,7 @@ static void svm_set_interrupt_shadow(struct kvm_vcpu *vcpu, int mask)
}
-static int skip_emulated_instruction(struct kvm_vcpu *vcpu)
+static int svm_skip_emulated_instruction(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -351,7 +351,7 @@ static void svm_queue_exception(struct kvm_vcpu *vcpu)
* raises a fault that is not intercepted. Still better than
* failing in all cases.
*/
- (void)skip_emulated_instruction(&svm->vcpu);
+ (void)svm_skip_emulated_instruction(&svm->vcpu);
rip = kvm_rip_read(&svm->vcpu);
svm->int3_rip = rip + svm->vmcb->save.cs.base;
svm->int3_injected = rip - old_rip;
@@ -1153,7 +1153,7 @@ static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
avic_update_vapic_bar(svm, APIC_DEFAULT_PHYS_BASE);
}
-static int svm_create_vcpu(struct kvm_vcpu *vcpu)
+static int svm_vcpu_create(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm;
struct page *page;
@@ -1232,7 +1232,7 @@ static void svm_clear_current_vmcb(struct vmcb *vmcb)
cmpxchg(&per_cpu(svm_data, i)->current_vmcb, vmcb, NULL);
}
-static void svm_free_vcpu(struct kvm_vcpu *vcpu)
+static void svm_vcpu_free(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -1585,7 +1585,7 @@ int svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4)
return 1;
if (npt_enabled && ((old_cr4 ^ cr4) & X86_CR4_PGE))
- svm_flush_tlb(vcpu);
+ svm_tlb_flush(vcpu);
vcpu->arch.cr4 = cr4;
if (!npt_enabled)
@@ -1627,7 +1627,7 @@ static void svm_set_segment(struct kvm_vcpu *vcpu,
mark_dirty(svm->vmcb, VMCB_SEG);
}
-static void update_bp_intercept(struct kvm_vcpu *vcpu)
+static void svm_update_bp_intercept(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -2143,7 +2143,7 @@ static int task_switch_interception(struct vcpu_svm *svm)
int_type == SVM_EXITINTINFO_TYPE_SOFT ||
(int_type == SVM_EXITINTINFO_TYPE_EXEPT &&
(int_vec == OF_VECTOR || int_vec == BP_VECTOR))) {
- if (!skip_emulated_instruction(&svm->vcpu))
+ if (!svm_skip_emulated_instruction(&svm->vcpu))
return 0;
}
@@ -2909,7 +2909,7 @@ static void svm_get_exit_info(struct kvm_vcpu *vcpu, u64 *info1, u64 *info2)
*info2 = control->exit_info_2;
}
-static int handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath)
+static int svm_handle_exit(struct kvm_vcpu *vcpu, fastpath_t exit_fastpath)
{
struct vcpu_svm *svm = to_svm(vcpu);
struct kvm_run *kvm_run = vcpu->run;
@@ -3023,7 +3023,7 @@ static void svm_inject_nmi(struct kvm_vcpu *vcpu)
++vcpu->stat.nmi_injections;
}
-static void svm_set_irq(struct kvm_vcpu *vcpu)
+static void svm_inject_irq(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3036,7 +3036,7 @@ static void svm_set_irq(struct kvm_vcpu *vcpu)
SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_INTR;
}
-static void update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
+static void svm_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3145,7 +3145,7 @@ static int svm_interrupt_allowed(struct kvm_vcpu *vcpu, bool for_injection)
return !svm_interrupt_blocked(vcpu);
}
-static void enable_irq_window(struct kvm_vcpu *vcpu)
+static void svm_enable_irq_window(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3169,7 +3169,7 @@ static void enable_irq_window(struct kvm_vcpu *vcpu)
}
}
-static void enable_nmi_window(struct kvm_vcpu *vcpu)
+static void svm_enable_nmi_window(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3202,7 +3202,7 @@ static int svm_set_identity_map_addr(struct kvm *kvm, u64 ident_addr)
return 0;
}
-void svm_flush_tlb(struct kvm_vcpu *vcpu)
+void svm_tlb_flush(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3219,7 +3219,7 @@ void svm_flush_tlb(struct kvm_vcpu *vcpu)
svm->asid_generation--;
}
-static void svm_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t gva)
+static void svm_tlb_flush_gva(struct kvm_vcpu *vcpu, gva_t gva)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3857,7 +3857,7 @@ static int svm_pre_leave_smm(struct kvm_vcpu *vcpu, const char *smstate)
return 0;
}
-static void enable_smi_window(struct kvm_vcpu *vcpu)
+static void svm_enable_smi_window(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
@@ -3968,124 +3968,126 @@ static int svm_vm_init(struct kvm *kvm)
return 0;
}
+#define KVM_X86_OP(name) .name = svm_##name
+
static struct kvm_x86_ops svm_x86_ops __initdata = {
- .hardware_teardown = svm_hardware_teardown,
- .hardware_enable = svm_hardware_enable,
- .hardware_disable = svm_hardware_disable,
- .cpu_has_accelerated_tpr = svm_cpu_has_accelerated_tpr,
- .has_emulated_msr = svm_has_emulated_msr,
+ KVM_X86_OP(hardware_teardown),
+ KVM_X86_OP(hardware_enable),
+ KVM_X86_OP(hardware_disable),
+ KVM_X86_OP(cpu_has_accelerated_tpr),
+ KVM_X86_OP(has_emulated_msr),
- .vcpu_create = svm_create_vcpu,
- .vcpu_free = svm_free_vcpu,
- .vcpu_reset = svm_vcpu_reset,
+ KVM_X86_OP(vcpu_create),
+ KVM_X86_OP(vcpu_free),
+ KVM_X86_OP(vcpu_reset),
.vm_size = sizeof(struct kvm_svm),
- .vm_init = svm_vm_init,
- .vm_destroy = svm_vm_destroy,
-
- .prepare_guest_switch = svm_prepare_guest_switch,
- .vcpu_load = svm_vcpu_load,
- .vcpu_put = svm_vcpu_put,
- .vcpu_blocking = svm_vcpu_blocking,
- .vcpu_unblocking = svm_vcpu_unblocking,
-
- .update_bp_intercept = update_bp_intercept,
- .get_msr_feature = svm_get_msr_feature,
- .get_msr = svm_get_msr,
- .set_msr = svm_set_msr,
- .get_segment_base = svm_get_segment_base,
- .get_segment = svm_get_segment,
- .set_segment = svm_set_segment,
- .get_cpl = svm_get_cpl,
+ KVM_X86_OP(vm_init),
+ KVM_X86_OP(vm_destroy),
+
+ KVM_X86_OP(prepare_guest_switch),
+ KVM_X86_OP(vcpu_load),
+ KVM_X86_OP(vcpu_put),
+ KVM_X86_OP(vcpu_blocking),
+ KVM_X86_OP(vcpu_unblocking),
+
+ KVM_X86_OP(update_bp_intercept),
+ KVM_X86_OP(get_msr_feature),
+ KVM_X86_OP(get_msr),
+ KVM_X86_OP(set_msr),
+ KVM_X86_OP(get_segment_base),
+ KVM_X86_OP(get_segment),
+ KVM_X86_OP(set_segment),
+ KVM_X86_OP(get_cpl),
.get_cs_db_l_bits = kvm_get_cs_db_l_bits,
- .set_cr0 = svm_set_cr0,
- .set_cr4 = svm_set_cr4,
- .set_efer = svm_set_efer,
- .get_idt = svm_get_idt,
- .set_idt = svm_set_idt,
- .get_gdt = svm_get_gdt,
- .set_gdt = svm_set_gdt,
- .set_dr7 = svm_set_dr7,
- .sync_dirty_debug_regs = svm_sync_dirty_debug_regs,
- .cache_reg = svm_cache_reg,
- .get_rflags = svm_get_rflags,
- .set_rflags = svm_set_rflags,
-
- .tlb_flush_all = svm_flush_tlb,
- .tlb_flush_current = svm_flush_tlb,
- .tlb_flush_gva = svm_flush_tlb_gva,
- .tlb_flush_guest = svm_flush_tlb,
-
- .vcpu_run = svm_vcpu_run,
- .handle_exit = handle_exit,
- .skip_emulated_instruction = skip_emulated_instruction,
+ KVM_X86_OP(set_cr0),
+ KVM_X86_OP(set_cr4),
+ KVM_X86_OP(set_efer),
+ KVM_X86_OP(get_idt),
+ KVM_X86_OP(set_idt),
+ KVM_X86_OP(get_gdt),
+ KVM_X86_OP(set_gdt),
+ KVM_X86_OP(set_dr7),
+ KVM_X86_OP(sync_dirty_debug_regs),
+ KVM_X86_OP(cache_reg),
+ KVM_X86_OP(get_rflags),
+ KVM_X86_OP(set_rflags),
+
+ .tlb_flush_all = svm_tlb_flush,
+ .tlb_flush_current = svm_tlb_flush,
+ KVM_X86_OP(tlb_flush_gva),
+ .tlb_flush_guest = svm_tlb_flush,
+
+ KVM_X86_OP(vcpu_run),
+ KVM_X86_OP(handle_exit),
+ KVM_X86_OP(skip_emulated_instruction),
.update_emulated_instruction = NULL,
- .set_interrupt_shadow = svm_set_interrupt_shadow,
- .get_interrupt_shadow = svm_get_interrupt_shadow,
- .patch_hypercall = svm_patch_hypercall,
- .inject_irq = svm_set_irq,
- .inject_nmi = svm_inject_nmi,
- .queue_exception = svm_queue_exception,
- .cancel_injection = svm_cancel_injection,
- .interrupt_allowed = svm_interrupt_allowed,
- .nmi_allowed = svm_nmi_allowed,
- .get_nmi_mask = svm_get_nmi_mask,
- .set_nmi_mask = svm_set_nmi_mask,
- .enable_nmi_window = enable_nmi_window,
- .enable_irq_window = enable_irq_window,
- .update_cr8_intercept = update_cr8_intercept,
- .set_virtual_apic_mode = svm_set_virtual_apic_mode,
- .refresh_apicv_exec_ctrl = svm_refresh_apicv_exec_ctrl,
- .check_apicv_inhibit_reasons = svm_check_apicv_inhibit_reasons,
- .pre_update_apicv_exec_ctrl = svm_pre_update_apicv_exec_ctrl,
- .load_eoi_exitmap = svm_load_eoi_exitmap,
- .hwapic_irr_update = svm_hwapic_irr_update,
- .hwapic_isr_update = svm_hwapic_isr_update,
+ KVM_X86_OP(set_interrupt_shadow),
+ KVM_X86_OP(get_interrupt_shadow),
+ KVM_X86_OP(patch_hypercall),
+ KVM_X86_OP(inject_irq),
+ KVM_X86_OP(inject_nmi),
+ KVM_X86_OP(queue_exception),
+ KVM_X86_OP(cancel_injection),
+ KVM_X86_OP(interrupt_allowed),
+ KVM_X86_OP(nmi_allowed),
+ KVM_X86_OP(get_nmi_mask),
+ KVM_X86_OP(set_nmi_mask),
+ KVM_X86_OP(enable_nmi_window),
+ KVM_X86_OP(enable_irq_window),
+ KVM_X86_OP(update_cr8_intercept),
+ KVM_X86_OP(set_virtual_apic_mode),
+ KVM_X86_OP(refresh_apicv_exec_ctrl),
+ KVM_X86_OP(check_apicv_inhibit_reasons),
+ KVM_X86_OP(pre_update_apicv_exec_ctrl),
+ KVM_X86_OP(load_eoi_exitmap),
+ KVM_X86_OP(hwapic_irr_update),
+ KVM_X86_OP(hwapic_isr_update),
.sync_pir_to_irr = kvm_lapic_find_highest_irr,
- .apicv_post_state_restore = avic_post_state_restore,
+ .apicv_post_state_restore = svm_avic_post_state_restore,
- .set_tss_addr = svm_set_tss_addr,
- .set_identity_map_addr = svm_set_identity_map_addr,
- .get_tdp_level = get_npt_level,
- .get_mt_mask = svm_get_mt_mask,
+ KVM_X86_OP(set_tss_addr),
+ KVM_X86_OP(set_identity_map_addr),
+ KVM_X86_OP(get_tdp_level),
+ KVM_X86_OP(get_mt_mask),
- .get_exit_info = svm_get_exit_info,
+ KVM_X86_OP(get_exit_info),
- .cpuid_update = svm_cpuid_update,
+ KVM_X86_OP(cpuid_update),
- .has_wbinvd_exit = svm_has_wbinvd_exit,
+ KVM_X86_OP(has_wbinvd_exit),
- .write_l1_tsc_offset = svm_write_l1_tsc_offset,
+ KVM_X86_OP(write_l1_tsc_offset),
- .load_mmu_pgd = svm_load_mmu_pgd,
+ KVM_X86_OP(load_mmu_pgd),
- .check_intercept = svm_check_intercept,
- .handle_exit_irqoff = svm_handle_exit_irqoff,
+ KVM_X86_OP(check_intercept),
+ KVM_X86_OP(handle_exit_irqoff),
.request_immediate_exit = __kvm_request_immediate_exit,
- .sched_in = svm_sched_in,
+ KVM_X86_OP(sched_in),
.pmu_ops = &amd_pmu_ops,
.nested_ops = &svm_nested_ops,
.deliver_posted_interrupt = svm_deliver_avic_intr,
- .dy_apicv_has_pending_interrupt = svm_dy_apicv_has_pending_interrupt,
- .update_pi_irte = svm_update_pi_irte,
- .setup_mce = svm_setup_mce,
+ KVM_X86_OP(dy_apicv_has_pending_interrupt),
+ KVM_X86_OP(update_pi_irte),
+ KVM_X86_OP(setup_mce),
- .smi_allowed = svm_smi_allowed,
- .pre_enter_smm = svm_pre_enter_smm,
- .pre_leave_smm = svm_pre_leave_smm,
- .enable_smi_window = enable_smi_window,
+ KVM_X86_OP(smi_allowed),
+ KVM_X86_OP(pre_enter_smm),
+ KVM_X86_OP(pre_leave_smm),
+ KVM_X86_OP(enable_smi_window),
- .mem_enc_op = svm_mem_enc_op,
- .mem_enc_register_region = svm_register_enc_region,
- .mem_enc_unregister_region = svm_unregister_enc_region,
+ KVM_X86_OP(mem_enc_op),
+ KVM_X86_OP(mem_enc_register_region),
+ KVM_X86_OP(mem_enc_unregister_region),
- .need_emulation_on_page_fault = svm_need_emulation_on_page_fault,
+ KVM_X86_OP(need_emulation_on_page_fault),
- .apic_init_signal_blocked = svm_apic_init_signal_blocked,
+ KVM_X86_OP(apic_init_signal_blocked),
};
static struct kvm_x86_init_ops svm_init_ops __initdata = {
diff --git a/arch/x86/kvm/svm/svm.h b/arch/x86/kvm/svm/svm.h
index 6ac4c00..e2d5029 100644
--- a/arch/x86/kvm/svm/svm.h
+++ b/arch/x86/kvm/svm/svm.h
@@ -352,7 +352,7 @@ static inline bool gif_set(struct vcpu_svm *svm)
void svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
void svm_set_cr0(struct kvm_vcpu *vcpu, unsigned long cr0);
int svm_set_cr4(struct kvm_vcpu *vcpu, unsigned long cr4);
-void svm_flush_tlb(struct kvm_vcpu *vcpu);
+void svm_tlb_flush(struct kvm_vcpu *vcpu);
void disable_nmi_singlestep(struct vcpu_svm *svm);
bool svm_smi_blocked(struct kvm_vcpu *vcpu);
bool svm_nmi_blocked(struct kvm_vcpu *vcpu);
@@ -444,7 +444,7 @@ static inline bool avic_vcpu_is_running(struct kvm_vcpu *vcpu)
int avic_init_vcpu(struct vcpu_svm *svm);
void avic_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
void avic_vcpu_put(struct kvm_vcpu *vcpu);
-void avic_post_state_restore(struct kvm_vcpu *vcpu);
+void svm_avic_post_state_restore(struct kvm_vcpu *vcpu);
void svm_set_virtual_apic_mode(struct kvm_vcpu *vcpu);
void svm_refresh_apicv_exec_ctrl(struct kvm_vcpu *vcpu);
bool svm_check_apicv_inhibit_reasons(ulong bit);
@@ -481,9 +481,9 @@ static inline bool svm_sev_enabled(void)
void sev_vm_destroy(struct kvm *kvm);
int svm_mem_enc_op(struct kvm *kvm, void __user *argp);
-int svm_register_enc_region(struct kvm *kvm,
+int svm_mem_enc_register_region(struct kvm *kvm,
struct kvm_enc_region *range);
-int svm_unregister_enc_region(struct kvm *kvm,
+int svm_mem_enc_unregister_region(struct kvm *kvm,
struct kvm_enc_region *range);
void pre_sev_run(struct vcpu_svm *svm, int cpu);
int __init sev_hardware_setup(void);
--
1.8.3.1
^ permalink raw reply related [flat|nested] 10+ messages in thread
* [PATCH 3/5 v2] KVM: nSVM: Fill in conforming svm_nested_ops via macro
2020-07-25 3:26 [PATCH 0/5 v2] KVM: x86: Fill in conforming {vmx|svm}_x86_ops and {vmx|svm}_nested_ops via macros Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 1/5 v2] KVM: x86: Change names of some of the kvm_x86_ops functions to make them more semantical and readable Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 2/5 v2] KVM: SVM: Fill in conforming svm_x86_ops via macro Krish Sadhukhan
@ 2020-07-25 3:26 ` Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 4/5 v2] KVM: VMX: Fill in conforming vmx_x86_ops " Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 5/5 v2] KVM: nVMX: Fill in conforming vmx_nested_ops " Krish Sadhukhan
4 siblings, 0 replies; 10+ messages in thread
From: Krish Sadhukhan @ 2020-07-25 3:26 UTC (permalink / raw)
To: kvm; +Cc: jmattson, sean.j.christopherson, pbonzini, vkuznets
The names of the nested_svm_ops functions do not have a corresponding
'nested_svm_' prefix. Generate the names using a macro so that the names are
conformant. Fixing the naming will help in better readability and
maintenance of the code.
Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
---
arch/x86/kvm/svm/nested.c | 16 +++++++++-------
1 file changed, 9 insertions(+), 7 deletions(-)
diff --git a/arch/x86/kvm/svm/nested.c b/arch/x86/kvm/svm/nested.c
index 3be6256..7cb834a 100644
--- a/arch/x86/kvm/svm/nested.c
+++ b/arch/x86/kvm/svm/nested.c
@@ -718,7 +718,7 @@ static int nested_svm_intercept(struct vcpu_svm *svm)
/*
* Host-intercepted exceptions have been checked already in
* nested_svm_exit_special. There is nothing to do here,
- * the vmexit is injected by svm_check_nested_events.
+ * the vmexit is injected by nested_svm_check_events().
*/
vmexit = NESTED_EXIT_DONE;
break;
@@ -850,7 +850,7 @@ static void nested_svm_init(struct vcpu_svm *svm)
}
-static int svm_check_nested_events(struct kvm_vcpu *vcpu)
+static int nested_svm_check_events(struct kvm_vcpu *vcpu)
{
struct vcpu_svm *svm = to_svm(vcpu);
bool block_nested_events =
@@ -933,7 +933,7 @@ int nested_svm_exit_special(struct vcpu_svm *svm)
return NESTED_EXIT_CONTINUE;
}
-static int svm_get_nested_state(struct kvm_vcpu *vcpu,
+static int nested_svm_get_state(struct kvm_vcpu *vcpu,
struct kvm_nested_state __user *user_kvm_nested_state,
u32 user_data_size)
{
@@ -990,7 +990,7 @@ static int svm_get_nested_state(struct kvm_vcpu *vcpu,
return kvm_state.size;
}
-static int svm_set_nested_state(struct kvm_vcpu *vcpu,
+static int nested_svm_set_state(struct kvm_vcpu *vcpu,
struct kvm_nested_state __user *user_kvm_nested_state,
struct kvm_nested_state *kvm_state)
{
@@ -1075,8 +1075,10 @@ static int svm_set_nested_state(struct kvm_vcpu *vcpu,
return 0;
}
+#define KVM_X86_NESTED_OP(name) .name = nested_svm_##name
+
struct kvm_x86_nested_ops svm_nested_ops = {
- .check_events = svm_check_nested_events,
- .get_state = svm_get_nested_state,
- .set_state = svm_set_nested_state,
+ KVM_X86_NESTED_OP(check_events),
+ KVM_X86_NESTED_OP(get_state),
+ KVM_X86_NESTED_OP(set_state),
};
--
1.8.3.1
^ permalink raw reply related [flat|nested] 10+ messages in thread
* [PATCH 4/5 v2] KVM: VMX: Fill in conforming vmx_x86_ops via macro
2020-07-25 3:26 [PATCH 0/5 v2] KVM: x86: Fill in conforming {vmx|svm}_x86_ops and {vmx|svm}_nested_ops via macros Krish Sadhukhan
` (2 preceding siblings ...)
2020-07-25 3:26 ` [PATCH 3/5 v2] KVM: nSVM: Fill in conforming svm_nested_ops " Krish Sadhukhan
@ 2020-07-25 3:26 ` Krish Sadhukhan
2020-07-25 3:26 ` [PATCH 5/5 v2] KVM: nVMX: Fill in conforming vmx_nested_ops " Krish Sadhukhan
4 siblings, 0 replies; 10+ messages in thread
From: Krish Sadhukhan @ 2020-07-25 3:26 UTC (permalink / raw)
To: kvm; +Cc: jmattson, sean.j.christopherson, pbonzini, vkuznets
The names of some of the vmx_x86_ops functions do not have a corresponding
'vmx_' prefix. Generate the names using a macro so that the names are
conformant. Fixing the naming will help in better readability and
maintenance of the code.
Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
---
arch/x86/kvm/vmx/nested.c | 2 +-
arch/x86/kvm/vmx/vmx.c | 234 +++++++++++++++++++++++-----------------------
arch/x86/kvm/vmx/vmx.h | 2 +-
3 files changed, 120 insertions(+), 118 deletions(-)
diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c
index d1af20b..a898b53 100644
--- a/arch/x86/kvm/vmx/nested.c
+++ b/arch/x86/kvm/vmx/nested.c
@@ -3016,7 +3016,7 @@ static int nested_vmx_check_vmentry_hw(struct kvm_vcpu *vcpu)
preempt_disable();
- vmx_prepare_switch_to_guest(vcpu);
+ vmx_prepare_guest_switch(vcpu);
/*
* Induce a consistency check VMExit by clearing bit 1 in GUEST_RFLAGS,
diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
index 90d91524..f6a6674 100644
--- a/arch/x86/kvm/vmx/vmx.c
+++ b/arch/x86/kvm/vmx/vmx.c
@@ -1125,7 +1125,7 @@ void vmx_set_host_fs_gs(struct vmcs_host_state *host, u16 fs_sel, u16 gs_sel,
}
}
-void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
+void vmx_prepare_guest_switch(struct kvm_vcpu *vcpu)
{
struct vcpu_vmx *vmx = to_vmx(vcpu);
struct vmcs_host_state *host_state;
@@ -2317,7 +2317,7 @@ static int kvm_cpu_vmxon(u64 vmxon_pointer)
return -EFAULT;
}
-static int hardware_enable(void)
+static int vmx_hardware_enable(void)
{
int cpu = raw_smp_processor_id();
u64 phys_addr = __pa(per_cpu(vmxarea, cpu));
@@ -2366,7 +2366,7 @@ static void kvm_cpu_vmxoff(void)
cr4_clear_bits(X86_CR4_VMXE);
}
-static void hardware_disable(void)
+static void vmx_hardware_disable(void)
{
vmclear_local_loaded_vmcss();
kvm_cpu_vmxoff();
@@ -2911,7 +2911,7 @@ static void exit_lmode(struct kvm_vcpu *vcpu)
#endif
-static void vmx_flush_tlb_all(struct kvm_vcpu *vcpu)
+static void vmx_tlb_flush_all(struct kvm_vcpu *vcpu)
{
struct vcpu_vmx *vmx = to_vmx(vcpu);
@@ -2934,7 +2934,7 @@ static void vmx_flush_tlb_all(struct kvm_vcpu *vcpu)
}
}
-static void vmx_flush_tlb_current(struct kvm_vcpu *vcpu)
+static void vmx_tlb_flush_current(struct kvm_vcpu *vcpu)
{
u64 root_hpa = vcpu->arch.mmu->root_hpa;
@@ -2950,16 +2950,16 @@ static void vmx_flush_tlb_current(struct kvm_vcpu *vcpu)
vpid_sync_context(nested_get_vpid02(vcpu));
}
-static void vmx_flush_tlb_gva(struct kvm_vcpu *vcpu, gva_t addr)
+static void vmx_tlb_flush_gva(struct kvm_vcpu *vcpu, gva_t addr)
{
/*
* vpid_sync_vcpu_addr() is a nop if vmx->vpid==0, see the comment in
- * vmx_flush_tlb_guest() for an explanation of why this is ok.
+ * vmx_tlb_flush_guest() for an explanation of why this is ok.
*/
vpid_sync_vcpu_addr(to_vmx(vcpu)->vpid, addr);
}
-static void vmx_flush_tlb_guest(struct kvm_vcpu *vcpu)
+static void vmx_tlb_flush_guest(struct kvm_vcpu *vcpu)
{
/*
* vpid_sync_context() is a nop if vmx->vpid==0, e.g. if enable_vpid==0
@@ -4455,16 +4455,16 @@ static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
vmx_clear_hlt(vcpu);
}
-static void enable_irq_window(struct kvm_vcpu *vcpu)
+static void vmx_enable_irq_window(struct kvm_vcpu *vcpu)
{
exec_controls_setbit(to_vmx(vcpu), CPU_BASED_INTR_WINDOW_EXITING);
}
-static void enable_nmi_window(struct kvm_vcpu *vcpu)
+static void vmx_enable_nmi_window(struct kvm_vcpu *vcpu)
{
if (!enable_vnmi ||
vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) & GUEST_INTR_STATE_STI) {
- enable_irq_window(vcpu);
+ vmx_enable_irq_window(vcpu);
return;
}
@@ -6173,7 +6173,7 @@ static void vmx_l1d_flush(struct kvm_vcpu *vcpu)
: "eax", "ebx", "ecx", "edx");
}
-static void update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
+static void vmx_update_cr8_intercept(struct kvm_vcpu *vcpu, int tpr, int irr)
{
struct vmcs12 *vmcs12 = get_vmcs12(vcpu);
int tpr_threshold;
@@ -6261,7 +6261,7 @@ static void vmx_set_apic_access_page_addr(struct kvm_vcpu *vcpu)
return;
vmcs_write64(APIC_ACCESS_ADDR, page_to_phys(page));
- vmx_flush_tlb_current(vcpu);
+ vmx_tlb_flush_current(vcpu);
/*
* Do not pin apic access page in memory, the MMU notifier
@@ -6837,7 +6837,7 @@ static fastpath_t vmx_vcpu_run(struct kvm_vcpu *vcpu)
return exit_fastpath;
}
-static void vmx_free_vcpu(struct kvm_vcpu *vcpu)
+static void vmx_vcpu_free(struct kvm_vcpu *vcpu)
{
struct vcpu_vmx *vmx = to_vmx(vcpu);
@@ -6848,7 +6848,7 @@ static void vmx_free_vcpu(struct kvm_vcpu *vcpu)
free_loaded_vmcs(vmx->loaded_vmcs);
}
-static int vmx_create_vcpu(struct kvm_vcpu *vcpu)
+static int vmx_vcpu_create(struct kvm_vcpu *vcpu)
{
struct vcpu_vmx *vmx;
unsigned long *msr_bitmap;
@@ -7802,7 +7802,7 @@ static int vmx_pre_leave_smm(struct kvm_vcpu *vcpu, const char *smstate)
return 0;
}
-static void enable_smi_window(struct kvm_vcpu *vcpu)
+static void vmx_enable_smi_window(struct kvm_vcpu *vcpu)
{
/* RSM will cause a vmexit anyway. */
}
@@ -7827,7 +7827,7 @@ static void vmx_migrate_timers(struct kvm_vcpu *vcpu)
}
}
-static void hardware_unsetup(void)
+static void vmx_hardware_teardown(void)
{
if (nested)
nested_vmx_hardware_unsetup();
@@ -7843,134 +7843,136 @@ static bool vmx_check_apicv_inhibit_reasons(ulong bit)
return supported & BIT(bit);
}
+#define KVM_X86_OP(name) .name = vmx_##name
+
static struct kvm_x86_ops vmx_x86_ops __initdata = {
- .hardware_teardown = hardware_unsetup,
+ KVM_X86_OP(hardware_teardown),
- .hardware_enable = hardware_enable,
- .hardware_disable = hardware_disable,
+ KVM_X86_OP(hardware_enable),
+ KVM_X86_OP(hardware_disable),
.cpu_has_accelerated_tpr = report_flexpriority,
- .has_emulated_msr = vmx_has_emulated_msr,
+ KVM_X86_OP(has_emulated_msr),
.vm_size = sizeof(struct kvm_vmx),
- .vm_init = vmx_vm_init,
+ KVM_X86_OP(vm_init),
- .vcpu_create = vmx_create_vcpu,
- .vcpu_free = vmx_free_vcpu,
- .vcpu_reset = vmx_vcpu_reset,
+ KVM_X86_OP(vcpu_create),
+ KVM_X86_OP(vcpu_free),
+ KVM_X86_OP(vcpu_reset),
- .prepare_guest_switch = vmx_prepare_switch_to_guest,
- .vcpu_load = vmx_vcpu_load,
- .vcpu_put = vmx_vcpu_put,
+ KVM_X86_OP(prepare_guest_switch),
+ KVM_X86_OP(vcpu_load),
+ KVM_X86_OP(vcpu_put),
.update_bp_intercept = update_exception_bitmap,
- .get_msr_feature = vmx_get_msr_feature,
- .get_msr = vmx_get_msr,
- .set_msr = vmx_set_msr,
- .get_segment_base = vmx_get_segment_base,
- .get_segment = vmx_get_segment,
- .set_segment = vmx_set_segment,
- .get_cpl = vmx_get_cpl,
- .get_cs_db_l_bits = vmx_get_cs_db_l_bits,
- .set_cr0 = vmx_set_cr0,
- .set_cr4 = vmx_set_cr4,
- .set_efer = vmx_set_efer,
- .get_idt = vmx_get_idt,
- .set_idt = vmx_set_idt,
- .get_gdt = vmx_get_gdt,
- .set_gdt = vmx_set_gdt,
- .set_dr7 = vmx_set_dr7,
- .sync_dirty_debug_regs = vmx_sync_dirty_debug_regs,
- .cache_reg = vmx_cache_reg,
- .get_rflags = vmx_get_rflags,
- .set_rflags = vmx_set_rflags,
-
- .tlb_flush_all = vmx_flush_tlb_all,
- .tlb_flush_current = vmx_flush_tlb_current,
- .tlb_flush_gva = vmx_flush_tlb_gva,
- .tlb_flush_guest = vmx_flush_tlb_guest,
-
- .vcpu_run = vmx_vcpu_run,
- .handle_exit = vmx_handle_exit,
- .skip_emulated_instruction = vmx_skip_emulated_instruction,
- .update_emulated_instruction = vmx_update_emulated_instruction,
- .set_interrupt_shadow = vmx_set_interrupt_shadow,
- .get_interrupt_shadow = vmx_get_interrupt_shadow,
- .patch_hypercall = vmx_patch_hypercall,
- .inject_irq = vmx_inject_irq,
- .inject_nmi = vmx_inject_nmi,
- .queue_exception = vmx_queue_exception,
- .cancel_injection = vmx_cancel_injection,
- .interrupt_allowed = vmx_interrupt_allowed,
- .nmi_allowed = vmx_nmi_allowed,
- .get_nmi_mask = vmx_get_nmi_mask,
- .set_nmi_mask = vmx_set_nmi_mask,
- .enable_nmi_window = enable_nmi_window,
- .enable_irq_window = enable_irq_window,
- .update_cr8_intercept = update_cr8_intercept,
- .set_virtual_apic_mode = vmx_set_virtual_apic_mode,
- .set_apic_access_page_addr = vmx_set_apic_access_page_addr,
- .refresh_apicv_exec_ctrl = vmx_refresh_apicv_exec_ctrl,
- .load_eoi_exitmap = vmx_load_eoi_exitmap,
- .apicv_post_state_restore = vmx_apicv_post_state_restore,
- .check_apicv_inhibit_reasons = vmx_check_apicv_inhibit_reasons,
- .hwapic_irr_update = vmx_hwapic_irr_update,
- .hwapic_isr_update = vmx_hwapic_isr_update,
- .guest_apic_has_interrupt = vmx_guest_apic_has_interrupt,
- .sync_pir_to_irr = vmx_sync_pir_to_irr,
- .deliver_posted_interrupt = vmx_deliver_posted_interrupt,
- .dy_apicv_has_pending_interrupt = vmx_dy_apicv_has_pending_interrupt,
-
- .set_tss_addr = vmx_set_tss_addr,
- .set_identity_map_addr = vmx_set_identity_map_addr,
- .get_tdp_level = vmx_get_tdp_level,
- .get_mt_mask = vmx_get_mt_mask,
-
- .get_exit_info = vmx_get_exit_info,
-
- .cpuid_update = vmx_cpuid_update,
+ KVM_X86_OP(get_msr_feature),
+ KVM_X86_OP(get_msr),
+ KVM_X86_OP(set_msr),
+ KVM_X86_OP(get_segment_base),
+ KVM_X86_OP(get_segment),
+ KVM_X86_OP(set_segment),
+ KVM_X86_OP(get_cpl),
+ KVM_X86_OP(get_cs_db_l_bits),
+ KVM_X86_OP(set_cr0),
+ KVM_X86_OP(set_cr4),
+ KVM_X86_OP(set_efer),
+ KVM_X86_OP(get_idt),
+ KVM_X86_OP(set_idt),
+ KVM_X86_OP(get_gdt),
+ KVM_X86_OP(set_gdt),
+ KVM_X86_OP(set_dr7),
+ KVM_X86_OP(sync_dirty_debug_regs),
+ KVM_X86_OP(cache_reg),
+ KVM_X86_OP(get_rflags),
+ KVM_X86_OP(set_rflags),
+
+ KVM_X86_OP(tlb_flush_all),
+ KVM_X86_OP(tlb_flush_current),
+ KVM_X86_OP(tlb_flush_gva),
+ KVM_X86_OP(tlb_flush_guest),
+
+ KVM_X86_OP(vcpu_run),
+ KVM_X86_OP(handle_exit),
+ KVM_X86_OP(skip_emulated_instruction),
+ KVM_X86_OP(update_emulated_instruction),
+ KVM_X86_OP(set_interrupt_shadow),
+ KVM_X86_OP(get_interrupt_shadow),
+ KVM_X86_OP(patch_hypercall),
+ KVM_X86_OP(inject_irq),
+ KVM_X86_OP(inject_nmi),
+ KVM_X86_OP(queue_exception),
+ KVM_X86_OP(cancel_injection),
+ KVM_X86_OP(interrupt_allowed),
+ KVM_X86_OP(nmi_allowed),
+ KVM_X86_OP(get_nmi_mask),
+ KVM_X86_OP(set_nmi_mask),
+ KVM_X86_OP(enable_nmi_window),
+ KVM_X86_OP(enable_irq_window),
+ KVM_X86_OP(update_cr8_intercept),
+ KVM_X86_OP(set_virtual_apic_mode),
+ KVM_X86_OP(set_apic_access_page_addr),
+ KVM_X86_OP(refresh_apicv_exec_ctrl),
+ KVM_X86_OP(load_eoi_exitmap),
+ KVM_X86_OP(apicv_post_state_restore),
+ KVM_X86_OP(check_apicv_inhibit_reasons),
+ KVM_X86_OP(hwapic_irr_update),
+ KVM_X86_OP(hwapic_isr_update),
+ KVM_X86_OP(guest_apic_has_interrupt),
+ KVM_X86_OP(sync_pir_to_irr),
+ KVM_X86_OP(deliver_posted_interrupt),
+ KVM_X86_OP(dy_apicv_has_pending_interrupt),
+
+ KVM_X86_OP(set_tss_addr),
+ KVM_X86_OP(set_identity_map_addr),
+ KVM_X86_OP(get_tdp_level),
+ KVM_X86_OP(get_mt_mask),
+
+ KVM_X86_OP(get_exit_info),
+
+ KVM_X86_OP(cpuid_update),
.has_wbinvd_exit = cpu_has_vmx_wbinvd_exit,
- .write_l1_tsc_offset = vmx_write_l1_tsc_offset,
+ KVM_X86_OP(write_l1_tsc_offset),
- .load_mmu_pgd = vmx_load_mmu_pgd,
+ KVM_X86_OP(load_mmu_pgd),
- .check_intercept = vmx_check_intercept,
- .handle_exit_irqoff = vmx_handle_exit_irqoff,
+ KVM_X86_OP(check_intercept),
+ KVM_X86_OP(handle_exit_irqoff),
- .request_immediate_exit = vmx_request_immediate_exit,
+ KVM_X86_OP(request_immediate_exit),
- .sched_in = vmx_sched_in,
+ KVM_X86_OP(sched_in),
- .slot_enable_log_dirty = vmx_slot_enable_log_dirty,
- .slot_disable_log_dirty = vmx_slot_disable_log_dirty,
- .flush_log_dirty = vmx_flush_log_dirty,
- .enable_log_dirty_pt_masked = vmx_enable_log_dirty_pt_masked,
+ KVM_X86_OP(slot_enable_log_dirty),
+ KVM_X86_OP(slot_disable_log_dirty),
+ KVM_X86_OP(flush_log_dirty),
+ KVM_X86_OP(enable_log_dirty_pt_masked),
.write_log_dirty = vmx_write_pml_buffer,
- .pre_block = vmx_pre_block,
- .post_block = vmx_post_block,
+ KVM_X86_OP(pre_block),
+ KVM_X86_OP(post_block),
.pmu_ops = &intel_pmu_ops,
.nested_ops = &vmx_nested_ops,
- .update_pi_irte = vmx_update_pi_irte,
+ KVM_X86_OP(update_pi_irte),
#ifdef CONFIG_X86_64
- .set_hv_timer = vmx_set_hv_timer,
- .cancel_hv_timer = vmx_cancel_hv_timer,
+ KVM_X86_OP(set_hv_timer),
+ KVM_X86_OP(cancel_hv_timer),
#endif
- .setup_mce = vmx_setup_mce,
+ KVM_X86_OP(setup_mce),
- .smi_allowed = vmx_smi_allowed,
- .pre_enter_smm = vmx_pre_enter_smm,
- .pre_leave_smm = vmx_pre_leave_smm,
- .enable_smi_window = enable_smi_window,
+ KVM_X86_OP(smi_allowed),
+ KVM_X86_OP(pre_enter_smm),
+ KVM_X86_OP(pre_leave_smm),
+ KVM_X86_OP(enable_smi_window),
- .need_emulation_on_page_fault = vmx_need_emulation_on_page_fault,
- .apic_init_signal_blocked = vmx_apic_init_signal_blocked,
- .migrate_timers = vmx_migrate_timers,
+ KVM_X86_OP(need_emulation_on_page_fault),
+ KVM_X86_OP(apic_init_signal_blocked),
+ KVM_X86_OP(migrate_timers),
};
static __init int hardware_setup(void)
diff --git a/arch/x86/kvm/vmx/vmx.h b/arch/x86/kvm/vmx/vmx.h
index 639798e..8084ce0 100644
--- a/arch/x86/kvm/vmx/vmx.h
+++ b/arch/x86/kvm/vmx/vmx.h
@@ -325,7 +325,7 @@ void vmx_vcpu_load_vmcs(struct kvm_vcpu *vcpu, int cpu,
int allocate_vpid(void);
void free_vpid(int vpid);
void vmx_set_constant_host_state(struct vcpu_vmx *vmx);
-void vmx_prepare_switch_to_guest(struct kvm_vcpu *vcpu);
+void vmx_prepare_guest_switch(struct kvm_vcpu *vcpu);
void vmx_set_host_fs_gs(struct vmcs_host_state *host, u16 fs_sel, u16 gs_sel,
unsigned long fs_base, unsigned long gs_base);
int vmx_get_cpl(struct kvm_vcpu *vcpu);
--
1.8.3.1
^ permalink raw reply related [flat|nested] 10+ messages in thread
* [PATCH 5/5 v2] KVM: nVMX: Fill in conforming vmx_nested_ops via macro
2020-07-25 3:26 [PATCH 0/5 v2] KVM: x86: Fill in conforming {vmx|svm}_x86_ops and {vmx|svm}_nested_ops via macros Krish Sadhukhan
` (3 preceding siblings ...)
2020-07-25 3:26 ` [PATCH 4/5 v2] KVM: VMX: Fill in conforming vmx_x86_ops " Krish Sadhukhan
@ 2020-07-25 3:26 ` Krish Sadhukhan
4 siblings, 0 replies; 10+ messages in thread
From: Krish Sadhukhan @ 2020-07-25 3:26 UTC (permalink / raw)
To: kvm; +Cc: jmattson, sean.j.christopherson, pbonzini, vkuznets
The names of some of the vmx_nested_ops functions do not have a corresponding
'nested_vmx_' prefix. Generate the names using a macro so that the names are
conformant. Fixing the naming will help in better readability and
maintenance of the code.
Suggested-by: Vitaly Kuznetsov <vkuznets@redhat.com>
Suggested-by: Paolo Bonzini <pbonzini@redhat.com>
Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: Krish Sadhukhan <krish.sadhukhan@oracle.com>
---
arch/x86/kvm/vmx/nested.c | 24 +++++++++++++-----------
arch/x86/kvm/vmx/nested.h | 2 +-
arch/x86/kvm/vmx/vmx.c | 4 ++--
3 files changed, 16 insertions(+), 14 deletions(-)
diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c
index a898b53..fc09bb0 100644
--- a/arch/x86/kvm/vmx/nested.c
+++ b/arch/x86/kvm/vmx/nested.c
@@ -3105,7 +3105,7 @@ static int nested_vmx_check_vmentry_hw(struct kvm_vcpu *vcpu)
return 0;
}
-static bool nested_get_vmcs12_pages(struct kvm_vcpu *vcpu)
+static bool nested_vmx_get_vmcs12_pages(struct kvm_vcpu *vcpu)
{
struct vmcs12 *vmcs12 = get_vmcs12(vcpu);
struct vcpu_vmx *vmx = to_vmx(vcpu);
@@ -3295,7 +3295,7 @@ enum nvmx_vmentry_status nested_vmx_enter_non_root_mode(struct kvm_vcpu *vcpu,
prepare_vmcs02_early(vmx, vmcs12);
if (from_vmentry) {
- if (unlikely(!nested_get_vmcs12_pages(vcpu)))
+ if (unlikely(!nested_vmx_get_vmcs12_pages(vcpu)))
return NVMX_VMENTRY_KVM_INTERNAL_ERROR;
if (nested_vmx_check_vmentry_hw(vcpu)) {
@@ -3711,7 +3711,7 @@ static bool nested_vmx_preemption_timer_pending(struct kvm_vcpu *vcpu)
to_vmx(vcpu)->nested.preemption_timer_expired;
}
-static int vmx_check_nested_events(struct kvm_vcpu *vcpu)
+static int nested_vmx_check_events(struct kvm_vcpu *vcpu)
{
struct vcpu_vmx *vmx = to_vmx(vcpu);
unsigned long exit_qual;
@@ -5907,7 +5907,7 @@ bool nested_vmx_reflect_vmexit(struct kvm_vcpu *vcpu)
return true;
}
-static int vmx_get_nested_state(struct kvm_vcpu *vcpu,
+static int nested_vmx_get_state(struct kvm_vcpu *vcpu,
struct kvm_nested_state __user *user_kvm_nested_state,
u32 user_data_size)
{
@@ -6031,7 +6031,7 @@ void vmx_leave_nested(struct kvm_vcpu *vcpu)
free_nested(vcpu);
}
-static int vmx_set_nested_state(struct kvm_vcpu *vcpu,
+static int nested_vmx_set_state(struct kvm_vcpu *vcpu,
struct kvm_nested_state __user *user_kvm_nested_state,
struct kvm_nested_state *kvm_state)
{
@@ -6448,7 +6448,7 @@ void nested_vmx_setup_ctls_msrs(struct nested_vmx_msrs *msrs, u32 ept_caps)
msrs->vmcs_enum = VMCS12_MAX_FIELD_INDEX << 1;
}
-void nested_vmx_hardware_unsetup(void)
+void nested_vmx_hardware_teardown(void)
{
int i;
@@ -6473,7 +6473,7 @@ __init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *))
vmx_bitmap[i] = (unsigned long *)
__get_free_page(GFP_KERNEL);
if (!vmx_bitmap[i]) {
- nested_vmx_hardware_unsetup();
+ nested_vmx_hardware_teardown();
return -ENOMEM;
}
}
@@ -6497,12 +6497,14 @@ __init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *))
return 0;
}
+#define KVM_X86_NESTED_OP(name) .name = nested_vmx_##name
+
struct kvm_x86_nested_ops vmx_nested_ops = {
- .check_events = vmx_check_nested_events,
+ KVM_X86_NESTED_OP(check_events),
.hv_timer_pending = nested_vmx_preemption_timer_pending,
- .get_state = vmx_get_nested_state,
- .set_state = vmx_set_nested_state,
- .get_vmcs12_pages = nested_get_vmcs12_pages,
+ KVM_X86_NESTED_OP(get_state),
+ KVM_X86_NESTED_OP(set_state),
+ KVM_X86_NESTED_OP(get_vmcs12_pages),
.enable_evmcs = nested_enable_evmcs,
.get_evmcs_version = nested_get_evmcs_version,
};
diff --git a/arch/x86/kvm/vmx/nested.h b/arch/x86/kvm/vmx/nested.h
index 758bccc..ac6b561 100644
--- a/arch/x86/kvm/vmx/nested.h
+++ b/arch/x86/kvm/vmx/nested.h
@@ -18,7 +18,7 @@ enum nvmx_vmentry_status {
void vmx_leave_nested(struct kvm_vcpu *vcpu);
void nested_vmx_setup_ctls_msrs(struct nested_vmx_msrs *msrs, u32 ept_caps);
-void nested_vmx_hardware_unsetup(void);
+void nested_vmx_hardware_teardown(void);
__init int nested_vmx_hardware_setup(int (*exit_handlers[])(struct kvm_vcpu *));
void nested_vmx_set_vmcs_shadowing_bitmap(void);
void nested_vmx_free_vcpu(struct kvm_vcpu *vcpu);
diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c
index f6a6674..6512e6e 100644
--- a/arch/x86/kvm/vmx/vmx.c
+++ b/arch/x86/kvm/vmx/vmx.c
@@ -7830,7 +7830,7 @@ static void vmx_migrate_timers(struct kvm_vcpu *vcpu)
static void vmx_hardware_teardown(void)
{
if (nested)
- nested_vmx_hardware_unsetup();
+ nested_vmx_hardware_teardown();
free_kvm_area();
}
@@ -8144,7 +8144,7 @@ static __init int hardware_setup(void)
r = alloc_kvm_area();
if (r)
- nested_vmx_hardware_unsetup();
+ nested_vmx_hardware_teardown();
return r;
}
--
1.8.3.1
^ permalink raw reply related [flat|nested] 10+ messages in thread