From: Paolo Bonzini <pbonzini@redhat.com>
To: Sean Christopherson <sean.j.christopherson@intel.com>
Cc: Vitaly Kuznetsov <vkuznets@redhat.com>,
Wanpeng Li <wanpengli@tencent.com>,
Jim Mattson <jmattson@google.com>, Joerg Roedel <joro@8bytes.org>,
kvm@vger.kernel.org, linux-kernel@vger.kernel.org
Subject: Re: [PATCH v2 6/7] KVM: x86/mmu: Rename kvm_mmu->get_cr3() to ->get_guest_cr3_or_eptp()
Date: Wed, 12 Feb 2020 13:00:59 +0100 [thread overview]
Message-ID: <1424348b-7f09-513a-960b-6d15ac3a9ae4@redhat.com> (raw)
In-Reply-To: <20200207173747.6243-7-sean.j.christopherson@intel.com>
On 07/02/20 18:37, Sean Christopherson wrote:
> Rename kvm_mmu->get_cr3() to call out that it is retrieving a guest
> value, as opposed to kvm_mmu->set_cr3(), which sets a host value, and to
> note that it will return L1's EPTP when nested EPT is in use. Hopefully
> the new name will also make it more obvious that L1's nested_cr3 is
> returned in SVM's nested NPT case.
>
> No functional change intended.
Should we call it "get_pgd", since that is how Linux calls the top-level
directory? I always get confused by PUD/PMD, but as long as we only
keep one /p.d/ moniker it should be fine.
Paolo
> Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
> ---
> arch/x86/include/asm/kvm_host.h | 2 +-
> arch/x86/kvm/mmu/mmu.c | 24 ++++++++++++------------
> arch/x86/kvm/mmu/paging_tmpl.h | 2 +-
> arch/x86/kvm/svm.c | 10 +++++-----
> arch/x86/kvm/vmx/nested.c | 8 ++++----
> arch/x86/kvm/x86.c | 2 +-
> 6 files changed, 24 insertions(+), 24 deletions(-)
>
> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
> index 4dffbc10d3f8..d3d69ad2e969 100644
> --- a/arch/x86/include/asm/kvm_host.h
> +++ b/arch/x86/include/asm/kvm_host.h
> @@ -383,7 +383,7 @@ struct kvm_mmu_root_info {
> */
> struct kvm_mmu {
> void (*set_cr3)(struct kvm_vcpu *vcpu, unsigned long root);
> - unsigned long (*get_cr3)(struct kvm_vcpu *vcpu);
> + unsigned long (*get_guest_cr3_or_eptp)(struct kvm_vcpu *vcpu);
> u64 (*get_pdptr)(struct kvm_vcpu *vcpu, int index);
> int (*page_fault)(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u32 err,
> bool prefault);
> diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c
> index 70f67bcab7db..13df4b4a5649 100644
> --- a/arch/x86/kvm/mmu/mmu.c
> +++ b/arch/x86/kvm/mmu/mmu.c
> @@ -3731,7 +3731,7 @@ static int mmu_alloc_direct_roots(struct kvm_vcpu *vcpu)
> vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->pae_root);
> } else
> BUG();
> - vcpu->arch.mmu->root_cr3 = vcpu->arch.mmu->get_cr3(vcpu);
> + vcpu->arch.mmu->root_cr3 = vcpu->arch.mmu->get_guest_cr3_or_eptp(vcpu);
>
> return 0;
> }
> @@ -3743,7 +3743,7 @@ static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu)
> gfn_t root_gfn, root_cr3;
> int i;
>
> - root_cr3 = vcpu->arch.mmu->get_cr3(vcpu);
> + root_cr3 = vcpu->arch.mmu->get_guest_cr3_or_eptp(vcpu);
> root_gfn = root_cr3 >> PAGE_SHIFT;
>
> if (mmu_check_root(vcpu, root_gfn))
> @@ -4080,7 +4080,7 @@ static int kvm_arch_setup_async_pf(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa,
> arch.token = (vcpu->arch.apf.id++ << 12) | vcpu->vcpu_id;
> arch.gfn = gfn;
> arch.direct_map = vcpu->arch.mmu->direct_map;
> - arch.cr3 = vcpu->arch.mmu->get_cr3(vcpu);
> + arch.cr3 = vcpu->arch.mmu->get_guest_cr3_or_eptp(vcpu);
>
> return kvm_setup_async_pf(vcpu, cr2_or_gpa,
> kvm_vcpu_gfn_to_hva(vcpu, gfn), &arch);
> @@ -4932,7 +4932,7 @@ static void init_kvm_tdp_mmu(struct kvm_vcpu *vcpu)
> context->shadow_root_level = kvm_x86_ops->get_tdp_level(vcpu);
> context->direct_map = true;
> context->set_cr3 = kvm_x86_ops->set_tdp_cr3;
> - context->get_cr3 = get_cr3;
> + context->get_guest_cr3_or_eptp = get_cr3;
> context->get_pdptr = kvm_pdptr_read;
> context->inject_page_fault = kvm_inject_page_fault;
>
> @@ -5080,10 +5080,10 @@ static void init_kvm_softmmu(struct kvm_vcpu *vcpu)
> struct kvm_mmu *context = vcpu->arch.mmu;
>
> kvm_init_shadow_mmu(vcpu);
> - context->set_cr3 = kvm_x86_ops->set_cr3;
> - context->get_cr3 = get_cr3;
> - context->get_pdptr = kvm_pdptr_read;
> - context->inject_page_fault = kvm_inject_page_fault;
> + context->set_cr3 = kvm_x86_ops->set_cr3;
> + context->get_guest_cr3_or_eptp = get_cr3;
> + context->get_pdptr = kvm_pdptr_read;
> + context->inject_page_fault = kvm_inject_page_fault;
> }
>
> static void init_kvm_nested_mmu(struct kvm_vcpu *vcpu)
> @@ -5095,10 +5095,10 @@ static void init_kvm_nested_mmu(struct kvm_vcpu *vcpu)
> if (new_role.as_u64 == g_context->mmu_role.as_u64)
> return;
>
> - g_context->mmu_role.as_u64 = new_role.as_u64;
> - g_context->get_cr3 = get_cr3;
> - g_context->get_pdptr = kvm_pdptr_read;
> - g_context->inject_page_fault = kvm_inject_page_fault;
> + g_context->mmu_role.as_u64 = new_role.as_u64;
> + g_context->get_guest_cr3_or_eptp = get_cr3;
> + g_context->get_pdptr = kvm_pdptr_read;
> + g_context->inject_page_fault = kvm_inject_page_fault;
>
> /*
> * Note that arch.mmu->gva_to_gpa translates l2_gpa to l1_gpa using
> diff --git a/arch/x86/kvm/mmu/paging_tmpl.h b/arch/x86/kvm/mmu/paging_tmpl.h
> index 6b15b58f3ecc..24dfa0fcba56 100644
> --- a/arch/x86/kvm/mmu/paging_tmpl.h
> +++ b/arch/x86/kvm/mmu/paging_tmpl.h
> @@ -333,7 +333,7 @@ static int FNAME(walk_addr_generic)(struct guest_walker *walker,
> trace_kvm_mmu_pagetable_walk(addr, access);
> retry_walk:
> walker->level = mmu->root_level;
> - pte = mmu->get_cr3(vcpu);
> + pte = mmu->get_guest_cr3_or_eptp(vcpu);
> have_ad = PT_HAVE_ACCESSED_DIRTY(mmu);
>
> #if PTTYPE == 64
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index a3e32d61d60c..1e2f05a79417 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -3026,11 +3026,11 @@ static void nested_svm_init_mmu_context(struct kvm_vcpu *vcpu)
>
> vcpu->arch.mmu = &vcpu->arch.guest_mmu;
> kvm_init_shadow_mmu(vcpu);
> - vcpu->arch.mmu->set_cr3 = nested_svm_set_tdp_cr3;
> - vcpu->arch.mmu->get_cr3 = nested_svm_get_tdp_cr3;
> - vcpu->arch.mmu->get_pdptr = nested_svm_get_tdp_pdptr;
> - vcpu->arch.mmu->inject_page_fault = nested_svm_inject_npf_exit;
> - vcpu->arch.mmu->shadow_root_level = get_npt_level(vcpu);
> + vcpu->arch.mmu->set_cr3 = nested_svm_set_tdp_cr3;
> + vcpu->arch.mmu->get_guest_cr3_or_eptp = nested_svm_get_tdp_cr3;
> + vcpu->arch.mmu->get_pdptr = nested_svm_get_tdp_pdptr;
> + vcpu->arch.mmu->inject_page_fault = nested_svm_inject_npf_exit;
> + vcpu->arch.mmu->shadow_root_level = get_npt_level(vcpu);
> reset_shadow_zero_bits_mask(vcpu, vcpu->arch.mmu);
> vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu;
> }
> diff --git a/arch/x86/kvm/vmx/nested.c b/arch/x86/kvm/vmx/nested.c
> index 4fb05c0e29fe..2d7b87b532f5 100644
> --- a/arch/x86/kvm/vmx/nested.c
> +++ b/arch/x86/kvm/vmx/nested.c
> @@ -354,10 +354,10 @@ static void nested_ept_init_mmu_context(struct kvm_vcpu *vcpu)
> VMX_EPT_EXECUTE_ONLY_BIT,
> nested_ept_ad_enabled(vcpu),
> nested_ept_get_eptp(vcpu));
> - vcpu->arch.mmu->set_cr3 = vmx_set_cr3;
> - vcpu->arch.mmu->get_cr3 = nested_ept_get_eptp;
> - vcpu->arch.mmu->inject_page_fault = nested_ept_inject_page_fault;
> - vcpu->arch.mmu->get_pdptr = kvm_pdptr_read;
> + vcpu->arch.mmu->set_cr3 = vmx_set_cr3;
> + vcpu->arch.mmu->get_guest_cr3_or_eptp = nested_ept_get_eptp;
> + vcpu->arch.mmu->inject_page_fault = nested_ept_inject_page_fault;
> + vcpu->arch.mmu->get_pdptr = kvm_pdptr_read;
>
> vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu;
> }
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index fbabb2f06273..1e6d8766fbdd 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -10179,7 +10179,7 @@ void kvm_arch_async_page_ready(struct kvm_vcpu *vcpu, struct kvm_async_pf *work)
> return;
>
> if (!vcpu->arch.mmu->direct_map &&
> - work->arch.cr3 != vcpu->arch.mmu->get_cr3(vcpu))
> + work->arch.cr3 != vcpu->arch.mmu->get_guest_cr3_or_eptp(vcpu))
> return;
>
> vcpu->arch.mmu->page_fault(vcpu, work->cr2_or_gpa, 0, true);
>
next prev parent reply other threads:[~2020-02-12 12:01 UTC|newest]
Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-02-07 17:37 [PATCH v2 0/7] KVM: x86/mmu: nVMX: 5-level paging fixes and enabling Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 1/7] KVM: nVMX: Use correct root level for nested EPT shadow page tables Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 2/7] KVM: x86/mmu: Fix struct guest_walker arrays for 5-level paging Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 3/7] KVM: nVMX: Allow L1 to use 5-level page walks for nested EPT Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 4/7] KVM: nVMX: Rename nested_ept_get_cr3() to nested_ept_get_eptp() Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 5/7] KVM: nVMX: Rename EPTP validity helper and associated variables Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 6/7] KVM: x86/mmu: Rename kvm_mmu->get_cr3() to ->get_guest_cr3_or_eptp() Sean Christopherson
2020-02-12 12:00 ` Paolo Bonzini [this message]
2020-02-12 16:28 ` Sean Christopherson
2020-02-12 16:42 ` Paolo Bonzini
2020-03-01 17:49 ` Sean Christopherson
2020-02-07 17:37 ` [PATCH v2 7/7] KVM: nVMX: Drop unnecessary check on ept caps for execute-only Sean Christopherson
2020-02-12 12:03 ` [PATCH v2 0/7] KVM: x86/mmu: nVMX: 5-level paging fixes and enabling Paolo Bonzini
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1424348b-7f09-513a-960b-6d15ac3a9ae4@redhat.com \
--to=pbonzini@redhat.com \
--cc=jmattson@google.com \
--cc=joro@8bytes.org \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=sean.j.christopherson@intel.com \
--cc=vkuznets@redhat.com \
--cc=wanpengli@tencent.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).