From: janani <janani@linux.ibm.com>
To: Bharata B Rao <bharata@linux.ibm.com>
Cc: linuxppc-dev@lists.ozlabs.org, linuxram@us.ibm.com,
cclaudio@linux.ibm.com, kvm-ppc@vger.kernel.org,
linux-mm@kvack.org, jglisse@redhat.com,
aneesh.kumar@linux.vnet.ibm.com, paulus@au1.ibm.com,
sukadev@linux.vnet.ibm.com,
Linuxppc-dev
<linuxppc-dev-bounces+janani=linux.ibm.com@lists.ozlabs.org>
Subject: Re: [RFC PATCH v5 5/7] kvmppc: Radix changes for secure guest
Date: Tue, 09 Jul 2019 14:53:47 -0500 [thread overview]
Message-ID: <5c7231766bc1f78e3cc1a467186e3356@linux.vnet.ibm.com> (raw)
In-Reply-To: <20190709102545.9187-6-bharata@linux.ibm.com>
On 2019-07-09 05:25, Bharata B Rao wrote:
> - After the guest becomes secure, when we handle a page fault of a page
> belonging to SVM in HV, send that page to UV via UV_PAGE_IN.
> - Whenever a page is unmapped on the HV side, inform UV via
> UV_PAGE_INVAL.
> - Ensure all those routines that walk the secondary page tables of
> the guest don't do so in case of secure VM. For secure guest, the
> active secondary page tables are in secure memory and the secondary
> page tables in HV are freed when guest becomes secure.
>
> Signed-off-by: Bharata B Rao <bharata@linux.ibm.com>
Reviewed-by: Janani Janakiraman <janani@linux.ibm.com>
> ---
> arch/powerpc/include/asm/kvm_host.h | 12 ++++++++++++
> arch/powerpc/include/asm/ultravisor-api.h | 1 +
> arch/powerpc/include/asm/ultravisor.h | 7 +++++++
> arch/powerpc/kvm/book3s_64_mmu_radix.c | 22 ++++++++++++++++++++++
> arch/powerpc/kvm/book3s_hv_hmm.c | 20 ++++++++++++++++++++
> 5 files changed, 62 insertions(+)
>
> diff --git a/arch/powerpc/include/asm/kvm_host.h
> b/arch/powerpc/include/asm/kvm_host.h
> index 0c49c3401c63..dcbf7480cb10 100644
> --- a/arch/powerpc/include/asm/kvm_host.h
> +++ b/arch/powerpc/include/asm/kvm_host.h
> @@ -865,6 +865,8 @@ static inline void
> kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {}
> #ifdef CONFIG_PPC_UV
> extern int kvmppc_hmm_init(void);
> extern void kvmppc_hmm_free(void);
> +extern bool kvmppc_is_guest_secure(struct kvm *kvm);
> +extern int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gpa);
> #else
> static inline int kvmppc_hmm_init(void)
> {
> @@ -872,6 +874,16 @@ static inline int kvmppc_hmm_init(void)
> }
>
> static inline void kvmppc_hmm_free(void) {}
> +
> +static inline bool kvmppc_is_guest_secure(struct kvm *kvm)
> +{
> + return false;
> +}
> +
> +static inline int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned
> long gpa)
> +{
> + return -EFAULT;
> +}
> #endif /* CONFIG_PPC_UV */
>
> #endif /* __POWERPC_KVM_HOST_H__ */
> diff --git a/arch/powerpc/include/asm/ultravisor-api.h
> b/arch/powerpc/include/asm/ultravisor-api.h
> index d6d6eb2e6e6b..9f5510b55892 100644
> --- a/arch/powerpc/include/asm/ultravisor-api.h
> +++ b/arch/powerpc/include/asm/ultravisor-api.h
> @@ -24,5 +24,6 @@
> #define UV_UNREGISTER_MEM_SLOT 0xF124
> #define UV_PAGE_IN 0xF128
> #define UV_PAGE_OUT 0xF12C
> +#define UV_PAGE_INVAL 0xF138
>
> #endif /* _ASM_POWERPC_ULTRAVISOR_API_H */
> diff --git a/arch/powerpc/include/asm/ultravisor.h
> b/arch/powerpc/include/asm/ultravisor.h
> index fe45be9ee63b..f4f674794b35 100644
> --- a/arch/powerpc/include/asm/ultravisor.h
> +++ b/arch/powerpc/include/asm/ultravisor.h
> @@ -77,6 +77,13 @@ static inline int uv_unregister_mem_slot(u64 lpid,
> u64 slotid)
>
> return ucall(UV_UNREGISTER_MEM_SLOT, retbuf, lpid, slotid);
> }
> +
> +static inline int uv_page_inval(u64 lpid, u64 gpa, u64 page_shift)
> +{
> + unsigned long retbuf[UCALL_BUFSIZE];
> +
> + return ucall(UV_PAGE_INVAL, retbuf, lpid, gpa, page_shift);
> +}
> #endif /* !__ASSEMBLY__ */
>
> #endif /* _ASM_POWERPC_ULTRAVISOR_H */
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> index f55ef071883f..c454600c454f 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> @@ -21,6 +21,8 @@
> #include <asm/pgtable.h>
> #include <asm/pgalloc.h>
> #include <asm/pte-walk.h>
> +#include <asm/ultravisor.h>
> +#include <asm/kvm_host.h>
>
> /*
> * Supported radix tree geometry.
> @@ -923,6 +925,9 @@ int kvmppc_book3s_radix_page_fault(struct kvm_run
> *run, struct kvm_vcpu *vcpu,
> if (!(dsisr & DSISR_PRTABLE_FAULT))
> gpa |= ea & 0xfff;
>
> + if (kvmppc_is_guest_secure(kvm))
> + return kvmppc_send_page_to_uv(kvm, gpa & PAGE_MASK);
> +
> /* Get the corresponding memslot */
> memslot = gfn_to_memslot(kvm, gfn);
>
> @@ -980,6 +985,11 @@ int kvm_unmap_radix(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> unsigned long gpa = gfn << PAGE_SHIFT;
> unsigned int shift;
>
> + if (kvmppc_is_guest_secure(kvm)) {
> + uv_page_inval(kvm->arch.lpid, gpa, PAGE_SIZE);
> + return 0;
> + }
> +
> ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift);
> if (ptep && pte_present(*ptep))
> kvmppc_unmap_pte(kvm, ptep, gpa, shift, memslot,
> @@ -997,6 +1007,9 @@ int kvm_age_radix(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> int ref = 0;
> unsigned long old, *rmapp;
>
> + if (kvmppc_is_guest_secure(kvm))
> + return ref;
> +
> ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift);
> if (ptep && pte_present(*ptep) && pte_young(*ptep)) {
> old = kvmppc_radix_update_pte(kvm, ptep, _PAGE_ACCESSED, 0,
> @@ -1021,6 +1034,9 @@ int kvm_test_age_radix(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> unsigned int shift;
> int ref = 0;
>
> + if (kvmppc_is_guest_secure(kvm))
> + return ref;
> +
> ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift);
> if (ptep && pte_present(*ptep) && pte_young(*ptep))
> ref = 1;
> @@ -1038,6 +1054,9 @@ static int kvm_radix_test_clear_dirty(struct kvm
> *kvm,
> int ret = 0;
> unsigned long old, *rmapp;
>
> + if (kvmppc_is_guest_secure(kvm))
> + return ret;
> +
> ptep = __find_linux_pte(kvm->arch.pgtable, gpa, NULL, &shift);
> if (ptep && pte_present(*ptep) && pte_dirty(*ptep)) {
> ret = 1;
> @@ -1090,6 +1109,9 @@ void kvmppc_radix_flush_memslot(struct kvm *kvm,
> unsigned long gpa;
> unsigned int shift;
>
> + if (kvmppc_is_guest_secure(kvm))
> + return;
> +
> gpa = memslot->base_gfn << PAGE_SHIFT;
> spin_lock(&kvm->mmu_lock);
> for (n = memslot->npages; n; --n) {
> diff --git a/arch/powerpc/kvm/book3s_hv_hmm.c
> b/arch/powerpc/kvm/book3s_hv_hmm.c
> index 55bab9c4e60a..9e6c88de456f 100644
> --- a/arch/powerpc/kvm/book3s_hv_hmm.c
> +++ b/arch/powerpc/kvm/book3s_hv_hmm.c
> @@ -62,6 +62,11 @@ struct kvmppc_hmm_migrate_args {
> unsigned long page_shift;
> };
>
> +bool kvmppc_is_guest_secure(struct kvm *kvm)
> +{
> + return !!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_DONE);
> +}
> +
> unsigned long kvmppc_h_svm_init_start(struct kvm *kvm)
> {
> struct kvm_memslots *slots;
> @@ -494,6 +499,21 @@ kvmppc_h_svm_page_out(struct kvm *kvm, unsigned
> long gpa,
> return ret;
> }
>
> +int kvmppc_send_page_to_uv(struct kvm *kvm, unsigned long gpa)
> +{
> + unsigned long pfn;
> + int ret;
> +
> + pfn = gfn_to_pfn(kvm, gpa >> PAGE_SHIFT);
> + if (is_error_noslot_pfn(pfn))
> + return -EFAULT;
> +
> + ret = uv_page_in(kvm->arch.lpid, pfn << PAGE_SHIFT, gpa, 0,
> PAGE_SHIFT);
> + kvm_release_pfn_clean(pfn);
> +
> + return (ret == U_SUCCESS) ? RESUME_GUEST : -EFAULT;
> +}
> +
> static u64 kvmppc_get_secmem_size(void)
> {
> struct device_node *np;
next prev parent reply other threads:[~2019-07-09 19:51 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-07-09 10:25 [PATCH v5 0/7] kvmppc: HMM driver to manage pages of secure guest Bharata B Rao
2019-07-09 10:25 ` [PATCH v5 1/7] kvmppc: HMM backend " Bharata B Rao
2019-07-09 18:55 ` janani
2019-07-10 13:47 ` Jason Gunthorpe
2019-07-11 5:08 ` Bharata B Rao
2019-07-19 6:46 ` Christoph Hellwig
2019-07-19 8:14 ` Bharata B Rao
2019-07-09 10:25 ` [PATCH v5 2/7] kvmppc: Shared pages support for secure guests Bharata B Rao
2019-07-09 19:35 ` janani
2019-07-09 10:25 ` [PATCH v5 3/7] kvmppc: H_SVM_INIT_START and H_SVM_INIT_DONE hcalls Bharata B Rao
2019-07-09 19:42 ` janani
2019-07-09 10:25 ` [PATCH v5 4/7] kvmppc: Handle memory plug/unplug to secure VM Bharata B Rao
2019-07-09 19:43 ` janani
2019-07-09 10:25 ` [RFC PATCH v5 5/7] kvmppc: Radix changes for secure guest Bharata B Rao
2019-07-09 19:53 ` janani [this message]
2019-07-09 10:25 ` [RFC PATCH v5 6/7] kvmppc: Support reset of " Bharata B Rao
2019-07-10 13:30 ` janani
2019-07-09 10:25 ` [PATCH v5 7/7] KVM: PPC: Ultravisor: Add PPC_UV config option Bharata B Rao
2019-07-10 13:24 ` janani
2019-07-10 14:15 ` Jason Gunthorpe
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=5c7231766bc1f78e3cc1a467186e3356@linux.vnet.ibm.com \
--to=janani@linux.ibm.com \
--cc=aneesh.kumar@linux.vnet.ibm.com \
--cc=bharata@linux.ibm.com \
--cc=cclaudio@linux.ibm.com \
--cc=jglisse@redhat.com \
--cc=kvm-ppc@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linuxppc-dev-bounces+janani=linux.ibm.com@lists.ozlabs.org \
--cc=linuxppc-dev@lists.ozlabs.org \
--cc=linuxram@us.ibm.com \
--cc=paulus@au1.ibm.com \
--cc=sukadev@linux.vnet.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).