From: Punit Agrawal <punit.agrawal@arm.com> To: kvmarm@lists.cs.columbia.edu Cc: Punit Agrawal <punit.agrawal@arm.com>, marc.zyngier@arm.com, will.deacon@arm.com, linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org, suzuki.poulose@arm.com, punitagrawal@gmail.com, Russell King <linux@armlinux.org.uk>, Catalin Marinas <catalin.marinas@arm.com> Subject: [PATCH v9 4/8] KVM: arm64: Support dirty page tracking for PUD hugepages Date: Wed, 31 Oct 2018 17:57:41 +0000 [thread overview] Message-ID: <20181031175745.18650-5-punit.agrawal@arm.com> (raw) In-Reply-To: <20181031175745.18650-1-punit.agrawal@arm.com> In preparation for creating PUD hugepages at stage 2, add support for write protecting PUD hugepages when they are encountered. Write protecting guest tables is used to track dirty pages when migrating VMs. Also, provide trivial implementations of required kvm_s2pud_* helpers to allow sharing of code with arm32. Signed-off-by: Punit Agrawal <punit.agrawal@arm.com> Reviewed-by: Christoffer Dall <christoffer.dall@arm.com> Reviewed-by: Suzuki K Poulose <suzuki.poulose@arm.com> Cc: Marc Zyngier <marc.zyngier@arm.com> Cc: Russell King <linux@armlinux.org.uk> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Will Deacon <will.deacon@arm.com> --- arch/arm/include/asm/kvm_mmu.h | 15 +++++++++++++++ arch/arm64/include/asm/kvm_mmu.h | 10 ++++++++++ virt/kvm/arm/mmu.c | 11 +++++++---- 3 files changed, 32 insertions(+), 4 deletions(-) diff --git a/arch/arm/include/asm/kvm_mmu.h b/arch/arm/include/asm/kvm_mmu.h index e6eff8bf5d7f..37bf85d39607 100644 --- a/arch/arm/include/asm/kvm_mmu.h +++ b/arch/arm/include/asm/kvm_mmu.h @@ -87,6 +87,21 @@ void kvm_clear_hyp_idmap(void); #define kvm_pmd_mkhuge(pmd) pmd_mkhuge(pmd) +/* + * The following kvm_*pud*() functions are provided strictly to allow + * sharing code with arm64. They should never be called in practice. + */ +static inline void kvm_set_s2pud_readonly(pud_t *pud) +{ + BUG(); +} + +static inline bool kvm_s2pud_readonly(pud_t *pud) +{ + BUG(); + return false; +} + static inline pte_t kvm_s2pte_mkwrite(pte_t pte) { pte_val(pte) |= L_PTE_S2_RDWR; diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h index 13d482710292..8da6d1b2a196 100644 --- a/arch/arm64/include/asm/kvm_mmu.h +++ b/arch/arm64/include/asm/kvm_mmu.h @@ -251,6 +251,16 @@ static inline bool kvm_s2pmd_exec(pmd_t *pmdp) return !(READ_ONCE(pmd_val(*pmdp)) & PMD_S2_XN); } +static inline void kvm_set_s2pud_readonly(pud_t *pudp) +{ + kvm_set_s2pte_readonly((pte_t *)pudp); +} + +static inline bool kvm_s2pud_readonly(pud_t *pudp) +{ + return kvm_s2pte_readonly((pte_t *)pudp); +} + #define hyp_pte_table_empty(ptep) kvm_page_empty(ptep) #ifdef __PAGETABLE_PMD_FOLDED diff --git a/virt/kvm/arm/mmu.c b/virt/kvm/arm/mmu.c index fb5325f7a1ac..1c669c3c1208 100644 --- a/virt/kvm/arm/mmu.c +++ b/virt/kvm/arm/mmu.c @@ -1347,9 +1347,12 @@ static void stage2_wp_puds(struct kvm *kvm, pgd_t *pgd, do { next = stage2_pud_addr_end(kvm, addr, end); if (!stage2_pud_none(kvm, *pud)) { - /* TODO:PUD not supported, revisit later if supported */ - BUG_ON(stage2_pud_huge(kvm, *pud)); - stage2_wp_pmds(kvm, pud, addr, next); + if (stage2_pud_huge(kvm, *pud)) { + if (!kvm_s2pud_readonly(pud)) + kvm_set_s2pud_readonly(pud); + } else { + stage2_wp_pmds(kvm, pud, addr, next); + } } } while (pud++, addr = next, addr != end); } @@ -1392,7 +1395,7 @@ static void stage2_wp_range(struct kvm *kvm, phys_addr_t addr, phys_addr_t end) * * Called to start logging dirty pages after memory region * KVM_MEM_LOG_DIRTY_PAGES operation is called. After this function returns - * all present PMD and PTEs are write protected in the memory region. + * all present PUD, PMD and PTEs are write protected in the memory region. * Afterwards read of dirty page log can be called. * * Acquires kvm_mmu_lock. Called with kvm->slots_lock mutex acquired, -- 2.19.1
WARNING: multiple messages have this Message-ID (diff)
From: punit.agrawal@arm.com (Punit Agrawal) To: linux-arm-kernel@lists.infradead.org Subject: [PATCH v9 4/8] KVM: arm64: Support dirty page tracking for PUD hugepages Date: Wed, 31 Oct 2018 17:57:41 +0000 [thread overview] Message-ID: <20181031175745.18650-5-punit.agrawal@arm.com> (raw) In-Reply-To: <20181031175745.18650-1-punit.agrawal@arm.com> In preparation for creating PUD hugepages at stage 2, add support for write protecting PUD hugepages when they are encountered. Write protecting guest tables is used to track dirty pages when migrating VMs. Also, provide trivial implementations of required kvm_s2pud_* helpers to allow sharing of code with arm32. Signed-off-by: Punit Agrawal <punit.agrawal@arm.com> Reviewed-by: Christoffer Dall <christoffer.dall@arm.com> Reviewed-by: Suzuki K Poulose <suzuki.poulose@arm.com> Cc: Marc Zyngier <marc.zyngier@arm.com> Cc: Russell King <linux@armlinux.org.uk> Cc: Catalin Marinas <catalin.marinas@arm.com> Cc: Will Deacon <will.deacon@arm.com> --- arch/arm/include/asm/kvm_mmu.h | 15 +++++++++++++++ arch/arm64/include/asm/kvm_mmu.h | 10 ++++++++++ virt/kvm/arm/mmu.c | 11 +++++++---- 3 files changed, 32 insertions(+), 4 deletions(-) diff --git a/arch/arm/include/asm/kvm_mmu.h b/arch/arm/include/asm/kvm_mmu.h index e6eff8bf5d7f..37bf85d39607 100644 --- a/arch/arm/include/asm/kvm_mmu.h +++ b/arch/arm/include/asm/kvm_mmu.h @@ -87,6 +87,21 @@ void kvm_clear_hyp_idmap(void); #define kvm_pmd_mkhuge(pmd) pmd_mkhuge(pmd) +/* + * The following kvm_*pud*() functions are provided strictly to allow + * sharing code with arm64. They should never be called in practice. + */ +static inline void kvm_set_s2pud_readonly(pud_t *pud) +{ + BUG(); +} + +static inline bool kvm_s2pud_readonly(pud_t *pud) +{ + BUG(); + return false; +} + static inline pte_t kvm_s2pte_mkwrite(pte_t pte) { pte_val(pte) |= L_PTE_S2_RDWR; diff --git a/arch/arm64/include/asm/kvm_mmu.h b/arch/arm64/include/asm/kvm_mmu.h index 13d482710292..8da6d1b2a196 100644 --- a/arch/arm64/include/asm/kvm_mmu.h +++ b/arch/arm64/include/asm/kvm_mmu.h @@ -251,6 +251,16 @@ static inline bool kvm_s2pmd_exec(pmd_t *pmdp) return !(READ_ONCE(pmd_val(*pmdp)) & PMD_S2_XN); } +static inline void kvm_set_s2pud_readonly(pud_t *pudp) +{ + kvm_set_s2pte_readonly((pte_t *)pudp); +} + +static inline bool kvm_s2pud_readonly(pud_t *pudp) +{ + return kvm_s2pte_readonly((pte_t *)pudp); +} + #define hyp_pte_table_empty(ptep) kvm_page_empty(ptep) #ifdef __PAGETABLE_PMD_FOLDED diff --git a/virt/kvm/arm/mmu.c b/virt/kvm/arm/mmu.c index fb5325f7a1ac..1c669c3c1208 100644 --- a/virt/kvm/arm/mmu.c +++ b/virt/kvm/arm/mmu.c @@ -1347,9 +1347,12 @@ static void stage2_wp_puds(struct kvm *kvm, pgd_t *pgd, do { next = stage2_pud_addr_end(kvm, addr, end); if (!stage2_pud_none(kvm, *pud)) { - /* TODO:PUD not supported, revisit later if supported */ - BUG_ON(stage2_pud_huge(kvm, *pud)); - stage2_wp_pmds(kvm, pud, addr, next); + if (stage2_pud_huge(kvm, *pud)) { + if (!kvm_s2pud_readonly(pud)) + kvm_set_s2pud_readonly(pud); + } else { + stage2_wp_pmds(kvm, pud, addr, next); + } } } while (pud++, addr = next, addr != end); } @@ -1392,7 +1395,7 @@ static void stage2_wp_range(struct kvm *kvm, phys_addr_t addr, phys_addr_t end) * * Called to start logging dirty pages after memory region * KVM_MEM_LOG_DIRTY_PAGES operation is called. After this function returns - * all present PMD and PTEs are write protected in the memory region. + * all present PUD, PMD and PTEs are write protected in the memory region. * Afterwards read of dirty page log can be called. * * Acquires kvm_mmu_lock. Called with kvm->slots_lock mutex acquired, -- 2.19.1
next prev parent reply other threads:[~2018-10-31 17:58 UTC|newest] Thread overview: 69+ messages / expand[flat|nested] mbox.gz Atom feed top 2018-10-31 17:57 [PATCH v9 0/8] KVM: Support PUD hugepage at stage 2 Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-10-31 17:57 ` [PATCH v9 1/8] KVM: arm/arm64: Share common code in user_mem_abort() Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-12-03 12:11 ` Anshuman Khandual 2018-12-03 12:11 ` Anshuman Khandual 2018-12-03 13:37 ` Suzuki K Poulose 2018-12-03 13:37 ` Suzuki K Poulose 2018-12-03 13:37 ` Suzuki K Poulose 2018-12-10 8:56 ` Christoffer Dall 2018-12-10 8:56 ` Christoffer Dall 2018-12-10 10:26 ` Suzuki K Poulose 2018-12-10 10:26 ` Suzuki K Poulose 2018-12-10 10:47 ` Suzuki K Poulose 2018-12-10 10:47 ` Suzuki K Poulose 2018-12-10 11:01 ` Christoffer Dall 2018-12-10 11:01 ` Christoffer Dall 2018-10-31 17:57 ` [PATCH v9 2/8] KVM: arm/arm64: Re-factor setting the Stage 2 entry to exec on fault Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-12-03 13:32 ` Anshuman Khandual 2018-12-03 13:32 ` Anshuman Khandual 2018-12-05 10:47 ` Suzuki K Poulose 2018-12-05 10:47 ` Suzuki K Poulose 2018-12-10 9:00 ` Christoffer Dall 2018-12-10 9:00 ` Christoffer Dall 2018-12-10 8:59 ` Christoffer Dall 2018-12-10 8:59 ` Christoffer Dall 2018-10-31 17:57 ` [PATCH v9 3/8] KVM: arm/arm64: Introduce helpers to manipulate page table entries Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-12-03 13:50 ` Anshuman Khandual 2018-12-03 13:50 ` Anshuman Khandual 2018-12-03 14:03 ` Suzuki K Poulose 2018-12-03 14:03 ` Suzuki K Poulose 2018-12-10 9:01 ` Christoffer Dall 2018-12-10 9:01 ` Christoffer Dall 2018-10-31 17:57 ` Punit Agrawal [this message] 2018-10-31 17:57 ` [PATCH v9 4/8] KVM: arm64: Support dirty page tracking for PUD hugepages Punit Agrawal 2018-12-03 14:17 ` Anshuman Khandual 2018-12-03 14:17 ` Anshuman Khandual 2018-12-03 14:21 ` Suzuki K Poulose 2018-12-03 14:21 ` Suzuki K Poulose 2018-10-31 17:57 ` [PATCH v9 5/8] KVM: arm64: Support PUD hugepage in stage2_is_exec() Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-11-01 13:38 ` Christoffer Dall 2018-11-01 13:38 ` Christoffer Dall 2018-12-05 17:57 ` Suzuki K Poulose 2018-12-05 17:57 ` Suzuki K Poulose 2018-12-10 9:06 ` Christoffer Dall 2018-12-10 9:06 ` Christoffer Dall 2018-12-03 14:37 ` Anshuman Khandual 2018-12-03 14:37 ` Anshuman Khandual 2018-10-31 17:57 ` [PATCH v9 6/8] KVM: arm64: Support handling access faults for PUD hugepages Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-11-01 13:40 ` Christoffer Dall 2018-11-01 13:40 ` Christoffer Dall 2018-11-01 13:40 ` Christoffer Dall 2018-12-03 15:10 ` Anshuman Khandual 2018-12-03 15:10 ` Anshuman Khandual 2018-10-31 17:57 ` [PATCH v9 7/8] KVM: arm64: Update age handlers to support " Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-12-03 15:19 ` Anshuman Khandual 2018-12-03 15:19 ` Anshuman Khandual 2018-10-31 17:57 ` [PATCH v9 8/8] KVM: arm64: Add support for creating PUD hugepages at stage 2 Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-10-31 17:57 ` Punit Agrawal 2018-12-03 15:46 ` Anshuman Khandual 2018-12-03 15:46 ` Anshuman Khandual
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20181031175745.18650-5-punit.agrawal@arm.com \ --to=punit.agrawal@arm.com \ --cc=catalin.marinas@arm.com \ --cc=kvmarm@lists.cs.columbia.edu \ --cc=linux-arm-kernel@lists.infradead.org \ --cc=linux-kernel@vger.kernel.org \ --cc=linux@armlinux.org.uk \ --cc=marc.zyngier@arm.com \ --cc=punitagrawal@gmail.com \ --cc=suzuki.poulose@arm.com \ --cc=will.deacon@arm.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.