linuxppc-dev.lists.ozlabs.org archive mirror
 help / color / mirror / Atom feed
From: Christophe Leroy <christophe.leroy@csgroup.eu>
To: Benjamin Herrenschmidt <benh@kernel.crashing.org>,
	Paul Mackerras <paulus@samba.org>,
	Michael Ellerman <mpe@ellerman.id.au>
Cc: linuxppc-dev@lists.ozlabs.org, linux-kernel@vger.kernel.org
Subject: [PATCH v3 23/45] powerpc/mm: Reduce hugepd size for 8M hugepages on 8xx
Date: Mon, 11 May 2020 11:25:49 +0000 (UTC)	[thread overview]
Message-ID: <af3ec2ee64e136cb6bffa4e657c4fa5e3202936d.1589196133.git.christophe.leroy@csgroup.eu> (raw)
In-Reply-To: <cover.1589196133.git.christophe.leroy@csgroup.eu>

Commit 55c8fc3f4930 ("powerpc/8xx: reintroduce 16K pages with HW
assistance") redefined pte_t as a struct of 4 pte_basic_t, because
in 16K pages mode there are four identical entries in the page table.
But hugepd entries for 8M pages require only one entry of size
pte_basic_t. So there is no point in creating a cache for 4 entries
page tables.

Calculate PTE_T_ORDER using the size of pte_basic_t instead of pte_t.

Define specific huge_pte helpers (set_huge_pte_at(), huge_pte_clear(),
huge_ptep_set_wrprotect()) to write the pte in a single entry instead
of using set_pte_at() which writes 4 identical entries in 16k pages
mode. Also make sure that __ptep_set_access_flags() properly handle
the huge_pte case.

Define set_pte_filter() inline otherwise GCC doesn't inline it anymore
because it is now used twice, and that gives a pretty suboptimal code
because of pte_t being a struct of 4 entries.

Those functions are also used for 512k pages which only require one
entry as well allthough replicating it four times was harmless as 512k
pages entries are spread every 128 bytes in the table.

Signed-off-by: Christophe Leroy <christophe.leroy@csgroup.eu>
---
 .../include/asm/nohash/32/hugetlb-8xx.h       | 20 ++++++++++++++
 arch/powerpc/include/asm/nohash/32/pgtable.h  |  3 ++-
 arch/powerpc/mm/hugetlbpage.c                 |  3 ++-
 arch/powerpc/mm/pgtable.c                     | 26 ++++++++++++++++---
 4 files changed, 46 insertions(+), 6 deletions(-)

diff --git a/arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h b/arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h
index a46616937d20..785437323576 100644
--- a/arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h
+++ b/arch/powerpc/include/asm/nohash/32/hugetlb-8xx.h
@@ -41,4 +41,24 @@ static inline int check_and_get_huge_psize(int shift)
 	return shift_to_mmu_psize(shift);
 }
 
+#define __HAVE_ARCH_HUGE_SET_HUGE_PTE_AT
+void set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte);
+
+#define __HAVE_ARCH_HUGE_PTE_CLEAR
+static inline void huge_pte_clear(struct mm_struct *mm, unsigned long addr,
+				  pte_t *ptep, unsigned long sz)
+{
+	pte_update(mm, addr, ptep, ~0UL, 0, 1);
+}
+
+#define __HAVE_ARCH_HUGE_PTEP_SET_WRPROTECT
+static inline void huge_ptep_set_wrprotect(struct mm_struct *mm,
+					   unsigned long addr, pte_t *ptep)
+{
+	unsigned long clr = ~pte_val(pte_wrprotect(__pte(~0)));
+	unsigned long set = pte_val(pte_wrprotect(__pte(0)));
+
+	pte_update(mm, addr, ptep, clr, set, 1);
+}
+
 #endif /* _ASM_POWERPC_NOHASH_32_HUGETLB_8XX_H */
diff --git a/arch/powerpc/include/asm/nohash/32/pgtable.h b/arch/powerpc/include/asm/nohash/32/pgtable.h
index 272963a05ab2..dd5835354e33 100644
--- a/arch/powerpc/include/asm/nohash/32/pgtable.h
+++ b/arch/powerpc/include/asm/nohash/32/pgtable.h
@@ -314,8 +314,9 @@ static inline void __ptep_set_access_flags(struct vm_area_struct *vma,
 	pte_t pte_clr = pte_mkyoung(pte_mkdirty(pte_mkwrite(pte_mkexec(__pte(~0)))));
 	unsigned long set = pte_val(entry) & pte_val(pte_set);
 	unsigned long clr = ~pte_val(entry) & ~pte_val(pte_clr);
+	int huge = psize > mmu_virtual_psize ? 1 : 0;
 
-	pte_update(vma->vm_mm, address, ptep, clr, set, 0);
+	pte_update(vma->vm_mm, address, ptep, clr, set, huge);
 
 	flush_tlb_page(vma, address);
 }
diff --git a/arch/powerpc/mm/hugetlbpage.c b/arch/powerpc/mm/hugetlbpage.c
index 33b3461d91e8..edf511c2a30a 100644
--- a/arch/powerpc/mm/hugetlbpage.c
+++ b/arch/powerpc/mm/hugetlbpage.c
@@ -30,7 +30,8 @@ bool hugetlb_disabled = false;
 
 #define hugepd_none(hpd)	(hpd_val(hpd) == 0)
 
-#define PTE_T_ORDER	(__builtin_ffs(sizeof(pte_t)) - __builtin_ffs(sizeof(void *)))
+#define PTE_T_ORDER	(__builtin_ffs(sizeof(pte_basic_t)) - \
+			 __builtin_ffs(sizeof(void *)))
 
 pte_t *huge_pte_offset(struct mm_struct *mm, unsigned long addr, unsigned long sz)
 {
diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
index e3759b69f81b..214a5f4beb6c 100644
--- a/arch/powerpc/mm/pgtable.c
+++ b/arch/powerpc/mm/pgtable.c
@@ -100,7 +100,7 @@ static pte_t set_pte_filter_hash(pte_t pte) { return pte; }
  * as we don't have two bits to spare for _PAGE_EXEC and _PAGE_HWEXEC so
  * instead we "filter out" the exec permission for non clean pages.
  */
-static pte_t set_pte_filter(pte_t pte)
+static inline pte_t set_pte_filter(pte_t pte)
 {
 	struct page *pg;
 
@@ -249,16 +249,34 @@ int huge_ptep_set_access_flags(struct vm_area_struct *vma,
 
 #else
 		/*
-		 * Not used on non book3s64 platforms. But 8xx
-		 * can possibly use tsize derived from hstate.
+		 * Not used on non book3s64 platforms.
+		 * 8xx compares it with mmu_virtual_psize to
+		 * know if it is a huge page or not.
 		 */
-		psize = 0;
+		psize = MMU_PAGE_COUNT;
 #endif
 		__ptep_set_access_flags(vma, ptep, pte, addr, psize);
 	}
 	return changed;
 #endif
 }
+
+#if defined(CONFIG_PPC_8xx)
+void set_huge_pte_at(struct mm_struct *mm, unsigned long addr, pte_t *ptep, pte_t pte)
+{
+	/*
+	 * Make sure hardware valid bit is not set. We don't do
+	 * tlb flush for this update.
+	 */
+	VM_WARN_ON(pte_hw_valid(*ptep) && !pte_protnone(*ptep));
+
+	pte = pte_mkpte(pte);
+
+	pte = set_pte_filter(pte);
+
+	ptep->pte = pte_val(pte);
+}
+#endif
 #endif /* CONFIG_HUGETLB_PAGE */
 
 #ifdef CONFIG_DEBUG_VM
-- 
2.25.0


  parent reply	other threads:[~2020-05-11 12:45 UTC|newest]

Thread overview: 46+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-05-11 11:25 [PATCH v3 00/45] Use hugepages to map kernel mem on 8xx Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 01/45] powerpc/kasan: Fix error detection on memory allocation Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 02/45] powerpc/kasan: Fix issues by lowering KASAN_SHADOW_END Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 03/45] powerpc/kasan: Fix shadow pages allocation failure Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 04/45] powerpc/kasan: Remove unnecessary page table locking Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 05/45] powerpc/kasan: Refactor update of early shadow mappings Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 06/45] powerpc/kasan: Declare kasan_init_region() weak Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 07/45] powerpc/ptdump: Limit size of flags text to 1/2 chars on PPC32 Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 08/45] powerpc/ptdump: Reorder flags Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 09/45] powerpc/ptdump: Add _PAGE_COHERENT flag Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 10/45] powerpc/ptdump: Display size of BATs Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 11/45] powerpc/ptdump: Standardise display of BAT flags Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 12/45] powerpc/ptdump: Properly handle non standard page size Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 13/45] powerpc/ptdump: Handle hugepd at PGD level Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 14/45] powerpc/32s: Don't warn when mapping RO data ROX Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 15/45] powerpc/mm: Allocate static page tables for fixmap Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 16/45] powerpc/mm: Fix conditions to perform MMU specific management by blocks on PPC32 Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 17/45] powerpc/mm: PTE_ATOMIC_UPDATES is only for 40x Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 18/45] powerpc/mm: Refactor pte_update() on nohash/32 Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 19/45] powerpc/mm: Refactor pte_update() on book3s/32 Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 20/45] powerpc/mm: Standardise __ptep_test_and_clear_young() params between PPC32 and PPC64 Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 21/45] powerpc/mm: Standardise pte_update() prototype " Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 22/45] powerpc/mm: Create a dedicated pte_update() for 8xx Christophe Leroy
2020-05-11 11:25 ` Christophe Leroy [this message]
2020-05-11 11:25 ` [PATCH v3 24/45] powerpc/8xx: Drop CONFIG_8xx_COPYBACK option Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 25/45] powerpc/8xx: Prepare handlers for _PAGE_HUGE for 512k pages Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 26/45] powerpc/8xx: Manage 512k huge pages as standard pages Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 27/45] powerpc/8xx: Only 8M pages are hugepte pages now Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 28/45] powerpc/8xx: MM_SLICE is not needed anymore Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 29/45] powerpc/8xx: Move PPC_PIN_TLB options into 8xx Kconfig Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 30/45] powerpc/8xx: Add function to set pinned TLBs Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 31/45] powerpc/8xx: Don't set IMMR map anymore at boot Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 32/45] powerpc/8xx: Always pin TLBs at startup Christophe Leroy
2020-05-11 11:25 ` [PATCH v3 33/45] powerpc/8xx: Drop special handling of Linear and IMMR mappings in I/D TLB handlers Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 34/45] powerpc/8xx: Remove now unused TLB miss functions Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 35/45] powerpc/8xx: Move DTLB perf handling closer Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 36/45] powerpc/mm: Don't be too strict with _etext alignment on PPC32 Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 37/45] powerpc/8xx: Refactor kernel address boundary comparison Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 38/45] powerpc/8xx: Add a function to early map kernel via huge pages Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 39/45] powerpc/8xx: Map IMMR with a huge page Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 40/45] powerpc/8xx: Map linear memory with huge pages Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 41/45] powerpc/8xx: Allow STRICT_KERNEL_RwX with pinned TLB Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 42/45] powerpc/8xx: Allow large TLBs with DEBUG_PAGEALLOC Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 43/45] powerpc/8xx: Implement dedicated kasan_init_region() Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 44/45] powerpc/32s: Allow mapping with BATs with DEBUG_PAGEALLOC Christophe Leroy
2020-05-11 11:26 ` [PATCH v3 45/45] powerpc/32s: Implement dedicated kasan_init_region() Christophe Leroy

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=af3ec2ee64e136cb6bffa4e657c4fa5e3202936d.1589196133.git.christophe.leroy@csgroup.eu \
    --to=christophe.leroy@csgroup.eu \
    --cc=benh@kernel.crashing.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=mpe@ellerman.id.au \
    --cc=paulus@samba.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).