From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from lists.ozlabs.org (lists.ozlabs.org [112.213.38.117]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id 25EB3C43334 for ; Fri, 24 Jun 2022 05:24:13 +0000 (UTC) Received: from boromir.ozlabs.org (localhost [IPv6:::1]) by lists.ozlabs.org (Postfix) with ESMTP id 4LTlsm4GRSz3fG6 for ; Fri, 24 Jun 2022 15:24:12 +1000 (AEST) Authentication-Results: lists.ozlabs.org; spf=pass (sender SPF authorized) smtp.mailfrom=arm.com (client-ip=217.140.110.172; helo=foss.arm.com; envelope-from=anshuman.khandual@arm.com; receiver=) Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by lists.ozlabs.org (Postfix) with ESMTP id 4LTlsJ3XlCz2yj3 for ; Fri, 24 Jun 2022 15:23:47 +1000 (AEST) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id BE747176B; Thu, 23 Jun 2022 22:23:16 -0700 (PDT) Received: from [10.162.41.7] (unknown [10.162.41.7]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 77AE93F66F; Thu, 23 Jun 2022 22:23:08 -0700 (PDT) Message-ID: <4e8b3213-2ca8-0bff-6ea9-3b03d1d1ad27@arm.com> Date: Fri, 24 Jun 2022 10:53:05 +0530 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:91.0) Gecko/20100101 Thunderbird/91.9.1 Subject: Re: [PATCH V4 03/26] powerpc/mm: Move protection_map[] inside the platform Content-Language: en-US To: Christophe Leroy , "linux-mm@kvack.org" References: <20220624044339.1533882-1-anshuman.khandual@arm.com> <20220624044339.1533882-4-anshuman.khandual@arm.com> From: Anshuman Khandual In-Reply-To: Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-BeenThere: linuxppc-dev@lists.ozlabs.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: Linux on PowerPC Developers Mail List List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: "linux-ia64@vger.kernel.org" , "linux-sh@vger.kernel.org" , "linux-kernel@vger.kernel.org" , Paul Mackerras , "sparclinux@vger.kernel.org" , "linux-riscv@lists.infradead.org" , "linux-s390@vger.kernel.org" , "linux-hexagon@vger.kernel.org" , "x86@kernel.org" , "linux-csky@vger.kernel.org" , "hch@infradead.org" , "linux-snps-arc@lists.infradead.org" , "linux-xtensa@linux-xtensa.org" , "linux-alpha@vger.kernel.org" , "linux-um@lists.infradead.org" , Nicholas Piggin , "linux-m68k@lists.linux-m68k.org" , "openrisc@lists.librecore s.org" , "linux-arm-kernel@lists.infradead.org" , "linux-parisc@vger.kernel.org" , "linux-mips@vger.kernel.org" , Andrew Morton , "linuxppc-dev@lists.ozlabs.org" Errors-To: linuxppc-dev-bounces+linuxppc-dev=archiver.kernel.org@lists.ozlabs.org Sender: "Linuxppc-dev" On 6/24/22 10:48, Christophe Leroy wrote: > > > Le 24/06/2022 à 06:43, Anshuman Khandual a écrit : >> This moves protection_map[] inside the platform and while here, also enable >> ARCH_HAS_VM_GET_PAGE_PROT on 32 bit platforms via DECLARE_VM_GET_PAGE_PROT. > > Not only 32 bit platforms, also nohash 64 (aka book3e/64) Sure, will update the commit message. > >> >> Cc: Michael Ellerman >> Cc: Paul Mackerras >> Cc: Nicholas Piggin >> Cc: linuxppc-dev@lists.ozlabs.org >> Cc: linux-kernel@vger.kernel.org >> Signed-off-by: Anshuman Khandual >> --- >> arch/powerpc/Kconfig | 2 +- >> arch/powerpc/include/asm/pgtable.h | 20 +------------------- >> arch/powerpc/mm/pgtable.c | 24 ++++++++++++++++++++++++ >> 3 files changed, 26 insertions(+), 20 deletions(-) >> >> diff --git a/arch/powerpc/Kconfig b/arch/powerpc/Kconfig >> index c2ce2e60c8f0..1035d172c7dd 100644 >> --- a/arch/powerpc/Kconfig >> +++ b/arch/powerpc/Kconfig >> @@ -140,7 +140,7 @@ config PPC >> select ARCH_HAS_TICK_BROADCAST if GENERIC_CLOCKEVENTS_BROADCAST >> select ARCH_HAS_UACCESS_FLUSHCACHE >> select ARCH_HAS_UBSAN_SANITIZE_ALL >> - select ARCH_HAS_VM_GET_PAGE_PROT if PPC_BOOK3S_64 >> + select ARCH_HAS_VM_GET_PAGE_PROT >> select ARCH_HAVE_NMI_SAFE_CMPXCHG >> select ARCH_KEEP_MEMBLOCK >> select ARCH_MIGHT_HAVE_PC_PARPORT >> diff --git a/arch/powerpc/include/asm/pgtable.h b/arch/powerpc/include/asm/pgtable.h >> index d564d0ecd4cd..bf98db844579 100644 >> --- a/arch/powerpc/include/asm/pgtable.h >> +++ b/arch/powerpc/include/asm/pgtable.h >> @@ -20,25 +20,6 @@ struct mm_struct; >> #include >> #endif /* !CONFIG_PPC_BOOK3S */ >> >> -/* Note due to the way vm flags are laid out, the bits are XWR */ >> -#define __P000 PAGE_NONE >> -#define __P001 PAGE_READONLY >> -#define __P010 PAGE_COPY >> -#define __P011 PAGE_COPY >> -#define __P100 PAGE_READONLY_X >> -#define __P101 PAGE_READONLY_X >> -#define __P110 PAGE_COPY_X >> -#define __P111 PAGE_COPY_X >> - >> -#define __S000 PAGE_NONE >> -#define __S001 PAGE_READONLY >> -#define __S010 PAGE_SHARED >> -#define __S011 PAGE_SHARED >> -#define __S100 PAGE_READONLY_X >> -#define __S101 PAGE_READONLY_X >> -#define __S110 PAGE_SHARED_X >> -#define __S111 PAGE_SHARED_X >> - >> #ifndef __ASSEMBLY__ >> >> #ifndef MAX_PTRS_PER_PGD >> @@ -79,6 +60,7 @@ extern void paging_init(void); >> void poking_init(void); >> >> extern unsigned long ioremap_bot; >> +extern pgprot_t protection_map[16] __ro_after_init; >> >> /* >> * kern_addr_valid is intended to indicate whether an address is a valid >> diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c >> index e6166b71d36d..618f30d35b17 100644 >> --- a/arch/powerpc/mm/pgtable.c >> +++ b/arch/powerpc/mm/pgtable.c >> @@ -472,3 +472,27 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea, >> return ret_pte; >> } >> EXPORT_SYMBOL_GPL(__find_linux_pte); >> + >> +/* Note due to the way vm flags are laid out, the bits are XWR */ >> +pgprot_t protection_map[16] __ro_after_init = { > > I can't see any place where protection_map[] gets modified. This could > be made const. Sure, will make it a const as in case for many other platforms as well. > >> + [VM_NONE] = PAGE_NONE, >> + [VM_READ] = PAGE_READONLY, >> + [VM_WRITE] = PAGE_COPY, >> + [VM_WRITE | VM_READ] = PAGE_COPY, >> + [VM_EXEC] = PAGE_READONLY_X, >> + [VM_EXEC | VM_READ] = PAGE_READONLY_X, >> + [VM_EXEC | VM_WRITE] = PAGE_COPY_X, >> + [VM_EXEC | VM_WRITE | VM_READ] = PAGE_COPY_X, >> + [VM_SHARED] = PAGE_NONE, >> + [VM_SHARED | VM_READ] = PAGE_READONLY, >> + [VM_SHARED | VM_WRITE] = PAGE_SHARED, >> + [VM_SHARED | VM_WRITE | VM_READ] = PAGE_SHARED, >> + [VM_SHARED | VM_EXEC] = PAGE_READONLY_X, >> + [VM_SHARED | VM_EXEC | VM_READ] = PAGE_READONLY_X, >> + [VM_SHARED | VM_EXEC | VM_WRITE] = PAGE_SHARED_X, >> + [VM_SHARED | VM_EXEC | VM_WRITE | VM_READ] = PAGE_SHARED_X >> +}; >> + >> +#ifndef CONFIG_PPC_BOOK3S_64 >> +DECLARE_VM_GET_PAGE_PROT >> +#endif