From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.7 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 63209C4361B for ; Tue, 15 Dec 2020 03:30:22 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 31C4F2246B for ; Tue, 15 Dec 2020 03:30:22 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726340AbgLODI3 (ORCPT ); Mon, 14 Dec 2020 22:08:29 -0500 Received: from mail.kernel.org ([198.145.29.99]:36124 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726310AbgLODIS (ORCPT ); Mon, 14 Dec 2020 22:08:18 -0500 Date: Mon, 14 Dec 2020 19:07:35 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=linux-foundation.org; s=korg; t=1608001657; bh=mgRRasolaYw/41v/+vVWUQXRy2eSjUl2768FMLL3tVI=; h=From:To:Subject:In-Reply-To:From; b=0rJBtD4C1/7OsWk+qyAAmFlPp1IMSyIf0Cc9PUYsmvD09ybDhxd+za2V6j7Puixi7 au0va4aCS83kRBiJ/Q+SjkK2zfrzUQ6i5vFwR1mAohox/bWmNNhtxW2dOIWA5SI1y8 dtUXcEvGDU6dynLN7gmmztnQjoRhAia+STQhO0B8= From: Andrew Morton To: akpm@linux-foundation.org, almasrymina@google.com, aneesh.kumar@linux.ibm.com, anshuman.khandual@arm.com, arnd@arndb.de, bgeffon@google.com, bp@alien8.de, catalin.marinas@arm.com, christian.brauner@ubuntu.com, dave.hansen@intel.com, frederic@kernel.org, gshan@redhat.com, hnaveed@wavecomp.com, hpa@zytor.com, jhubbard@nvidia.com, justin.he@arm.com, kaleshsingh@google.com, keescook@chromium.org, kirill.shutemov@linux.intel.com, krzk@kernel.org, linux-mm@kvack.org, linuxram@us.ibm.com, lokeshgidra@google.com, mark.rutland@arm.com, masahiroy@kernel.org, mhiramat@kernel.org, minchan@google.com, mingo@redhat.com, mm-commits@vger.kernel.org, peterz@infradead.org, rcampbell@nvidia.com, rppt@kernel.org, samitolvanen@google.com, sandipan@linux.ibm.com, shuah@kernel.org, sjpark@amazon.de, steven.price@arm.com, surenb@google.com, tglx@linutronix.de, torvalds@linux-foundation.org, will@kernel.org, ziy@nvidia.com Subject: [patch 076/200] arm64: mremap speedup - enable HAVE_MOVE_PUD Message-ID: <20201215030735.fiUldIRrJ%akpm@linux-foundation.org> In-Reply-To: <20201214190237.a17b70ae14f129e2dca3d204@linux-foundation.org> User-Agent: s-nail v14.8.16 Precedence: bulk Reply-To: linux-kernel@vger.kernel.org List-ID: X-Mailing-List: mm-commits@vger.kernel.org From: Kalesh Singh Subject: arm64: mremap speedup - enable HAVE_MOVE_PUD HAVE_MOVE_PUD enables remapping pages at the PUD level if both the source and destination addresses are PUD-aligned. With HAVE_MOVE_PUD enabled it can be inferred that there is approximately a 19x improvement in performance on arm64. (See data below). ------- Test Results --------- The following results were obtained using a 5.4 kernel, by remapping a PUD-aligned, 1GB sized region to a PUD-aligned destination. The results from 10 iterations of the test are given below: Total mremap times for 1GB data on arm64. All times are in nanoseconds. Control HAVE_MOVE_PUD 1247761 74271 1219896 46771 1094792 59687 1227760 48385 1043698 76666 1101771 50365 1159896 52500 1143594 75261 1025833 61354 1078125 48697 1134312.6 59395.7 <-- Mean time in nanoseconds A 1GB mremap completion time drops from ~1.1 milliseconds to ~59 microseconds on arm64. (~19x speed up). Link: https://lkml.kernel.org/r/20201014005320.2233162-5-kaleshsingh@google.com Signed-off-by: Kalesh Singh Acked-by: Kirill A. Shutemov Cc: Catalin Marinas Cc: Will Deacon Cc: Aneesh Kumar K.V Cc: Anshuman Khandual Cc: Arnd Bergmann Cc: Borislav Petkov Cc: Brian Geffon Cc: Christian Brauner Cc: Dave Hansen Cc: Frederic Weisbecker Cc: Gavin Shan Cc: Hassan Naveed Cc: "H. Peter Anvin" Cc: Ingo Molnar Cc: Jia He Cc: John Hubbard Cc: Kees Cook Cc: Krzysztof Kozlowski Cc: Lokesh Gidra Cc: Mark Rutland Cc: Masahiro Yamada Cc: Masami Hiramatsu Cc: Mike Rapoport Cc: Mina Almasry Cc: Minchan Kim Cc: Peter Zijlstra (Intel) Cc: Ralph Campbell Cc: Ram Pai Cc: Sami Tolvanen Cc: Sandipan Das Cc: SeongJae Park Cc: Shuah Khan Cc: Steven Price Cc: Suren Baghdasaryan Cc: Thomas Gleixner Cc: Zi Yan Signed-off-by: Andrew Morton --- arch/arm64/Kconfig | 1 + arch/arm64/include/asm/pgtable.h | 1 + 2 files changed, 2 insertions(+) --- a/arch/arm64/include/asm/pgtable.h~arm64-mremap-speedup-enable-have_move_pud +++ a/arch/arm64/include/asm/pgtable.h @@ -462,6 +462,7 @@ static inline pmd_t pmd_mkdevmap(pmd_t p #define pfn_pud(pfn,prot) __pud(__phys_to_pud_val((phys_addr_t)(pfn) << PAGE_SHIFT) | pgprot_val(prot)) #define set_pmd_at(mm, addr, pmdp, pmd) set_pte_at(mm, addr, (pte_t *)pmdp, pmd_pte(pmd)) +#define set_pud_at(mm, addr, pudp, pud) set_pte_at(mm, addr, (pte_t *)pudp, pud_pte(pud)) #define __p4d_to_phys(p4d) __pte_to_phys(p4d_pte(p4d)) #define __phys_to_p4d_val(phys) __phys_to_pte_val(phys) --- a/arch/arm64/Kconfig~arm64-mremap-speedup-enable-have_move_pud +++ a/arch/arm64/Kconfig @@ -125,6 +125,7 @@ config ARM64 select HANDLE_DOMAIN_IRQ select HARDIRQS_SW_RESEND select HAVE_MOVE_PMD + select HAVE_MOVE_PUD select HAVE_PCI select HAVE_ACPI_APEI if (ACPI && EFI) select HAVE_ALIGNED_STRUCT_PAGE if SLUB _