From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.3 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id B9973C54FCC for ; Tue, 21 Apr 2020 14:19:17 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 745D420663 for ; Tue, 21 Apr 2020 14:19:17 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 745D420663 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=huawei.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 12EA68E0005; Tue, 21 Apr 2020 10:19:17 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 0DF968E0003; Tue, 21 Apr 2020 10:19:17 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 01BF88E0005; Tue, 21 Apr 2020 10:19:16 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0203.hostedemail.com [216.40.44.203]) by kanga.kvack.org (Postfix) with ESMTP id DB34B8E0003 for ; Tue, 21 Apr 2020 10:19:16 -0400 (EDT) Received: from smtpin03.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay03.hostedemail.com (Postfix) with ESMTP id 9BCCC82499A8 for ; Tue, 21 Apr 2020 14:19:16 +0000 (UTC) X-FDA: 76732069512.03.pies27_6df2f603a443b X-HE-Tag: pies27_6df2f603a443b X-Filterd-Recvd-Size: 7090 Received: from huawei.com (szxga05-in.huawei.com [45.249.212.191]) by imf12.hostedemail.com (Postfix) with ESMTP for ; Tue, 21 Apr 2020 14:19:15 +0000 (UTC) Received: from DGGEMS407-HUB.china.huawei.com (unknown [172.30.72.59]) by Forcepoint Email with ESMTP id 37F28F4CD612CA17545A; Tue, 21 Apr 2020 22:19:09 +0800 (CST) Received: from [127.0.0.1] (10.173.220.25) by DGGEMS407-HUB.china.huawei.com (10.3.19.207) with Microsoft SMTP Server id 14.3.487.0; Tue, 21 Apr 2020 22:18:59 +0800 Subject: Re: [PATCH v1 5/6] mm: tlb: Provide flush_*_tlb_range wrappers To: Peter Zijlstra CC: , , , , , , , , , , , , , , , , , , , , , , , , References: <20200403090048.938-1-yezhenyu2@huawei.com> <20200403090048.938-6-yezhenyu2@huawei.com> <20200420120916.GE20696@hirez.programming.kicks-ass.net> From: Zhenyu Ye Message-ID: <8f6e877c-1462-441b-23ec-2a6bd3308099@huawei.com> Date: Tue, 21 Apr 2020 22:18:57 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Thunderbird/68.3.0 MIME-Version: 1.0 In-Reply-To: <20200420120916.GE20696@hirez.programming.kicks-ass.net> Content-Type: text/plain; charset="gbk" Content-Transfer-Encoding: 7bit X-Originating-IP: [10.173.220.25] X-CFilter-Loop: Reflected X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Hi Peter, On 2020/4/20 20:09, Peter Zijlstra wrote: > On Fri, Apr 03, 2020 at 05:00:47PM +0800, Zhenyu Ye wrote: >> This patch provides flush_{pte|pmd|pud|p4d}_tlb_range() in generic >> code, which are expressed through the mmu_gather APIs. These >> interface set tlb->cleared_* and finally call tlb_flush(), so we >> can do the tlb invalidation according to the information in >> struct mmu_gather. >> >> Signed-off-by: Zhenyu Ye >> --- >> include/asm-generic/pgtable.h | 12 +++++++-- >> mm/pgtable-generic.c | 50 +++++++++++++++++++++++++++++++++++ >> 2 files changed, 60 insertions(+), 2 deletions(-) >> >> diff --git a/include/asm-generic/pgtable.h b/include/asm-generic/pgtable.h >> index e2e2bef07dd2..2bedeee94131 100644 >> --- a/include/asm-generic/pgtable.h >> +++ b/include/asm-generic/pgtable.h >> @@ -1160,11 +1160,19 @@ static inline int pmd_free_pte_page(pmd_t *pmd, unsigned long addr) >> * invalidate the entire TLB which is not desitable. >> * e.g. see arch/arc: flush_pmd_tlb_range >> */ >> -#define flush_pmd_tlb_range(vma, addr, end) flush_tlb_range(vma, addr, end) >> -#define flush_pud_tlb_range(vma, addr, end) flush_tlb_range(vma, addr, end) >> +extern void flush_pte_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end); >> +extern void flush_pmd_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end); >> +extern void flush_pud_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end); >> +extern void flush_p4d_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end); >> #else >> +#define flush_pte_tlb_range(vma, addr, end) BUILD_BUG() >> #define flush_pmd_tlb_range(vma, addr, end) BUILD_BUG() >> #define flush_pud_tlb_range(vma, addr, end) BUILD_BUG() >> +#define flush_p4d_tlb_range(vma, addr, end) BUILD_BUG() >> #endif >> #endif > > Ideally you'd make __HAVE_ARCH_FLUSH_PMD_TLB_RANGE go away. Power > certainly doesnt need it with the below. > However, arch `arc` also uses __HAVE_ARCH_FLUSH_PMD_TLB_RANGE : grep -nr __HAVE_ARCH_FLUSH_PMD_TLB_RANGE mm/pgtable-generic.c:104:#ifndef __HAVE_ARCH_FLUSH_PMD_TLB_RANGE mm/pgtable-generic.c:152:#endif /* __HAVE_ARCH_FLUSH_PMD_TLB_RANGE */ include/asm-generic/pgtable.h:1153:#ifndef __HAVE_ARCH_FLUSH_PMD_TLB_RANGE arch/powerpc/include/asm/book3s/64/tlbflush.h:49:#define __HAVE_ARCH_FLUSH_PMD_TLB_RANGE arch/arc/include/asm/hugepage.h:69:#define __HAVE_ARCH_FLUSH_PMD_TLB_RANGE So I am not sure if we can remove it. And if we remove the __HAVE_ARCH_FLUSH_PMD_TLB_RANGE, how to ensure not redefine flush_pXX_tlb_range() ? >> diff --git a/mm/pgtable-generic.c b/mm/pgtable-generic.c >> index 3d7c01e76efc..0f5414a4a2ec 100644 >> --- a/mm/pgtable-generic.c >> +++ b/mm/pgtable-generic.c >> @@ -101,6 +101,56 @@ pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, >> >> #ifdef CONFIG_TRANSPARENT_HUGEPAGE >> >> +#ifndef __HAVE_ARCH_FLUSH_PMD_TLB_RANGE >> +void flush_pte_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end) >> +{ >> + struct mmu_gather tlb; >> + >> + tlb_gather_mmu(&tlb, vma->vm_mm, addr, end); >> + tlb_start_vma(&tlb, vma); >> + tlb_set_pte_range(&tlb, addr, end - addr); >> + tlb_end_vma(&tlb, vma); >> + tlb_finish_mmu(&tlb, addr, end); >> +} >> + >> +void flush_pmd_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end) >> +{ >> + struct mmu_gather tlb; >> + >> + tlb_gather_mmu(&tlb, vma->vm_mm, addr, end); >> + tlb_start_vma(&tlb, vma); >> + tlb_set_pmd_range(&tlb, addr, end - addr); >> + tlb_end_vma(&tlb, vma); >> + tlb_finish_mmu(&tlb, addr, end); >> +} >> + >> +void flush_pud_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end) >> +{ >> + struct mmu_gather tlb; >> + >> + tlb_gather_mmu(&tlb, vma->vm_mm, addr, end); >> + tlb_start_vma(&tlb, vma); >> + tlb_set_pud_range(&tlb, addr, end - addr); >> + tlb_end_vma(&tlb, vma); >> + tlb_finish_mmu(&tlb, addr, end); >> +} >> + >> +void flush_p4d_tlb_range(struct vm_area_struct *vma, >> + unsigned long addr, unsigned long end) >> +{ >> + struct mmu_gather tlb; >> + >> + tlb_gather_mmu(&tlb, vma->vm_mm, addr, end); >> + tlb_start_vma(&tlb, vma); >> + tlb_set_p4d_range(&tlb, addr, end - addr); >> + tlb_end_vma(&tlb, vma); >> + tlb_finish_mmu(&tlb, addr, end); >> +} >> +#endif /* __HAVE_ARCH_FLUSH_PMD_TLB_RANGE */ > > You're nowhere near lazy enough: > > #define FLUSH_Pxx_TLB_RANGE(_pxx) \ > void flush_##_pxx##_tlb_range(struct vm_area_struct *vma, \ > unsigned long addr, unsigned long end) \ > { \ > struct mmu_gather tlb; \ > \ > tlb_gather_mmu(&tlb, vma->vm_mm, addr, end); \ > tlb_start_vma(&tlb, vma); \ > tlb_flush_##_pxx##_range(&tlb, addr, end-addr); \ > tlb_end_vma(&tlb, vma); \ > tlb_finish_mmu(&tlb, addr, end); \ > } > > FLUSH_Pxx_TLB_RANGE(pte) > FLUSH_Pxx_TLB_RANGE(pmd) > FLUSH_Pxx_TLB_RANGE(pud) > FLUSH_Pxx_TLB_RANGE(p4d) > > > . >