From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.7 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE, SPF_PASS,URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E57B3C5519F for ; Tue, 17 Nov 2020 10:50:40 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 7382622447 for ; Tue, 17 Nov 2020 10:50:40 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=bytedance-com.20150623.gappssmtp.com header.i=@bytedance-com.20150623.gappssmtp.com header.b="HJYzn76F" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727882AbgKQKuT (ORCPT ); Tue, 17 Nov 2020 05:50:19 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:36158 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726310AbgKQKuS (ORCPT ); Tue, 17 Nov 2020 05:50:18 -0500 Received: from mail-pj1-x1036.google.com (mail-pj1-x1036.google.com [IPv6:2607:f8b0:4864:20::1036]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id C7FF0C0617A6 for ; Tue, 17 Nov 2020 02:50:18 -0800 (PST) Received: by mail-pj1-x1036.google.com with SMTP id f12so237151pjp.4 for ; Tue, 17 Nov 2020 02:50:18 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=bytedance-com.20150623.gappssmtp.com; s=20150623; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=mjUc1dVyS4RxJ3DlTkh1mxQ2bUOoW5hWXb6UapOFcG4=; b=HJYzn76Ffe692t55Wxr8B2H6SMK8BhzgrnTcrtgEb6ddIABw7D5Z/6ZK0MpgWPERFh zvyugk7rIXsxodADPNjxM/Y5GzZGVkUMwo8qpymcvRvcJGFAjhkicm4inTZ846wM9R8o iR82euIO0RrXqNHZe6lbis/8rpXQF2Murs4Q49K0znjitABu/BToSHL+/6ZKBq/FkYzp cM/VlKcaoUyQyW0Ta2jwlSKTToLobnPr0PrBZlgRFHwSg2avpwDBdMr4awslTTN9GuMv MNiYwU1jvUykPQtdOg1nf37VkGZPbLCbyGABRncmt5OQcRerI8zdbdJe+CZ8POcnj5uk yx0A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=mjUc1dVyS4RxJ3DlTkh1mxQ2bUOoW5hWXb6UapOFcG4=; b=bGHUKx76QDHBjq3sUF/I2RGgUH+x2FkbMbavn2AVLz8UoEagobkF2k/DDLkd3rawKN b8ezuECnRA6jKzBO/iSn2CfUZ4QHRge9hYK/Q0YO9fENBgiAyFsFUeAERniRvOBkC9XK pYvaqM2nIULWRPa0vHbwkyJ9ZgihfohW3pDKhGOQEDZMhhDNhp2Kd/joU6ocFB0PDrH4 QEK79grlPVmcETKz4UXfn8cb8MuCzJyAGwHedXYu1+YaJmYof81/rMdJStXcebQuscgW JuYSU020aKeuYE69kStguX3kuaFEV3uqjrkZLSY1zQQqQAUnCGcELLVa046gaJm8jTGD sFvQ== X-Gm-Message-State: AOAM530+Ua6t96hHWgxjZ1XgYqiw3U63zHibBYrWlVQ7K5Gumjxt8lmm W7N4owDwMlI3XrLQxvgqtYztYVN788feoDIxPr72yg== X-Google-Smtp-Source: ABdhPJyeuObIhGqno+9CkTgaJp3e6ToPI9JCgIWD55SynfomA6oDrbXK04ts8Y748KqNvY8eoRlcEA8+daxR70CrwKU= X-Received: by 2002:a17:902:c14b:b029:d6:ab18:108d with SMTP id 11-20020a170902c14bb02900d6ab18108dmr17511270plj.20.1605610218141; Tue, 17 Nov 2020 02:50:18 -0800 (PST) MIME-Version: 1.0 References: <20201113105952.11638-1-songmuchun@bytedance.com> <349168819c1249d4bceea26597760b0a@hisilicon.com> In-Reply-To: <349168819c1249d4bceea26597760b0a@hisilicon.com> From: Muchun Song Date: Tue, 17 Nov 2020 18:49:39 +0800 Message-ID: Subject: Re: [External] RE: [PATCH v4 00/21] Free some vmemmap pages of hugetlb page To: "Song Bao Hua (Barry Song)" Cc: "corbet@lwn.net" , "mike.kravetz@oracle.com" , "tglx@linutronix.de" , "mingo@redhat.com" , "bp@alien8.de" , "x86@kernel.org" , "hpa@zytor.com" , "dave.hansen@linux.intel.com" , "luto@kernel.org" , "peterz@infradead.org" , "viro@zeniv.linux.org.uk" , "akpm@linux-foundation.org" , "paulmck@kernel.org" , "mchehab+huawei@kernel.org" , "pawan.kumar.gupta@linux.intel.com" , "rdunlap@infradead.org" , "oneukum@suse.com" , "anshuman.khandual@arm.com" , "jroedel@suse.de" , "almasrymina@google.com" , "rientjes@google.com" , "willy@infradead.org" , "osalvador@suse.de" , "mhocko@suse.com" , "duanxiongchun@bytedance.com" , "linux-doc@vger.kernel.org" , "linux-kernel@vger.kernel.org" , "linux-mm@kvack.org" , "linux-fsdevel@vger.kernel.org" Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Nov 17, 2020 at 6:16 PM Song Bao Hua (Barry Song) wrote: > > > > > -----Original Message----- > > From: owner-linux-mm@kvack.org [mailto:owner-linux-mm@kvack.org] On > > Behalf Of Muchun Song > > Sent: Saturday, November 14, 2020 12:00 AM > > To: corbet@lwn.net; mike.kravetz@oracle.com; tglx@linutronix.de; > > mingo@redhat.com; bp@alien8.de; x86@kernel.org; hpa@zytor.com; > > dave.hansen@linux.intel.com; luto@kernel.org; peterz@infradead.org; > > viro@zeniv.linux.org.uk; akpm@linux-foundation.org; paulmck@kernel.org; > > mchehab+huawei@kernel.org; pawan.kumar.gupta@linux.intel.com; > > rdunlap@infradead.org; oneukum@suse.com; anshuman.khandual@arm.com; > > jroedel@suse.de; almasrymina@google.com; rientjes@google.com; > > willy@infradead.org; osalvador@suse.de; mhocko@suse.com > > Cc: duanxiongchun@bytedance.com; linux-doc@vger.kernel.org; > > linux-kernel@vger.kernel.org; linux-mm@kvack.org; > > linux-fsdevel@vger.kernel.org; Muchun Song > > Subject: [PATCH v4 00/21] Free some vmemmap pages of hugetlb page > > > > Hi all, > > > > This patch series will free some vmemmap pages(struct page structures) > > associated with each hugetlbpage when preallocated to save memory. > > > > Nowadays we track the status of physical page frames using struct page > > structures arranged in one or more arrays. And here exists one-to-one > > mapping between the physical page frame and the corresponding struct page > > structure. > > > > The HugeTLB support is built on top of multiple page size support that > > is provided by most modern architectures. For example, x86 CPUs normally > > support 4K and 2M (1G if architecturally supported) page sizes. Every > > HugeTLB has more than one struct page structure. The 2M HugeTLB has 512 > > struct page structure and 1G HugeTLB has 4096 struct page structures. But > > in the core of HugeTLB only uses the first 4 (Use of first 4 struct page > > structures comes from HUGETLB_CGROUP_MIN_ORDER.) struct page > > structures to > > store metadata associated with each HugeTLB. The rest of the struct page > > structures are usually read the compound_head field which are all the same > > value. If we can free some struct page memory to buddy system so that we > > can save a lot of memory. > > > > When the system boot up, every 2M HugeTLB has 512 struct page structures > > which size is 8 pages(sizeof(struct page) * 512 / PAGE_SIZE). > > > > hugetlbpage struct pages(8 pages) page > > frame(8 pages) > > +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+ > > | | | 0 | -------------> | 0 > > | > > | | | 1 | -------------> | 1 > > | > > | | | 2 | -------------> | 2 > > | > > | | | 3 | -------------> | 3 > > | > > | | | 4 | -------------> | 4 > > | > > | 2M | | 5 | -------------> | > > 5 | > > | | | 6 | -------------> | 6 > > | > > | | | 7 | -------------> | 7 > > | > > | | +-----------+ > > +-----------+ > > | | > > | | > > +-----------+ > > > > > > When a hugetlbpage is preallocated, we can change the mapping from above > > to > > bellow. > > > > hugetlbpage struct pages(8 pages) page > > frame(8 pages) > > +-----------+ ---virt_to_page---> +-----------+ mapping to +-----------+ > > | | | 0 | -------------> | 0 > > | > > | | | 1 | -------------> | 1 > > | > > | | | 2 | -------------> > > +-----------+ > > | | | 3 | -----------------^ ^ ^ ^ > > ^ > > | | | 4 | -------------------+ | | > > | > > | 2M | | 5 | ---------------------+ | > > | > > | | | 6 | -----------------------+ | > > | | | 7 | -------------------------+ > > | | +-----------+ > > | | > > | | > > +-----------+ > > > > For tail pages, the value of compound_head is the same. So we can reuse > > first page of tail page structs. We map the virtual addresses of the > > remaining 6 pages of tail page structs to the first tail page struct, > > and then free these 6 pages. Therefore, we need to reserve at least 2 > > pages as vmemmap areas. > > > > When a hugetlbpage is freed to the buddy system, we should allocate six > > pages for vmemmap pages and restore the previous mapping relationship. > > > > If we uses the 1G hugetlbpage, we can save 4088 pages(There are 4096 pages > > for > > struct page structures, we reserve 2 pages for vmemmap and 8 pages for page > > tables. So we can save 4088 pages). This is a very substantial gain. On our > > server, run some SPDK/QEMU applications which will use 1024GB hugetlbpage. > > With this feature enabled, we can save ~16GB(1G hugepage)/~11GB(2MB > > hugepage) > > Hi Muchun, > > Do we really save 11GB for 2MB hugepage? > How much do we save if we only get one 2MB hugetlb from one 128MB mem_section? > It seems we need to get at least one page for the PTEs since we are splitting PMD of > vmemmap into PTE? There are 524288(1024GB/2MB) 2MB HugeTLB pages. We can save 6 pages for each 2MB HugeTLB page. So we can save 3145728 pages. But we need to split PMD page table for every one 128MB mem_section and every section need one page as PTE page table. So we need 8192(1024GB/128MB) pages as PTE page tables. Finally, we can save 3137536(3145728-8192) pages which is 11.97GB. Thanks Barry. > > > memory. > > > > Because there are vmemmap page tables reconstruction on the > > freeing/allocating > > path, it increases some overhead. Here are some overhead analysis. > > > > 1) Allocating 10240 2MB hugetlb pages. > > > > a) With this patch series applied: > > # time echo 10240 > /proc/sys/vm/nr_hugepages > > > > real 0m0.166s > > user 0m0.000s > > sys 0m0.166s > > > > # bpftrace -e 'kprobe:alloc_fresh_huge_page { @start[tid] = nsecs; } > > kretprobe:alloc_fresh_huge_page /@start[tid]/ { @latency = hist(nsecs - > > @start[tid]); delete(@start[tid]); }' > > Attaching 2 probes... > > > > @latency: > > [8K, 16K) 8360 > > |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ > > @@@@@@@@@@@@@@@@| > > [16K, 32K) 1868 |@@@@@@@@@@@ > > | > > [32K, 64K) 10 | > > | > > [64K, 128K) 2 | > > | > > > > b) Without this patch series: > > # time echo 10240 > /proc/sys/vm/nr_hugepages > > > > real 0m0.066s > > user 0m0.000s > > sys 0m0.066s > > > > # bpftrace -e 'kprobe:alloc_fresh_huge_page { @start[tid] = nsecs; } > > kretprobe:alloc_fresh_huge_page /@start[tid]/ { @latency = hist(nsecs - > > @start[tid]); delete(@start[tid]); }' > > Attaching 2 probes... > > > > @latency: > > [4K, 8K) 10176 > > |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ > > @@@@@@@@@@@@@@@@| > > [8K, 16K) 62 | > > | > > [16K, 32K) 2 | > > | > > > > Summarize: this feature is about ~2x slower than before. > > > > 2) Freeing 10240 @MB hugetlb pages. > > > > a) With this patch series applied: > > # time echo 0 > /proc/sys/vm/nr_hugepages > > > > real 0m0.004s > > user 0m0.000s > > sys 0m0.002s > > > > Something is wrong here, it is faster than the case without this patchset: > 0.004s vs. 0m0.077s Yeah, it is faster. And Why the 'real' time of patched is smaller than before? Because in this patch series, the freeing HugeTLB is asynchronous(through worker). > > > # bpftrace -e 'kprobe:__free_hugepage { @start[tid] = nsecs; } > > kretprobe:__free_hugepage /@start[tid]/ { @latency = hist(nsecs - @start[tid]); > > delete(@start[tid]); }' > > Attaching 2 probes... > > > > @latency: > > [16K, 32K) 10240 > > |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ > > @@@@@@@@@@@@@@@@| > > > > b) Without this patch series: > > # time echo 0 > /proc/sys/vm/nr_hugepages > > > > real 0m0.077s > > user 0m0.001s > > sys 0m0.075s > > > > # bpftrace -e 'kprobe:__free_hugepage { @start[tid] = nsecs; } > > kretprobe:__free_hugepage /@start[tid]/ { @latency = hist(nsecs - @start[tid]); > > delete(@start[tid]); }' > > Attaching 2 probes... > > > > @latency: > > [4K, 8K) 9950 > > |@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@@ > > @@@@@@@@@@@@@@@@| > > [8K, 16K) 287 |@ > > | > > [16K, 32K) 3 | > > | > > > > Summarize: The overhead of __free_hugepage is about ~2-4x slower than > > before. > > But according to the allocation test above, I think that here is > > also ~2x slower than before. > > > > But why the 'real' time of patched is smaller than before? > > Because > > In this patch series, the freeing hugetlb is asynchronous(through > > kwoker). > > > > Although the overhead has increased, the overhead is not significant. Like MIke > > said, "However, remember that the majority of use cases create hugetlb pages > > at > > or shortly after boot time and add them to the pool. So, additional overhead is > > at pool creation time. There is no change to 'normal run time' operations of > > getting a page from or returning a page to the pool (think page fault/unmap)". > > > > It seems it is true. At runtime, people normally don't change hugetlb. > > > changelog in v4: > > 1. Move all the vmemmap functions to hugetlb_vmemmap.c. > > 2. Make the CONFIG_HUGETLB_PAGE_FREE_VMEMMAP default to y, if we > > want to > > disable this feature, we should disable it by a boot/kernel command line. > > 3. Remove vmemmap_pgtable_{init, deposit, withdraw}() helper functions. > > 4. Initialize page table lock for vmemmap through core_initcall mechanism. > > > > Thanks for Mike and Oscar's suggestions. > > > > changelog in v3: > > 1. Rename some helps function name. Thanks Mike. > > 2. Rework some code. Thanks Mike and Oscar. > > 3. Remap the tail vmemmap page with PAGE_KERNEL_RO instead of > > PAGE_KERNEL. Thanks Matthew. > > 4. Add some overhead analysis in the cover letter. > > 5. Use vmemap pmd table lock instead of a hugetlb specific global lock. > > > > changelog in v2: > > 1. Fix do not call dissolve_compound_page in alloc_huge_page_vmemmap(). > > 2. Fix some typo and code style problems. > > 3. Remove unused handle_vmemmap_fault(). > > 4. Merge some commits to one commit suggested by Mike. > > > > Muchun Song (21): > > mm/memory_hotplug: Move bootmem info registration API to > > bootmem_info.c > > mm/memory_hotplug: Move {get,put}_page_bootmem() to bootmem_info.c > > mm/hugetlb: Introduce a new config HUGETLB_PAGE_FREE_VMEMMAP > > mm/hugetlb: Introduce nr_free_vmemmap_pages in the struct hstate > > mm/hugetlb: Introduce pgtable allocation/freeing helpers > > mm/bootmem_info: Introduce {free,prepare}_vmemmap_page() > > mm/bootmem_info: Combine bootmem info and type into page->freelist > > mm/hugetlb: Initialize page table lock for vmemmap > > mm/hugetlb: Free the vmemmap pages associated with each hugetlb page > > mm/hugetlb: Defer freeing of hugetlb pages > > mm/hugetlb: Allocate the vmemmap pages associated with each hugetlb > > page > > mm/hugetlb: Introduce remap_huge_page_pmd_vmemmap helper > > mm/hugetlb: Use PG_slab to indicate split pmd > > mm/hugetlb: Support freeing vmemmap pages of gigantic page > > mm/hugetlb: Set the PageHWPoison to the raw error page > > mm/hugetlb: Flush work when dissolving hugetlb page > > mm/hugetlb: Add a kernel parameter hugetlb_free_vmemmap > > mm/hugetlb: Merge pte to huge pmd only for gigantic page > > mm/hugetlb: Gather discrete indexes of tail page > > mm/hugetlb: Add BUILD_BUG_ON to catch invalid usage of tail struct > > page > > mm/hugetlb: Disable freeing vmemmap if struct page size is not power > > of two > > > > Documentation/admin-guide/kernel-parameters.txt | 9 + > > Documentation/admin-guide/mm/hugetlbpage.rst | 3 + > > arch/x86/include/asm/hugetlb.h | 17 + > > arch/x86/include/asm/pgtable_64_types.h | 8 + > > arch/x86/mm/init_64.c | 7 +- > > fs/Kconfig | 14 + > > include/linux/bootmem_info.h | 78 +++ > > include/linux/hugetlb.h | 19 + > > include/linux/hugetlb_cgroup.h | 15 +- > > include/linux/memory_hotplug.h | 27 - > > mm/Makefile | 2 + > > mm/bootmem_info.c | 124 ++++ > > mm/hugetlb.c | 163 +++++- > > mm/hugetlb_vmemmap.c | 732 > > ++++++++++++++++++++++++ > > mm/hugetlb_vmemmap.h | 104 ++++ > > mm/memory_hotplug.c | 116 ---- > > mm/sparse.c | 5 +- > > 17 files changed, 1263 insertions(+), 180 deletions(-) > > create mode 100644 include/linux/bootmem_info.h > > create mode 100644 mm/bootmem_info.c > > create mode 100644 mm/hugetlb_vmemmap.c > > create mode 100644 mm/hugetlb_vmemmap.h > > > > Thanks > Barry > -- Yours, Muchun