From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751467AbdFZSar (ORCPT ); Mon, 26 Jun 2017 14:30:47 -0400 Received: from aserp1040.oracle.com ([141.146.126.69]:40791 "EHLO aserp1040.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751308AbdFZSal (ORCPT ); Mon, 26 Jun 2017 14:30:41 -0400 Subject: Re: [PATCH 1/3] mm, hugetlb: unclutter hugetlb allocation layers To: Michal Hocko , Andrew Morton References: <20170622193034.28972-1-mhocko@kernel.org> <20170622193034.28972-2-mhocko@kernel.org> Cc: linux-mm@kvack.org, Naoya Horiguchi , Mel Gorman , Vlastimil Babka , LKML , Michal Hocko From: Mike Kravetz Message-ID: <0b9f7cea-64b1-d20a-0b2a-87fe3d09651f@oracle.com> Date: Mon, 26 Jun 2017 11:30:25 -0700 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:45.0) Gecko/20100101 Thunderbird/45.8.0 MIME-Version: 1.0 In-Reply-To: <20170622193034.28972-2-mhocko@kernel.org> Content-Type: text/plain; charset=windows-1252 Content-Transfer-Encoding: 7bit X-Source-IP: userv0022.oracle.com [156.151.31.74] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 06/22/2017 12:30 PM, Michal Hocko wrote: > From: Michal Hocko > > Hugetlb allocation path for fresh huge pages is unnecessarily complex > and it mixes different interfaces between layers. __alloc_buddy_huge_page > is the central place to perform a new allocation. It checks for the > hugetlb overcommit and then relies on __hugetlb_alloc_buddy_huge_page to > invoke the page allocator. This is all good except that > __alloc_buddy_huge_page pushes vma and address down the callchain and > so __hugetlb_alloc_buddy_huge_page has to deal with two different > allocation modes - one for memory policy and other node specific (or to > make it more obscure node non-specific) requests. This just screams for a > reorganization. > > This patch pulls out all the vma specific handling up to > __alloc_buddy_huge_page_with_mpol where it belongs. > __alloc_buddy_huge_page will get nodemask argument and > __hugetlb_alloc_buddy_huge_page will become a trivial wrapper over the > page allocator. > > In short: > __alloc_buddy_huge_page_with_mpol - memory policy handling > __alloc_buddy_huge_page - overcommit handling and accounting > __hugetlb_alloc_buddy_huge_page - page allocator layer > > Also note that __hugetlb_alloc_buddy_huge_page and its cpuset retry loop > is not really needed because the page allocator already handles the > cpusets update. > > Finally __hugetlb_alloc_buddy_huge_page had a special case for node > specific allocations (when no policy is applied and there is a node > given). This has relied on __GFP_THISNODE to not fallback to a different > node. alloc_huge_page_node is the only caller which relies on this > behavior so move the __GFP_THISNODE there. > > Not only this removes quite some code it also should make those layers > easier to follow and clear wrt responsibilities. > > Changes since v1 > - pulled gfp mask out of __hugetlb_alloc_buddy_huge_page and make it an > explicit argument to allow __GFP_THISNODE in alloc_huge_page_node per > Vlastimil > Reviewed-by: Mike Kravetz Tested-by: Mike Kravetz -- Mike Kravetz > Acked-by: Vlastimil Babka > Signed-off-by: Michal Hocko > --- > include/linux/hugetlb.h | 2 +- > mm/hugetlb.c | 133 +++++++++++------------------------------------- > 2 files changed, 30 insertions(+), 105 deletions(-) > > diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h > index 8fd0725d3f30..66b621469f52 100644 > --- a/include/linux/hugetlb.h > +++ b/include/linux/hugetlb.h > @@ -349,7 +349,7 @@ struct page *alloc_huge_page(struct vm_area_struct *vma, > struct page *alloc_huge_page_node(struct hstate *h, int nid); > struct page *alloc_huge_page_noerr(struct vm_area_struct *vma, > unsigned long addr, int avoid_reserve); > -struct page *alloc_huge_page_nodemask(struct hstate *h, const nodemask_t *nmask); > +struct page *alloc_huge_page_nodemask(struct hstate *h, nodemask_t *nmask); > int huge_add_to_page_cache(struct page *page, struct address_space *mapping, > pgoff_t idx); > > diff --git a/mm/hugetlb.c b/mm/hugetlb.c > index 907786581812..fd6e0c50f949 100644 > --- a/mm/hugetlb.c > +++ b/mm/hugetlb.c > @@ -1521,82 +1521,19 @@ int dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn) > return rc; > } > > -/* > - * There are 3 ways this can get called: > - * 1. With vma+addr: we use the VMA's memory policy > - * 2. With !vma, but nid=NUMA_NO_NODE: We try to allocate a huge > - * page from any node, and let the buddy allocator itself figure > - * it out. > - * 3. With !vma, but nid!=NUMA_NO_NODE. We allocate a huge page > - * strictly from 'nid' > - */ > static struct page *__hugetlb_alloc_buddy_huge_page(struct hstate *h, > - struct vm_area_struct *vma, unsigned long addr, int nid) > + gfp_t gfp_mask, int nid, nodemask_t *nmask) > { > int order = huge_page_order(h); > - gfp_t gfp = htlb_alloc_mask(h)|__GFP_COMP|__GFP_REPEAT|__GFP_NOWARN; > - unsigned int cpuset_mems_cookie; > > - /* > - * We need a VMA to get a memory policy. If we do not > - * have one, we use the 'nid' argument. > - * > - * The mempolicy stuff below has some non-inlined bits > - * and calls ->vm_ops. That makes it hard to optimize at > - * compile-time, even when NUMA is off and it does > - * nothing. This helps the compiler optimize it out. > - */ > - if (!IS_ENABLED(CONFIG_NUMA) || !vma) { > - /* > - * If a specific node is requested, make sure to > - * get memory from there, but only when a node > - * is explicitly specified. > - */ > - if (nid != NUMA_NO_NODE) > - gfp |= __GFP_THISNODE; > - /* > - * Make sure to call something that can handle > - * nid=NUMA_NO_NODE > - */ > - return alloc_pages_node(nid, gfp, order); > - } > - > - /* > - * OK, so we have a VMA. Fetch the mempolicy and try to > - * allocate a huge page with it. We will only reach this > - * when CONFIG_NUMA=y. > - */ > - do { > - struct page *page; > - struct mempolicy *mpol; > - int nid; > - nodemask_t *nodemask; > - > - cpuset_mems_cookie = read_mems_allowed_begin(); > - nid = huge_node(vma, addr, gfp, &mpol, &nodemask); > - mpol_cond_put(mpol); > - page = __alloc_pages_nodemask(gfp, order, nid, nodemask); > - if (page) > - return page; > - } while (read_mems_allowed_retry(cpuset_mems_cookie)); > - > - return NULL; > + gfp_mask |= __GFP_COMP|__GFP_REPEAT|__GFP_NOWARN; > + if (nid == NUMA_NO_NODE) > + nid = numa_mem_id(); > + return __alloc_pages_nodemask(gfp_mask, order, nid, nmask); > } > > -/* > - * There are two ways to allocate a huge page: > - * 1. When you have a VMA and an address (like a fault) > - * 2. When you have no VMA (like when setting /proc/.../nr_hugepages) > - * > - * 'vma' and 'addr' are only for (1). 'nid' is always NUMA_NO_NODE in > - * this case which signifies that the allocation should be done with > - * respect for the VMA's memory policy. > - * > - * For (2), we ignore 'vma' and 'addr' and use 'nid' exclusively. This > - * implies that memory policies will not be taken in to account. > - */ > -static struct page *__alloc_buddy_huge_page(struct hstate *h, > - struct vm_area_struct *vma, unsigned long addr, int nid) > +static struct page *__alloc_buddy_huge_page(struct hstate *h, gfp_t gfp_mask, > + int nid, nodemask_t *nmask) > { > struct page *page; > unsigned int r_nid; > @@ -1605,15 +1542,6 @@ static struct page *__alloc_buddy_huge_page(struct hstate *h, > return NULL; > > /* > - * Make sure that anyone specifying 'nid' is not also specifying a VMA. > - * This makes sure the caller is picking _one_ of the modes with which > - * we can call this function, not both. > - */ > - if (vma || (addr != -1)) { > - VM_WARN_ON_ONCE(addr == -1); > - VM_WARN_ON_ONCE(nid != NUMA_NO_NODE); > - } > - /* > * Assume we will successfully allocate the surplus page to > * prevent racing processes from causing the surplus to exceed > * overcommit > @@ -1646,7 +1574,7 @@ static struct page *__alloc_buddy_huge_page(struct hstate *h, > } > spin_unlock(&hugetlb_lock); > > - page = __hugetlb_alloc_buddy_huge_page(h, vma, addr, nid); > + page = __hugetlb_alloc_buddy_huge_page(h, gfp_mask, nid, nmask); > > spin_lock(&hugetlb_lock); > if (page) { > @@ -1671,26 +1599,23 @@ static struct page *__alloc_buddy_huge_page(struct hstate *h, > } > > /* > - * Allocate a huge page from 'nid'. Note, 'nid' may be > - * NUMA_NO_NODE, which means that it may be allocated > - * anywhere. > - */ > -static > -struct page *__alloc_buddy_huge_page_no_mpol(struct hstate *h, int nid) > -{ > - unsigned long addr = -1; > - > - return __alloc_buddy_huge_page(h, NULL, addr, nid); > -} > - > -/* > * Use the VMA's mpolicy to allocate a huge page from the buddy. > */ > static > struct page *__alloc_buddy_huge_page_with_mpol(struct hstate *h, > struct vm_area_struct *vma, unsigned long addr) > { > - return __alloc_buddy_huge_page(h, vma, addr, NUMA_NO_NODE); > + struct page *page; > + struct mempolicy *mpol; > + gfp_t gfp_mask = htlb_alloc_mask(h); > + int nid; > + nodemask_t *nodemask; > + > + nid = huge_node(vma, addr, gfp_mask, &mpol, &nodemask); > + page = __alloc_buddy_huge_page(h, gfp_mask, nid, nodemask); > + mpol_cond_put(mpol); > + > + return page; > } > > /* > @@ -1700,21 +1625,26 @@ struct page *__alloc_buddy_huge_page_with_mpol(struct hstate *h, > */ > struct page *alloc_huge_page_node(struct hstate *h, int nid) > { > + gfp_t gfp_mask = htlb_alloc_mask(h); > struct page *page = NULL; > > + if (nid != NUMA_NO_NODE) > + gfp_mask |= __GFP_THISNODE; > + > spin_lock(&hugetlb_lock); > if (h->free_huge_pages - h->resv_huge_pages > 0) > page = dequeue_huge_page_node(h, nid); > spin_unlock(&hugetlb_lock); > > if (!page) > - page = __alloc_buddy_huge_page_no_mpol(h, nid); > + page = __alloc_buddy_huge_page(h, gfp_mask, nid, NULL); > > return page; > } > > -struct page *alloc_huge_page_nodemask(struct hstate *h, const nodemask_t *nmask) > +struct page *alloc_huge_page_nodemask(struct hstate *h, nodemask_t *nmask) > { > + gfp_t gfp_mask = htlb_alloc_mask(h); > struct page *page = NULL; > int node; > > @@ -1731,13 +1661,7 @@ struct page *alloc_huge_page_nodemask(struct hstate *h, const nodemask_t *nmask) > return page; > > /* No reservations, try to overcommit */ > - for_each_node_mask(node, *nmask) { > - page = __alloc_buddy_huge_page_no_mpol(h, node); > - if (page) > - return page; > - } > - > - return NULL; > + return __alloc_buddy_huge_page(h, gfp_mask, NUMA_NO_NODE, nmask); > } > > /* > @@ -1765,7 +1689,8 @@ static int gather_surplus_pages(struct hstate *h, int delta) > retry: > spin_unlock(&hugetlb_lock); > for (i = 0; i < needed; i++) { > - page = __alloc_buddy_huge_page_no_mpol(h, NUMA_NO_NODE); > + page = __alloc_buddy_huge_page(h, htlb_alloc_mask(h), > + NUMA_NO_NODE, NULL); > if (!page) { > alloc_ok = false; > break; >