From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-9.8 required=3.0 tests=DKIM_INVALID,DKIM_SIGNED, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id BF407C2D0EC for ; Sun, 12 Apr 2020 19:54:00 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 77BEC20CC7 for ; Sun, 12 Apr 2020 19:54:00 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (1024-bit key) header.d=kernel.org header.i=@kernel.org header.b="LhJ43W7x" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 77BEC20CC7 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=kernel.org Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 22BF78E00E8; Sun, 12 Apr 2020 15:54:00 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 2012B8E00D0; Sun, 12 Apr 2020 15:54:00 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 0F0328E00E8; Sun, 12 Apr 2020 15:54:00 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0035.hostedemail.com [216.40.44.35]) by kanga.kvack.org (Postfix) with ESMTP id EAA8C8E00D0 for ; Sun, 12 Apr 2020 15:53:59 -0400 (EDT) Received: from smtpin04.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay03.hostedemail.com (Postfix) with ESMTP id A3426824556B for ; Sun, 12 Apr 2020 19:53:59 +0000 (UTC) X-FDA: 76700253798.04.girl53_22cefbee1713d X-HE-Tag: girl53_22cefbee1713d X-Filterd-Recvd-Size: 10011 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by imf24.hostedemail.com (Postfix) with ESMTP for ; Sun, 12 Apr 2020 19:53:59 +0000 (UTC) Received: from aquarius.haifa.ibm.com (nesher1.haifa.il.ibm.com [195.110.40.7]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-SHA256 (128/128 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 802D5208E0; Sun, 12 Apr 2020 19:53:44 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1586721238; bh=b61jXh3YXZS+2XpmCP+m9oEdP1VkDOnLWvH5ZgMQlFk=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=LhJ43W7xtnqQQNK2hYGw3ieQ3iZHXnp0QUM2x2StqqTWVLjUlGAG2cusDiztOcCBQ swFoDidRxfG7L+Dexnyg+1kssnHNv4bhxgQitoE6H2HzF0UBng6Q6YawUUQxtqzKUF IqdYVp1UjDKAnVXtZwry8mP9lRQ0jrns9Qz0xzKo= From: Mike Rapoport To: linux-kernel@vger.kernel.org Cc: Andrew Morton , Baoquan He , Brian Cain , Catalin Marinas , "David S. Miller" , Geert Uytterhoeven , Greentime Hu , Greg Ungerer , Guan Xuetao , Guo Ren , Heiko Carstens , Helge Deller , Hoan Tran , "James E.J. Bottomley" , Jonathan Corbet , Ley Foon Tan , Mark Salter , Matt Turner , Max Filippov , Michael Ellerman , Michal Hocko , Michal Simek , Mike Rapoport , Nick Hu , Paul Walmsley , Richard Weinberger , Rich Felker , Russell King , Stafford Horne , Thomas Bogendoerfer , Tony Luck , Vineet Gupta , x86@kernel.org, Yoshinori Sato , linux-alpha@vger.kernel.org, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-c6x-dev@linux-c6x.org, linux-csky@vger.kernel.org, linux-doc@vger.kernel.org, linux-hexagon@vger.kernel.org, linux-ia64@vger.kernel.org, linux-m68k@lists.linux-m68k.org, linux-mips@vger.kernel.org, linux-mm@kvack.org, linux-parisc@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-riscv@lists.infradead.org, linux-s390@vger.kernel.org, linux-sh@vger.kernel.org, linux-snps-arc@lists.infradead.org, linux-um@lists.infradead.org, linux-xtensa@linux-xtensa.org, openrisc@lists.librecores.org, sparclinux@vger.kernel.org, uclinux-h8-devel@lists.sourceforge.jp, Mike Rapoport Subject: [PATCH 19/21] mm: clean up free_area_init_node() and its helpers Date: Sun, 12 Apr 2020 22:48:57 +0300 Message-Id: <20200412194859.12663-20-rppt@kernel.org> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20200412194859.12663-1-rppt@kernel.org> References: <20200412194859.12663-1-rppt@kernel.org> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: From: Mike Rapoport The free_area_init_node() now always uses memblock info and the zone PFN limits so it does not need the backwards compatibility functions to calculate the zone spanned and absent pages. The removal of the compat_ versions of zone_{abscent,spanned}_pages_in_node() in turn, makes zone_si= ze and zhole_size parameters unused. The node_start_pfn is determined by get_pfn_range_for_nid(), so there is = no need to pass it to free_area_init_node(). As the result, the only required parameter to free_area_init_node() is th= e node ID, all the rest are removed along with no longer used compat_zone_{abscent,spanned}_pages_in_node() helpers. Signed-off-by: Mike Rapoport --- mm/page_alloc.c | 104 ++++++++++-------------------------------------- 1 file changed, 22 insertions(+), 82 deletions(-) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index e46232ec4849..9af27ee784c7 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -6441,8 +6441,7 @@ static unsigned long __init zone_spanned_pages_in_n= ode(int nid, unsigned long node_start_pfn, unsigned long node_end_pfn, unsigned long *zone_start_pfn, - unsigned long *zone_end_pfn, - unsigned long *ignored) + unsigned long *zone_end_pfn) { unsigned long zone_low =3D arch_zone_lowest_possible_pfn[zone_type]; unsigned long zone_high =3D arch_zone_highest_possible_pfn[zone_type]; @@ -6506,8 +6505,7 @@ unsigned long __init absent_pages_in_range(unsigned= long start_pfn, static unsigned long __init zone_absent_pages_in_node(int nid, unsigned long zone_type, unsigned long node_start_pfn, - unsigned long node_end_pfn, - unsigned long *ignored) + unsigned long node_end_pfn) { unsigned long zone_low =3D arch_zone_lowest_possible_pfn[zone_type]; unsigned long zone_high =3D arch_zone_highest_possible_pfn[zone_type]; @@ -6554,43 +6552,9 @@ static unsigned long __init zone_absent_pages_in_n= ode(int nid, return nr_absent; } =20 -static inline unsigned long __init compat_zone_spanned_pages_in_node(int= nid, - unsigned long zone_type, - unsigned long node_start_pfn, - unsigned long node_end_pfn, - unsigned long *zone_start_pfn, - unsigned long *zone_end_pfn, - unsigned long *zones_size) -{ - unsigned int zone; - - *zone_start_pfn =3D node_start_pfn; - for (zone =3D 0; zone < zone_type; zone++) - *zone_start_pfn +=3D zones_size[zone]; - - *zone_end_pfn =3D *zone_start_pfn + zones_size[zone_type]; - - return zones_size[zone_type]; -} - -static inline unsigned long __init compat_zone_absent_pages_in_node(int = nid, - unsigned long zone_type, - unsigned long node_start_pfn, - unsigned long node_end_pfn, - unsigned long *zholes_size) -{ - if (!zholes_size) - return 0; - - return zholes_size[zone_type]; -} - static void __init calculate_node_totalpages(struct pglist_data *pgdat, unsigned long node_start_pfn, - unsigned long node_end_pfn, - unsigned long *zones_size, - unsigned long *zholes_size, - bool compat) + unsigned long node_end_pfn) { unsigned long realtotalpages =3D 0, totalpages =3D 0; enum zone_type i; @@ -6601,31 +6565,14 @@ static void __init calculate_node_totalpages(stru= ct pglist_data *pgdat, unsigned long spanned, absent; unsigned long size, real_size; =20 - if (compat) { - spanned =3D compat_zone_spanned_pages_in_node( - pgdat->node_id, i, - node_start_pfn, - node_end_pfn, - &zone_start_pfn, - &zone_end_pfn, - zones_size); - absent =3D compat_zone_absent_pages_in_node( - pgdat->node_id, i, - node_start_pfn, - node_end_pfn, - zholes_size); - } else { - spanned =3D zone_spanned_pages_in_node(pgdat->node_id, i, - node_start_pfn, - node_end_pfn, - &zone_start_pfn, - &zone_end_pfn, - zones_size); - absent =3D zone_absent_pages_in_node(pgdat->node_id, i, - node_start_pfn, - node_end_pfn, - zholes_size); - } + spanned =3D zone_spanned_pages_in_node(pgdat->node_id, i, + node_start_pfn, + node_end_pfn, + &zone_start_pfn, + &zone_end_pfn); + absent =3D zone_absent_pages_in_node(pgdat->node_id, i, + node_start_pfn, + node_end_pfn); =20 size =3D spanned; real_size =3D size - absent; @@ -6947,10 +6894,7 @@ static inline void pgdat_set_deferred_range(pg_dat= a_t *pgdat) static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {} #endif =20 -static void __init __free_area_init_node(int nid, unsigned long *zones_s= ize, - unsigned long node_start_pfn, - unsigned long *zholes_size, - bool compat) +static void __init free_area_init_node(int nid) { pg_data_t *pgdat =3D NODE_DATA(nid); unsigned long start_pfn =3D 0; @@ -6959,19 +6903,16 @@ static void __init __free_area_init_node(int nid,= unsigned long *zones_size, /* pg_data_t should be reset to zero when it's allocated */ WARN_ON(pgdat->nr_zones || pgdat->kswapd_classzone_idx); =20 + get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); + pgdat->node_id =3D nid; - pgdat->node_start_pfn =3D node_start_pfn; + pgdat->node_start_pfn =3D start_pfn; pgdat->per_cpu_nodestats =3D NULL; - if (!compat) { - get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); - pr_info("Initmem setup node %d [mem %#018Lx-%#018Lx]\n", nid, - (u64)start_pfn << PAGE_SHIFT, - end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0); - } else { - start_pfn =3D node_start_pfn; - } - calculate_node_totalpages(pgdat, start_pfn, end_pfn, - zones_size, zholes_size, compat); + + pr_info("Initmem setup node %d [mem %#018Lx-%#018Lx]\n", nid, + (u64)start_pfn << PAGE_SHIFT, + end_pfn ? ((u64)end_pfn << PAGE_SHIFT) - 1 : 0); + calculate_node_totalpages(pgdat, start_pfn, end_pfn); =20 alloc_node_mem_map(pgdat); pgdat_set_deferred_range(pgdat); @@ -6981,7 +6922,7 @@ static void __init __free_area_init_node(int nid, u= nsigned long *zones_size, =20 void __init free_area_init_memoryless_node(int nid) { - __free_area_init_node(nid, NULL, 0, NULL, false); + free_area_init_node(nid); } =20 #if !defined(CONFIG_FLAT_NODE_MEM_MAP) @@ -7509,8 +7450,7 @@ void __init free_area_init(unsigned long *max_zone_= pfn) init_unavailable_mem(); for_each_online_node(nid) { pg_data_t *pgdat =3D NODE_DATA(nid); - __free_area_init_node(nid, NULL, - find_min_pfn_for_node(nid), NULL, false); + free_area_init_node(nid); =20 /* Any memory on that node */ if (pgdat->node_present_pages) --=20 2.25.1