From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.2 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id CA68FC2BA83 for ; Sun, 9 Feb 2020 23:05:55 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 9DB17207FF for ; Sun, 9 Feb 2020 23:05:55 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727422AbgBIXFy (ORCPT ); Sun, 9 Feb 2020 18:05:54 -0500 Received: from mga11.intel.com ([192.55.52.93]:46152 "EHLO mga11.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726798AbgBIXFy (ORCPT ); Sun, 9 Feb 2020 18:05:54 -0500 X-Amp-Result: UNKNOWN X-Amp-Original-Verdict: FILE UNKNOWN X-Amp-File-Uploaded: False Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga102.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 09 Feb 2020 15:05:53 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.70,423,1574150400"; d="scan'208";a="225977840" Received: from richard.sh.intel.com (HELO localhost) ([10.239.159.54]) by orsmga008.jf.intel.com with ESMTP; 09 Feb 2020 15:05:39 -0800 Date: Mon, 10 Feb 2020 07:05:56 +0800 From: Wei Yang To: Baoquan He Cc: linux-kernel@vger.kernel.org, linux-mm@kvack.org, akpm@linux-foundation.org, dan.j.williams@intel.com, richardw.yang@linux.intel.com, david@redhat.com Subject: Re: [PATCH 1/7] mm/sparse.c: Introduce new function fill_subsection_map() Message-ID: <20200209230556.GA7326@richard> Reply-To: Wei Yang References: <20200209104826.3385-1-bhe@redhat.com> <20200209104826.3385-2-bhe@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200209104826.3385-2-bhe@redhat.com> User-Agent: Mutt/1.9.4 (2018-02-28) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Sun, Feb 09, 2020 at 06:48:20PM +0800, Baoquan He wrote: >Wrap the codes filling subsection map in section_activate() into >fill_subsection_map(), this makes section_activate() cleaner and >easier to follow. > This looks a preparation for #ifdef the code for VMEMMAP, then why not take the usage handling into this function too? >Signed-off-by: Baoquan He >--- > mm/sparse.c | 45 ++++++++++++++++++++++++++++++++++----------- > 1 file changed, 34 insertions(+), 11 deletions(-) > >diff --git a/mm/sparse.c b/mm/sparse.c >index c184b69460b7..9ad741ccbeb6 100644 >--- a/mm/sparse.c >+++ b/mm/sparse.c >@@ -788,24 +788,28 @@ static void section_deactivate(unsigned long pfn, unsigned long nr_pages, > depopulate_section_memmap(pfn, nr_pages, altmap); > } > >-static struct page * __meminit section_activate(int nid, unsigned long pfn, >- unsigned long nr_pages, struct vmem_altmap *altmap) >+/** >+ * fill_subsection_map - fill subsection map of a memory region >+ * @pfn - start pfn of the memory range >+ * @nr_pages - number of pfns to add in the region >+ * >+ * This clears the related subsection map inside one section, and only s/clears/fills/ ? >+ * intended for hotplug. >+ * >+ * Return: >+ * * 0 - On success. >+ * * -EINVAL - Invalid memory region. >+ * * -EEXIST - Subsection map has been set. >+ */ >+static int fill_subsection_map(unsigned long pfn, unsigned long nr_pages) > { >- DECLARE_BITMAP(map, SUBSECTIONS_PER_SECTION) = { 0 }; > struct mem_section *ms = __pfn_to_section(pfn); >- struct mem_section_usage *usage = NULL; >+ DECLARE_BITMAP(map, SUBSECTIONS_PER_SECTION) = { 0 }; > unsigned long *subsection_map; >- struct page *memmap; > int rc = 0; > > subsection_mask_set(map, pfn, nr_pages); > >- if (!ms->usage) { >- usage = kzalloc(mem_section_usage_size(), GFP_KERNEL); >- if (!usage) >- return ERR_PTR(-ENOMEM); >- ms->usage = usage; >- } > subsection_map = &ms->usage->subsection_map[0]; > > if (bitmap_empty(map, SUBSECTIONS_PER_SECTION)) >@@ -816,6 +820,25 @@ static struct page * __meminit section_activate(int nid, unsigned long pfn, > bitmap_or(subsection_map, map, subsection_map, > SUBSECTIONS_PER_SECTION); > >+ return rc; >+} >+ >+static struct page * __meminit section_activate(int nid, unsigned long pfn, >+ unsigned long nr_pages, struct vmem_altmap *altmap) >+{ >+ struct mem_section *ms = __pfn_to_section(pfn); >+ struct mem_section_usage *usage = NULL; >+ struct page *memmap; >+ int rc = 0; >+ >+ if (!ms->usage) { >+ usage = kzalloc(mem_section_usage_size(), GFP_KERNEL); >+ if (!usage) >+ return ERR_PTR(-ENOMEM); >+ ms->usage = usage; >+ } >+ >+ rc = fill_subsection_map(pfn, nr_pages); > if (rc) { > if (usage) > ms->usage = NULL; >-- >2.17.2 -- Wei Yang Help you, Help me