From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: ARC-Seal: i=1; a=rsa-sha256; t=1522228447; cv=none; d=google.com; s=arc-20160816; b=zZxuDxHBxmFa+2MiGQuNfRqlEQBmlbMygzHrQFcswhLY8NqULjWJ9TCQnQkobAWWgM h3QNOjHuL7EfczqSoYRCEmvf0ftTjNCRwsynEkzMNAyBAR2SXwhsDlKZj8y7zM6kwPvG r9ipjY7tws2ltbg4AEDa/vLoGxpwoaH9Rmu+IhK7GZitkMs3Ta7StB9QWzstfUpH4gib ZAuaq7cYaQPh8OHZu4/2EnLv2dM0lz8ORcuvAop8SsXbGP/TgzYXwNbSF9PkZE8lYDeb EG5wXmS4RUEf6VakX37B677xLw6LBFLlGJf6b1gEM1MYEC9pXmfmd/2zNC93rk/DBM1N Kdfg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=user-agent:in-reply-to:content-disposition:mime-version:references :reply-to:message-id:subject:cc:to:from:date:dkim-signature :arc-authentication-results; bh=hxiq/bleI/RiF1pIcTH/GOo4Dtqmo4x8aQIaJ+XOexY=; b=v5Q16Yvpg6wMCPBn9xXNxmTI0Day4gRs2VOhUFjC6Iz6ReczbJhGIprVS6ihqXVMDw nALsFmkYY4bi/0x0m8iyEzoh/58+sSRX6FWPsiskuk6ZJf8paz3jVndaO7s2B288Pov4 dRikcZMAlcXGWtwwPYJL7sVJLX/mk4K+/haqQRsO89t8niusR6lLq4/vTM8BnS7gYud5 xfVW1XGJj4XQJEuFdKfcAnOlnmN1uBNFzZdL7fKzxrcSGlwi8RWo/989DQCWJH4W+XDD TfUBdwgGrsSiQZbYf/FYG2le4RDU+1t5+6nw+8/rp5Y5RW8OXshUZjSntc63dO4k1DCJ 0JwQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=FIVeavD/; spf=pass (google.com: domain of richard.weiyang@gmail.com designates 209.85.220.65 as permitted sender) smtp.mailfrom=richard.weiyang@gmail.com; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=FIVeavD/; spf=pass (google.com: domain of richard.weiyang@gmail.com designates 209.85.220.65 as permitted sender) smtp.mailfrom=richard.weiyang@gmail.com; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com X-Google-Smtp-Source: AIpwx4+tt1CVY/wqXjsbDDecw/N5yk0ty2cs5eO8CvIEh1Ht8DvKWGG6WawnPijrGxJB9j98jINdkw== Date: Wed, 28 Mar 2018 17:13:57 +0800 From: Wei Yang To: Jia He Cc: Andrew Morton , Michal Hocko , Catalin Marinas , Mel Gorman , Will Deacon , Mark Rutland , Thomas Gleixner , Ingo Molnar , "H. Peter Anvin" , Pavel Tatashin , Daniel Jordan , AKASHI Takahiro , Gioh Kim , Steven Sistare , Daniel Vacek , Eugeniu Rosca , Vlastimil Babka , linux-kernel@vger.kernel.org, linux-mm@kvack.org, James Morse , Ard Biesheuvel , Steve Capper , x86@kernel.org, Greg Kroah-Hartman , Kate Stewart , Philippe Ombredanne , Johannes Weiner , Kemi Wang , Petr Tesarik , YASUAKI ISHIMATSU , Andrey Ryabinin , Nikolay Borisov , Jia He Subject: Re: [PATCH v2 1/5] mm: page_alloc: remain memblock_next_valid_pfn() when CONFIG_HAVE_ARCH_PFN_VALID is enable Message-ID: <20180328091357.GA97260@WeideMacBook-Pro.local> Reply-To: Wei Yang References: <1521894282-6454-1-git-send-email-hejianet@gmail.com> <1521894282-6454-2-git-send-email-hejianet@gmail.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1521894282-6454-2-git-send-email-hejianet@gmail.com> User-Agent: Mutt/1.9.1 (2017-09-22) X-getmail-retrieved-from-mailbox: INBOX X-GMAIL-THRID: =?utf-8?q?1595821857199891353?= X-GMAIL-MSGID: =?utf-8?q?1596172216849691695?= X-Mailing-List: linux-kernel@vger.kernel.org List-ID: On Sat, Mar 24, 2018 at 05:24:38AM -0700, Jia He wrote: >Commit b92df1de5d28 ("mm: page_alloc: skip over regions of invalid pfns >where possible") optimized the loop in memmap_init_zone(). But it causes >possible panic bug. So Daniel Vacek reverted it later. > Why this has a bug? Do you have some link about it? If the audience could know the potential risk, it would be helpful to review the code and decide whether to take it back. >But memblock_next_valid_pfn is valid when CONFIG_HAVE_ARCH_PFN_VALID is >enabled. And as verified by Eugeniu Rosca, arm can benifit from this >commit. So remain the memblock_next_valid_pfn. > >Signed-off-by: Jia He >--- > include/linux/memblock.h | 4 ++++ > mm/memblock.c | 29 +++++++++++++++++++++++++++++ > mm/page_alloc.c | 11 ++++++++++- > 3 files changed, 43 insertions(+), 1 deletion(-) > >diff --git a/include/linux/memblock.h b/include/linux/memblock.h >index 0257aee..efbbe4b 100644 >--- a/include/linux/memblock.h >+++ b/include/linux/memblock.h >@@ -203,6 +203,10 @@ void __next_mem_pfn_range(int *idx, int nid, unsigned long *out_start_pfn, > i >= 0; __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid)) > #endif /* CONFIG_HAVE_MEMBLOCK_NODE_MAP */ > >+#ifdef CONFIG_HAVE_ARCH_PFN_VALID >+unsigned long memblock_next_valid_pfn(unsigned long pfn); >+#endif >+ > /** > * for_each_free_mem_range - iterate through free memblock areas > * @i: u64 used as loop variable >diff --git a/mm/memblock.c b/mm/memblock.c >index ba7c878..bea5a9c 100644 >--- a/mm/memblock.c >+++ b/mm/memblock.c >@@ -1102,6 +1102,35 @@ void __init_memblock __next_mem_pfn_range(int *idx, int nid, > *out_nid = r->nid; > } > >+#ifdef CONFIG_HAVE_ARCH_PFN_VALID >+unsigned long __init_memblock memblock_next_valid_pfn(unsigned long pfn) >+{ >+ struct memblock_type *type = &memblock.memory; >+ unsigned int right = type->cnt; >+ unsigned int mid, left = 0; >+ phys_addr_t addr = PFN_PHYS(++pfn); >+ >+ do { >+ mid = (right + left) / 2; >+ >+ if (addr < type->regions[mid].base) >+ right = mid; >+ else if (addr >= (type->regions[mid].base + >+ type->regions[mid].size)) >+ left = mid + 1; >+ else { >+ /* addr is within the region, so pfn is valid */ >+ return pfn; >+ } >+ } while (left < right); >+ >+ if (right == type->cnt) >+ return -1UL; >+ else >+ return PHYS_PFN(type->regions[right].base); >+} >+#endif /*CONFIG_HAVE_ARCH_PFN_VALID*/ >+ > /** > * memblock_set_node - set node ID on memblock regions > * @base: base of area to set node ID for >diff --git a/mm/page_alloc.c b/mm/page_alloc.c >index c19f5ac..2a967f7 100644 >--- a/mm/page_alloc.c >+++ b/mm/page_alloc.c >@@ -5483,8 +5483,17 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone, > if (context != MEMMAP_EARLY) > goto not_early; > >- if (!early_pfn_valid(pfn)) >+ if (!early_pfn_valid(pfn)) { >+#if (defined CONFIG_HAVE_MEMBLOCK) && (defined CONFIG_HAVE_ARCH_PFN_VALID) In commit b92df1de5d28, it use CONFIG_HAVE_MEMBLOCK_NODE_MAP. Not get the point of your change. >+ /* >+ * Skip to the pfn preceding the next valid one (or >+ * end_pfn), such that we hit a valid pfn (or end_pfn) >+ * on our next iteration of the loop. >+ */ >+ pfn = memblock_next_valid_pfn(pfn) - 1; >+#endif > continue; >+ } > if (!early_pfn_in_nid(pfn, nid)) > continue; > if (!update_defer_init(pgdat, pfn, end_pfn, &nr_initialised)) >-- >2.7.4 -- Wei Yang Help you, Help me