From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753165AbdHKNE5 (ORCPT ); Fri, 11 Aug 2017 09:04:57 -0400 Received: from mx2.suse.de ([195.135.220.15]:54674 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1752855AbdHKNEy (ORCPT ); Fri, 11 Aug 2017 09:04:54 -0400 Date: Fri, 11 Aug 2017 15:04:49 +0200 From: Michal Hocko To: Pavel Tatashin Cc: linux-kernel@vger.kernel.org, sparclinux@vger.kernel.org, linux-mm@kvack.org, linuxppc-dev@lists.ozlabs.org, linux-s390@vger.kernel.org, linux-arm-kernel@lists.infradead.org, x86@kernel.org, kasan-dev@googlegroups.com, borntraeger@de.ibm.com, heiko.carstens@de.ibm.com, davem@davemloft.net, willy@infradead.org, ard.biesheuvel@linaro.org, will.deacon@arm.com, catalin.marinas@arm.com, sam@ravnborg.org Subject: Re: [v6 13/15] mm: stop zeroing memory during allocation in vmemmap Message-ID: <20170811130449.GL30811@dhcp22.suse.cz> References: <1502138329-123460-1-git-send-email-pasha.tatashin@oracle.com> <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> User-Agent: Mutt/1.5.23 (2014-03-12) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon 07-08-17 16:38:47, Pavel Tatashin wrote: > Replace allocators in sprase-vmemmap to use the non-zeroing version. So, > we will get the performance improvement by zeroing the memory in parallel > when struct pages are zeroed. First of all this should be probably merged with the previous patch. The I think vmemmap_alloc_block would be better to split up into __vmemmap_alloc_block which doesn't zero and vmemmap_alloc_block which does zero which would reduce the memset callsites and it would make it slightly more robust interface. > Signed-off-by: Pavel Tatashin > Reviewed-by: Steven Sistare > Reviewed-by: Daniel Jordan > Reviewed-by: Bob Picco > --- > mm/sparse-vmemmap.c | 6 +++--- > mm/sparse.c | 6 +++--- > 2 files changed, 6 insertions(+), 6 deletions(-) > > diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c > index d40c721ab19f..3b646b5ce1b6 100644 > --- a/mm/sparse-vmemmap.c > +++ b/mm/sparse-vmemmap.c > @@ -41,7 +41,7 @@ static void * __ref __earlyonly_bootmem_alloc(int node, > unsigned long align, > unsigned long goal) > { > - return memblock_virt_alloc_try_nid(size, align, goal, > + return memblock_virt_alloc_try_nid_raw(size, align, goal, > BOOTMEM_ALLOC_ACCESSIBLE, node); > } > > @@ -56,11 +56,11 @@ void * __meminit vmemmap_alloc_block(unsigned long size, int node) > > if (node_state(node, N_HIGH_MEMORY)) > page = alloc_pages_node( > - node, GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + node, GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > else > page = alloc_pages( > - GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > if (page) > return page_address(page); > diff --git a/mm/sparse.c b/mm/sparse.c > index 7b4be3fd5cac..0e315766ad11 100644 > --- a/mm/sparse.c > +++ b/mm/sparse.c > @@ -441,9 +441,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, > } > > size = PAGE_ALIGN(size); > - map = memblock_virt_alloc_try_nid(size * map_count, > - PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > - BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > + map = memblock_virt_alloc_try_nid_raw(size * map_count, > + PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > + BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > if (map) { > for (pnum = pnum_begin; pnum < pnum_end; pnum++) { > if (!present_section_nr(pnum)) > -- > 2.14.0 -- Michal Hocko SUSE Labs From mboxrd@z Thu Jan 1 00:00:00 1970 From: Michal Hocko Date: Fri, 11 Aug 2017 13:04:49 +0000 Subject: Re: [v6 13/15] mm: stop zeroing memory during allocation in vmemmap Message-Id: <20170811130449.GL30811@dhcp22.suse.cz> List-Id: References: <1502138329-123460-1-git-send-email-pasha.tatashin@oracle.com> <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> In-Reply-To: <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit To: linux-arm-kernel@lists.infradead.org On Mon 07-08-17 16:38:47, Pavel Tatashin wrote: > Replace allocators in sprase-vmemmap to use the non-zeroing version. So, > we will get the performance improvement by zeroing the memory in parallel > when struct pages are zeroed. First of all this should be probably merged with the previous patch. The I think vmemmap_alloc_block would be better to split up into __vmemmap_alloc_block which doesn't zero and vmemmap_alloc_block which does zero which would reduce the memset callsites and it would make it slightly more robust interface. > Signed-off-by: Pavel Tatashin > Reviewed-by: Steven Sistare > Reviewed-by: Daniel Jordan > Reviewed-by: Bob Picco > --- > mm/sparse-vmemmap.c | 6 +++--- > mm/sparse.c | 6 +++--- > 2 files changed, 6 insertions(+), 6 deletions(-) > > diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c > index d40c721ab19f..3b646b5ce1b6 100644 > --- a/mm/sparse-vmemmap.c > +++ b/mm/sparse-vmemmap.c > @@ -41,7 +41,7 @@ static void * __ref __earlyonly_bootmem_alloc(int node, > unsigned long align, > unsigned long goal) > { > - return memblock_virt_alloc_try_nid(size, align, goal, > + return memblock_virt_alloc_try_nid_raw(size, align, goal, > BOOTMEM_ALLOC_ACCESSIBLE, node); > } > > @@ -56,11 +56,11 @@ void * __meminit vmemmap_alloc_block(unsigned long size, int node) > > if (node_state(node, N_HIGH_MEMORY)) > page = alloc_pages_node( > - node, GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + node, GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > else > page = alloc_pages( > - GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > if (page) > return page_address(page); > diff --git a/mm/sparse.c b/mm/sparse.c > index 7b4be3fd5cac..0e315766ad11 100644 > --- a/mm/sparse.c > +++ b/mm/sparse.c > @@ -441,9 +441,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, > } > > size = PAGE_ALIGN(size); > - map = memblock_virt_alloc_try_nid(size * map_count, > - PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > - BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > + map = memblock_virt_alloc_try_nid_raw(size * map_count, > + PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > + BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > if (map) { > for (pnum = pnum_begin; pnum < pnum_end; pnum++) { > if (!present_section_nr(pnum)) > -- > 2.14.0 -- Michal Hocko SUSE Labs From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wr0-f199.google.com (mail-wr0-f199.google.com [209.85.128.199]) by kanga.kvack.org (Postfix) with ESMTP id 709716B0292 for ; Fri, 11 Aug 2017 09:04:54 -0400 (EDT) Received: by mail-wr0-f199.google.com with SMTP id q50so4882569wrb.14 for ; Fri, 11 Aug 2017 06:04:54 -0700 (PDT) Received: from mx1.suse.de (mx2.suse.de. [195.135.220.15]) by mx.google.com with ESMTPS id r7si688384wra.40.2017.08.11.06.04.53 for (version=TLS1 cipher=AES128-SHA bits=128/128); Fri, 11 Aug 2017 06:04:53 -0700 (PDT) Date: Fri, 11 Aug 2017 15:04:49 +0200 From: Michal Hocko Subject: Re: [v6 13/15] mm: stop zeroing memory during allocation in vmemmap Message-ID: <20170811130449.GL30811@dhcp22.suse.cz> References: <1502138329-123460-1-git-send-email-pasha.tatashin@oracle.com> <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> Sender: owner-linux-mm@kvack.org List-ID: To: Pavel Tatashin Cc: linux-kernel@vger.kernel.org, sparclinux@vger.kernel.org, linux-mm@kvack.org, linuxppc-dev@lists.ozlabs.org, linux-s390@vger.kernel.org, linux-arm-kernel@lists.infradead.org, x86@kernel.org, kasan-dev@googlegroups.com, borntraeger@de.ibm.com, heiko.carstens@de.ibm.com, davem@davemloft.net, willy@infradead.org, ard.biesheuvel@linaro.org, will.deacon@arm.com, catalin.marinas@arm.com, sam@ravnborg.org On Mon 07-08-17 16:38:47, Pavel Tatashin wrote: > Replace allocators in sprase-vmemmap to use the non-zeroing version. So, > we will get the performance improvement by zeroing the memory in parallel > when struct pages are zeroed. First of all this should be probably merged with the previous patch. The I think vmemmap_alloc_block would be better to split up into __vmemmap_alloc_block which doesn't zero and vmemmap_alloc_block which does zero which would reduce the memset callsites and it would make it slightly more robust interface. > Signed-off-by: Pavel Tatashin > Reviewed-by: Steven Sistare > Reviewed-by: Daniel Jordan > Reviewed-by: Bob Picco > --- > mm/sparse-vmemmap.c | 6 +++--- > mm/sparse.c | 6 +++--- > 2 files changed, 6 insertions(+), 6 deletions(-) > > diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c > index d40c721ab19f..3b646b5ce1b6 100644 > --- a/mm/sparse-vmemmap.c > +++ b/mm/sparse-vmemmap.c > @@ -41,7 +41,7 @@ static void * __ref __earlyonly_bootmem_alloc(int node, > unsigned long align, > unsigned long goal) > { > - return memblock_virt_alloc_try_nid(size, align, goal, > + return memblock_virt_alloc_try_nid_raw(size, align, goal, > BOOTMEM_ALLOC_ACCESSIBLE, node); > } > > @@ -56,11 +56,11 @@ void * __meminit vmemmap_alloc_block(unsigned long size, int node) > > if (node_state(node, N_HIGH_MEMORY)) > page = alloc_pages_node( > - node, GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + node, GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > else > page = alloc_pages( > - GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > if (page) > return page_address(page); > diff --git a/mm/sparse.c b/mm/sparse.c > index 7b4be3fd5cac..0e315766ad11 100644 > --- a/mm/sparse.c > +++ b/mm/sparse.c > @@ -441,9 +441,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, > } > > size = PAGE_ALIGN(size); > - map = memblock_virt_alloc_try_nid(size * map_count, > - PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > - BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > + map = memblock_virt_alloc_try_nid_raw(size * map_count, > + PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > + BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > if (map) { > for (pnum = pnum_begin; pnum < pnum_end; pnum++) { > if (!present_section_nr(pnum)) > -- > 2.14.0 -- Michal Hocko SUSE Labs -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org From mboxrd@z Thu Jan 1 00:00:00 1970 From: mhocko@kernel.org (Michal Hocko) Date: Fri, 11 Aug 2017 15:04:49 +0200 Subject: [v6 13/15] mm: stop zeroing memory during allocation in vmemmap In-Reply-To: <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> References: <1502138329-123460-1-git-send-email-pasha.tatashin@oracle.com> <1502138329-123460-14-git-send-email-pasha.tatashin@oracle.com> Message-ID: <20170811130449.GL30811@dhcp22.suse.cz> To: linux-arm-kernel@lists.infradead.org List-Id: linux-arm-kernel.lists.infradead.org On Mon 07-08-17 16:38:47, Pavel Tatashin wrote: > Replace allocators in sprase-vmemmap to use the non-zeroing version. So, > we will get the performance improvement by zeroing the memory in parallel > when struct pages are zeroed. First of all this should be probably merged with the previous patch. The I think vmemmap_alloc_block would be better to split up into __vmemmap_alloc_block which doesn't zero and vmemmap_alloc_block which does zero which would reduce the memset callsites and it would make it slightly more robust interface. > Signed-off-by: Pavel Tatashin > Reviewed-by: Steven Sistare > Reviewed-by: Daniel Jordan > Reviewed-by: Bob Picco > --- > mm/sparse-vmemmap.c | 6 +++--- > mm/sparse.c | 6 +++--- > 2 files changed, 6 insertions(+), 6 deletions(-) > > diff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c > index d40c721ab19f..3b646b5ce1b6 100644 > --- a/mm/sparse-vmemmap.c > +++ b/mm/sparse-vmemmap.c > @@ -41,7 +41,7 @@ static void * __ref __earlyonly_bootmem_alloc(int node, > unsigned long align, > unsigned long goal) > { > - return memblock_virt_alloc_try_nid(size, align, goal, > + return memblock_virt_alloc_try_nid_raw(size, align, goal, > BOOTMEM_ALLOC_ACCESSIBLE, node); > } > > @@ -56,11 +56,11 @@ void * __meminit vmemmap_alloc_block(unsigned long size, int node) > > if (node_state(node, N_HIGH_MEMORY)) > page = alloc_pages_node( > - node, GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + node, GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > else > page = alloc_pages( > - GFP_KERNEL | __GFP_ZERO | __GFP_RETRY_MAYFAIL, > + GFP_KERNEL | __GFP_RETRY_MAYFAIL, > get_order(size)); > if (page) > return page_address(page); > diff --git a/mm/sparse.c b/mm/sparse.c > index 7b4be3fd5cac..0e315766ad11 100644 > --- a/mm/sparse.c > +++ b/mm/sparse.c > @@ -441,9 +441,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, > } > > size = PAGE_ALIGN(size); > - map = memblock_virt_alloc_try_nid(size * map_count, > - PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > - BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > + map = memblock_virt_alloc_try_nid_raw(size * map_count, > + PAGE_SIZE, __pa(MAX_DMA_ADDRESS), > + BOOTMEM_ALLOC_ACCESSIBLE, nodeid); > if (map) { > for (pnum = pnum_begin; pnum < pnum_end; pnum++) { > if (!present_section_nr(pnum)) > -- > 2.14.0 -- Michal Hocko SUSE Labs