* [failures] mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch removed from -mm tree
@ 2021-05-12 20:29 akpm
2021-05-12 22:55 ` Stephen Rothwell
0 siblings, 1 reply; 3+ messages in thread
From: akpm @ 2021-05-12 20:29 UTC (permalink / raw)
To: hdanton, mgorman, mhocko, mm-commits, npiggin,
oleksiy.avramchenko, rostedt, sfr, urezki, willy
The patch titled
Subject: mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
has been removed from the -mm tree. Its filename was
mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch
This patch was dropped because it had testing failures
------------------------------------------------------
From: "Uladzislau Rezki (Sony)" <urezki@gmail.com>
Subject: mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
Recently there has been introduced a page bulk allocator for users which
need to get number of pages per one call request.
For order-0 pages switch to __alloc_pages_bulk() instead of
alloc_pages_node(), the reason is the former is not capable of allocating
set of pages, thus a one call is per one page.
Second, according to my tests the bulk allocator uses less cycles even for
scenarios when only one page is requested. Running the "perf" on same
test case shows below difference:
<default>
- 45.18% __vmalloc_node
- __vmalloc_node_range
- 35.60% __alloc_pages
- get_page_from_freelist
3.36% __list_del_entry_valid
3.00% check_preemption_disabled
1.42% prep_new_page
<default>
<patch>
- 31.00% __vmalloc_node
- __vmalloc_node_range
- 14.48% __alloc_pages_bulk
3.22% __list_del_entry_valid
- 0.83% __alloc_pages
get_page_from_freelist
<patch>
The "test_vmalloc.sh" also shows performance improvements:
fix_size_alloc_test_4MB loops: 1000000 avg: 89105095 usec
fix_size_alloc_test loops: 1000000 avg: 513672 usec
full_fit_alloc_test loops: 1000000 avg: 748900 usec
long_busy_list_alloc_test loops: 1000000 avg: 8043038 usec
random_size_alloc_test loops: 1000000 avg: 4028582 usec
fix_align_alloc_test loops: 1000000 avg: 1457671 usec
fix_size_alloc_test_4MB loops: 1000000 avg: 62083711 usec
fix_size_alloc_test loops: 1000000 avg: 449207 usec
full_fit_alloc_test loops: 1000000 avg: 735985 usec
long_busy_list_alloc_test loops: 1000000 avg: 5176052 usec
random_size_alloc_test loops: 1000000 avg: 2589252 usec
fix_align_alloc_test loops: 1000000 avg: 1365009 usec
For example 4MB allocations illustrates ~30% gain, all the
rest is also better.
Link: https://lkml.kernel.org/r/20210509193844.2562-1-urezki@gmail.com
Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Cc: Steven Rostedt <rostedt@goodmis.org>
Cc: Stephen Rothwell <sfr@canb.auug.org.au>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
mm/vmalloc.c | 74 ++++++++++++++++++++++++++-----------------------
1 file changed, 40 insertions(+), 34 deletions(-)
--- a/mm/vmalloc.c~mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node
+++ a/mm/vmalloc.c
@@ -2765,8 +2765,6 @@ static void *__vmalloc_area_node(struct
unsigned long array_size;
unsigned int nr_small_pages = size >> PAGE_SHIFT;
unsigned int page_order;
- struct page **pages;
- unsigned int i;
array_size = (unsigned long)nr_small_pages * sizeof(struct page *);
gfp_mask |= __GFP_NOWARN;
@@ -2775,13 +2773,13 @@ static void *__vmalloc_area_node(struct
/* Please note that the recursion is strictly bounded. */
if (array_size > PAGE_SIZE) {
- pages = __vmalloc_node(array_size, 1, nested_gfp, node,
+ area->pages = __vmalloc_node(array_size, 1, nested_gfp, node,
area->caller);
} else {
- pages = kmalloc_node(array_size, nested_gfp, node);
+ area->pages = kmalloc_node(array_size, nested_gfp, node);
}
- if (!pages) {
+ if (!area->pages) {
free_vm_area(area);
warn_alloc(gfp_mask, NULL,
"vmalloc size %lu allocation failure: "
@@ -2790,43 +2788,51 @@ static void *__vmalloc_area_node(struct
return NULL;
}
- area->pages = pages;
- area->nr_pages = nr_small_pages;
set_vm_area_page_order(area, page_shift - PAGE_SHIFT);
-
page_order = vm_area_page_order(area);
- /*
- * Careful, we allocate and map page_order pages, but tracking is done
- * per PAGE_SIZE page so as to keep the vm_struct APIs independent of
- * the physical/mapped size.
- */
- for (i = 0; i < area->nr_pages; i += 1U << page_order) {
- struct page *page;
- int p;
-
- /* Compound pages required for remap_vmalloc_page */
- page = alloc_pages_node(node, gfp_mask | __GFP_COMP, page_order);
- if (unlikely(!page)) {
- /* Successfully allocated i pages, free them in __vfree() */
- area->nr_pages = i;
- atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
- warn_alloc(gfp_mask, NULL,
- "vmalloc size %lu allocation failure: "
- "page order %u allocation failed",
- area->nr_pages * PAGE_SIZE, page_order);
- goto fail;
- }
+ if (!page_order) {
+ area->nr_pages = __alloc_pages_bulk(gfp_mask, node,
+ NULL, nr_small_pages, NULL, area->pages);
+ } else {
+ /*
+ * Careful, we allocate and map page_order pages, but tracking is done
+ * per PAGE_SIZE page so as to keep the vm_struct APIs independent of
+ * the physical/mapped size.
+ */
+ for (area->nr_pages = 0; area->nr_pages < nr_small_pages;
+ area->nr_pages += 1U << page_order) {
+ struct page *page;
+ int i;
+
+ /* Compound pages required for remap_vmalloc_page */
+ page = alloc_pages_node(node, gfp_mask | __GFP_COMP, page_order);
+ if (unlikely(!page))
+ break;
- for (p = 0; p < (1U << page_order); p++)
- area->pages[i + p] = page + p;
+ for (i = 0; i < (1U << page_order); i++)
+ area->pages[area->nr_pages + i] = page + i;
- if (gfpflags_allow_blocking(gfp_mask))
- cond_resched();
+ if (gfpflags_allow_blocking(gfp_mask))
+ cond_resched();
+ }
}
+
atomic_long_add(area->nr_pages, &nr_vmalloc_pages);
- if (vmap_pages_range(addr, addr + size, prot, pages, page_shift) < 0) {
+ /*
+ * If not enough pages were obtained to accomplish an
+ * allocation request, free them via __vfree() if any.
+ */
+ if (area->nr_pages != nr_small_pages) {
+ warn_alloc(gfp_mask, NULL,
+ "vmalloc size %lu allocation failure: "
+ "page order %u allocation failed",
+ area->nr_pages * PAGE_SIZE, page_order);
+ goto fail;
+ }
+
+ if (vmap_pages_range(addr, addr + size, prot, area->pages, page_shift) < 0) {
warn_alloc(gfp_mask, NULL,
"vmalloc size %lu allocation failure: "
"failed to map pages",
_
Patches currently in -mm which might be from urezki@gmail.com are
mm-vmalloc-print-a-warning-message-first-on-failure.patch
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [failures] mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch removed from -mm tree
2021-05-12 20:29 [failures] mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch removed from -mm tree akpm
@ 2021-05-12 22:55 ` Stephen Rothwell
2021-05-13 10:32 ` Uladzislau Rezki
0 siblings, 1 reply; 3+ messages in thread
From: Stephen Rothwell @ 2021-05-12 22:55 UTC (permalink / raw)
To: akpm
Cc: hdanton, mgorman, mhocko, mm-commits, npiggin,
oleksiy.avramchenko, rostedt, urezki, willy
[-- Attachment #1: Type: text/plain, Size: 439 bytes --]
Hi Andrew,
On Wed, 12 May 2021 13:29:49 -0700 akpm@linux-foundation.org wrote:
>
> The patch titled
> Subject: mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
> has been removed from the -mm tree. Its filename was
> mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch
>
> This patch was dropped because it had testing failures
Removed from linux-next.
--
Cheers,
Stephen Rothwell
[-- Attachment #2: OpenPGP digital signature --]
[-- Type: application/pgp-signature, Size: 488 bytes --]
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [failures] mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch removed from -mm tree
2021-05-12 22:55 ` Stephen Rothwell
@ 2021-05-13 10:32 ` Uladzislau Rezki
0 siblings, 0 replies; 3+ messages in thread
From: Uladzislau Rezki @ 2021-05-13 10:32 UTC (permalink / raw)
To: Stephen Rothwell, akpm
Cc: akpm, hdanton, mgorman, mhocko, mm-commits, npiggin,
oleksiy.avramchenko, rostedt, urezki, willy
On Thu, May 13, 2021 at 08:55:24AM +1000, Stephen Rothwell wrote:
> Hi Andrew,
>
> On Wed, 12 May 2021 13:29:49 -0700 akpm@linux-foundation.org wrote:
> >
> > The patch titled
> > Subject: mm/vmalloc: switch to bulk allocator in __vmalloc_area_node()
> > has been removed from the -mm tree. Its filename was
> > mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch
> >
> > This patch was dropped because it had testing failures
>
> Removed from linux-next.
>
What can of testing failures does it trigger? Where can i find the
details, logs or tracers of it?
Thanks.
--
Vlad Rezki
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2021-05-13 10:33 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-05-12 20:29 [failures] mm-vmalloc-switch-to-bulk-allocator-in-__vmalloc_area_node.patch removed from -mm tree akpm
2021-05-12 22:55 ` Stephen Rothwell
2021-05-13 10:32 ` Uladzislau Rezki
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.