Explicitly initialize the pagesets after the per cpu areas have been initialized. This is necessary in order to be able to use per cpu operations in later patches. Cc: Mel Gorman Signed-off-by: Christoph Lameter --- arch/ia64/kernel/setup.c | 1 + arch/powerpc/kernel/setup_64.c | 1 + arch/sparc/kernel/smp_64.c | 1 + arch/x86/kernel/setup_percpu.c | 2 ++ include/linux/mm.h | 1 + init/main.c | 1 + mm/page_alloc.c | 40 +++++++++++++++++++++++++++++----------- 7 files changed, 36 insertions(+), 11 deletions(-) Index: linux-2.6/init/main.c =================================================================== --- linux-2.6.orig/init/main.c 2009-06-17 14:06:22.000000000 -0500 +++ linux-2.6/init/main.c 2009-06-17 14:11:30.000000000 -0500 @@ -398,6 +398,7 @@ static void __init setup_per_cpu_areas(v memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start); ptr += size; } + setup_pagesets(); } #endif /* CONFIG_HAVE_SETUP_PER_CPU_AREA */ Index: linux-2.6/mm/page_alloc.c =================================================================== --- linux-2.6.orig/mm/page_alloc.c 2009-06-17 14:11:04.000000000 -0500 +++ linux-2.6/mm/page_alloc.c 2009-06-17 14:11:30.000000000 -0500 @@ -3129,23 +3129,42 @@ int zone_wait_table_init(struct zone *zo return 0; } -static __meminit void zone_pcp_init(struct zone *zone) +/* + * Early setup of pagesets. + * + * In the NUMA case the pageset setup simply results in all zones pcp + * pointer being directed at a per cpu pageset with zero batchsize. + * + * This means that every free and every allocation occurs directly from + * the buddy allocator tables. + * + * The pageset never queues pages during early boot and is therefore usable + * for every type of zone. + */ +__meminit void setup_pagesets(void) { int cpu; - unsigned long batch = zone_batchsize(zone); + struct zone *zone; - for (cpu = 0; cpu < NR_CPUS; cpu++) { + for_each_zone(zone) { #ifdef CONFIG_NUMA - /* Early boot. Slab allocator not functional yet */ - zone_pcp(zone, cpu) = &boot_pageset[cpu]; - setup_pageset(&boot_pageset[cpu],0); + unsigned long batch = 0; + + for (cpu = 0; cpu < NR_CPUS; cpu++) { + /* Early boot. Slab allocator not functional yet */ + zone_pcp(zone, cpu) = &boot_pageset[cpu]; + } #else - setup_pageset(zone_pcp(zone,cpu), batch); + unsigned long batch = zone_batchsize(zone); #endif + + for_each_possible_cpu(cpu) + setup_pageset(zone_pcp(zone, cpu), batch); + + if (zone->present_pages) + printk(KERN_DEBUG " %s zone: %lu pages, LIFO batch:%lu\n", + zone->name, zone->present_pages, batch); } - if (zone->present_pages) - printk(KERN_DEBUG " %s zone: %lu pages, LIFO batch:%lu\n", - zone->name, zone->present_pages, batch); } __meminit int init_currently_empty_zone(struct zone *zone, @@ -3700,7 +3719,6 @@ static void __paginginit free_area_init_ zone->prev_priority = DEF_PRIORITY; - zone_pcp_init(zone); for_each_lru(l) { INIT_LIST_HEAD(&zone->lru[l].list); zone->lru[l].nr_saved_scan = 0; Index: linux-2.6/include/linux/mm.h =================================================================== --- linux-2.6.orig/include/linux/mm.h 2009-06-17 14:06:22.000000000 -0500 +++ linux-2.6/include/linux/mm.h 2009-06-17 14:11:30.000000000 -0500 @@ -1060,6 +1060,7 @@ extern void show_mem(void); extern void si_meminfo(struct sysinfo * val); extern void si_meminfo_node(struct sysinfo *val, int nid); extern int after_bootmem; +extern void setup_pagesets(void); #ifdef CONFIG_NUMA extern void setup_per_cpu_pageset(void); Index: linux-2.6/arch/ia64/kernel/setup.c =================================================================== --- linux-2.6.orig/arch/ia64/kernel/setup.c 2009-06-17 14:06:22.000000000 -0500 +++ linux-2.6/arch/ia64/kernel/setup.c 2009-06-17 14:11:30.000000000 -0500 @@ -859,6 +859,7 @@ void __init setup_per_cpu_areas (void) { /* start_kernel() requires this... */ + setup_pagesets(); } /* Index: linux-2.6/arch/powerpc/kernel/setup_64.c =================================================================== --- linux-2.6.orig/arch/powerpc/kernel/setup_64.c 2009-06-17 14:06:22.000000000 -0500 +++ linux-2.6/arch/powerpc/kernel/setup_64.c 2009-06-17 14:11:30.000000000 -0500 @@ -588,6 +588,7 @@ void __init setup_per_cpu_areas(void) paca[i].data_offset = ptr - __per_cpu_start; memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start); } + setup_pagesets(); } #endif Index: linux-2.6/arch/sparc/kernel/smp_64.c =================================================================== --- linux-2.6.orig/arch/sparc/kernel/smp_64.c 2009-06-17 14:06:22.000000000 -0500 +++ linux-2.6/arch/sparc/kernel/smp_64.c 2009-06-17 14:11:30.000000000 -0500 @@ -1543,4 +1543,5 @@ void __init setup_per_cpu_areas(void) of_fill_in_cpu_data(); if (tlb_type == hypervisor) mdesc_fill_in_cpu_data(cpu_all_mask); + setup_pagesets(); } Index: linux-2.6/arch/x86/kernel/setup_percpu.c =================================================================== --- linux-2.6.orig/arch/x86/kernel/setup_percpu.c 2009-06-17 14:06:22.000000000 -0500 +++ linux-2.6/arch/x86/kernel/setup_percpu.c 2009-06-17 14:11:30.000000000 -0500 @@ -438,4 +438,6 @@ void __init setup_per_cpu_areas(void) /* Setup cpu initialized, callin, callout masks */ setup_cpu_local_masks(); + + setup_pagesets(); } -- -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org