From: Mel Gorman <mel@csn.ul.ie> To: Mel Gorman <mel@csn.ul.ie>, Linux Memory Management List <linux-mm@kvack.org> Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>, Christoph Lameter <cl@linux-foundation.org>, Nick Piggin <npiggin@suse.de>, Linux Kernel Mailing List <linux-kernel@vger.kernel.org>, Lin Ming <ming.m.lin@intel.com>, Zhang Yanmin <yanmin_zhang@linux.intel.com>, Peter Zijlstra <peterz@infradead.org>, Pekka Enberg <penberg@cs.helsinki.fi>, Andrew Morton <akpm@linux-foundation.org> Subject: [PATCH 09/22] Calculate the alloc_flags for allocation only once Date: Wed, 22 Apr 2009 14:53:14 +0100 [thread overview] Message-ID: <1240408407-21848-10-git-send-email-mel@csn.ul.ie> (raw) In-Reply-To: <1240408407-21848-1-git-send-email-mel@csn.ul.ie> Factor out the mapping between GFP and alloc_flags only once. Once factored out, it only needs to be calculated once but some care must be taken. [neilb@suse.de says] As the test: - if (((p->flags & PF_MEMALLOC) || unlikely(test_thread_flag(TIF_MEMDIE))) - && !in_interrupt()) { - if (!(gfp_mask & __GFP_NOMEMALLOC)) { has been replaced with a slightly weaker one: + if (alloc_flags & ALLOC_NO_WATERMARKS) { Without care, this would allow recursion into the allocator via direct reclaim. This patch ensures we do not recurse when PF_MEMALLOC is set but TF_MEMDIE callers are now allowed to directly reclaim where they would have been prevented in the past. From: Peter Zijlstra <a.p.zijlstra@chello.nl> Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Acked-by: Pekka Enberg <penberg@cs.helsinki.fi> --- mm/page_alloc.c | 94 +++++++++++++++++++++++++++++------------------------- 1 files changed, 50 insertions(+), 44 deletions(-) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index eb1548c..0d23795 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -1577,15 +1577,6 @@ __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, return page; } -static inline int -is_allocation_high_priority(struct task_struct *p, gfp_t gfp_mask) -{ - if (((p->flags & PF_MEMALLOC) || unlikely(test_thread_flag(TIF_MEMDIE))) - && !in_interrupt()) - return 1; - return 0; -} - /* * This is called in the allocator slow-path if the allocation request is of * sufficient urgency to ignore watermarks and take other desperate measures @@ -1621,6 +1612,42 @@ void wake_all_kswapd(unsigned int order, struct zonelist *zonelist, wakeup_kswapd(zone, order); } +static inline int +gfp_to_alloc_flags(gfp_t gfp_mask) +{ + struct task_struct *p = current; + int alloc_flags = ALLOC_WMARK_MIN | ALLOC_CPUSET; + const gfp_t wait = gfp_mask & __GFP_WAIT; + + /* + * The caller may dip into page reserves a bit more if the caller + * cannot run direct reclaim, or if the caller has realtime scheduling + * policy or is asking for __GFP_HIGH memory. GFP_ATOMIC requests will + * set both ALLOC_HARDER (!wait) and ALLOC_HIGH (__GFP_HIGH). + */ + if (gfp_mask & __GFP_HIGH) + alloc_flags |= ALLOC_HIGH; + + if (!wait) { + alloc_flags |= ALLOC_HARDER; + /* + * Ignore cpuset if GFP_ATOMIC (!wait) rather than fail alloc. + * See also cpuset_zone_allowed() comment in kernel/cpuset.c. + */ + alloc_flags &= ~ALLOC_CPUSET; + } else if (unlikely(rt_task(p))) + alloc_flags |= ALLOC_HARDER; + + if (likely(!(gfp_mask & __GFP_NOMEMALLOC))) { + if (!in_interrupt() && + ((p->flags & PF_MEMALLOC) || + unlikely(test_thread_flag(TIF_MEMDIE)))) + alloc_flags |= ALLOC_NO_WATERMARKS; + } + + return alloc_flags; +} + static inline struct page * __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, @@ -1651,56 +1678,35 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, * OK, we're below the kswapd watermark and have kicked background * reclaim. Now things get more complex, so set up alloc_flags according * to how we want to proceed. - * - * The caller may dip into page reserves a bit more if the caller - * cannot run direct reclaim, or if the caller has realtime scheduling - * policy or is asking for __GFP_HIGH memory. GFP_ATOMIC requests will - * set both ALLOC_HARDER (!wait) and ALLOC_HIGH (__GFP_HIGH). */ - alloc_flags = ALLOC_WMARK_MIN; - if ((unlikely(rt_task(p)) && !in_interrupt()) || !wait) - alloc_flags |= ALLOC_HARDER; - if (gfp_mask & __GFP_HIGH) - alloc_flags |= ALLOC_HIGH; - if (wait) - alloc_flags |= ALLOC_CPUSET; + alloc_flags = gfp_to_alloc_flags(gfp_mask); restart: - /* - * Go through the zonelist again. Let __GFP_HIGH and allocations - * coming from realtime tasks go deeper into reserves. - * - * This is the last chance, in general, before the goto nopage. - * Ignore cpuset if GFP_ATOMIC (!wait) rather than fail alloc. - * See also cpuset_zone_allowed() comment in kernel/cpuset.c. - */ + /* This is the last chance, in general, before the goto nopage. */ page = get_page_from_freelist(gfp_mask, nodemask, order, zonelist, - high_zoneidx, alloc_flags, - preferred_zone, - migratetype); + high_zoneidx, alloc_flags & ~ALLOC_NO_WATERMARKS, + preferred_zone, migratetype); if (page) goto got_pg; rebalance: /* Allocate without watermarks if the context allows */ - if (is_allocation_high_priority(p, gfp_mask)) { - /* Do not dip into emergency reserves if specified */ - if (!(gfp_mask & __GFP_NOMEMALLOC)) { - page = __alloc_pages_high_priority(gfp_mask, order, - zonelist, high_zoneidx, nodemask, preferred_zone, - migratetype); - if (page) - goto got_pg; - } - - /* Ensure no recursion into the allocator */ - goto nopage; + if (alloc_flags & ALLOC_NO_WATERMARKS) { + page = __alloc_pages_high_priority(gfp_mask, order, + zonelist, high_zoneidx, nodemask, + preferred_zone, migratetype); + if (page) + goto got_pg; } /* Atomic allocations - we can't balance anything */ if (!wait) goto nopage; + /* Avoid recursion of direct reclaim */ + if (p->flags & PF_MEMALLOC) + goto nopage; + /* Try direct reclaim and then allocating */ page = __alloc_pages_direct_reclaim(gfp_mask, order, zonelist, high_zoneidx, -- 1.5.6.5
WARNING: multiple messages have this Message-ID (diff)
From: Mel Gorman <mel@csn.ul.ie> To: Mel Gorman <mel@csn.ul.ie>, Linux Memory Management List <linux-mm@kvack.org> Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>, Christoph Lameter <cl@linux-foundation.org>, Nick Piggin <npiggin@suse.de>, Linux Kernel Mailing List <linux-kernel@vger.kernel.org>, Lin Ming <ming.m.lin@intel.com>, Zhang Yanmin <yanmin_zhang@linux.intel.com>, Peter Zijlstra <peterz@infradead.org>, Pekka Enberg <penberg@cs.helsinki.fi>, Andrew Morton <akpm@linux-foundation.org> Subject: [PATCH 09/22] Calculate the alloc_flags for allocation only once Date: Wed, 22 Apr 2009 14:53:14 +0100 [thread overview] Message-ID: <1240408407-21848-10-git-send-email-mel@csn.ul.ie> (raw) In-Reply-To: <1240408407-21848-1-git-send-email-mel@csn.ul.ie> Factor out the mapping between GFP and alloc_flags only once. Once factored out, it only needs to be calculated once but some care must be taken. [neilb@suse.de says] As the test: - if (((p->flags & PF_MEMALLOC) || unlikely(test_thread_flag(TIF_MEMDIE))) - && !in_interrupt()) { - if (!(gfp_mask & __GFP_NOMEMALLOC)) { has been replaced with a slightly weaker one: + if (alloc_flags & ALLOC_NO_WATERMARKS) { Without care, this would allow recursion into the allocator via direct reclaim. This patch ensures we do not recurse when PF_MEMALLOC is set but TF_MEMDIE callers are now allowed to directly reclaim where they would have been prevented in the past. From: Peter Zijlstra <a.p.zijlstra@chello.nl> Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Acked-by: Pekka Enberg <penberg@cs.helsinki.fi> --- mm/page_alloc.c | 94 +++++++++++++++++++++++++++++------------------------- 1 files changed, 50 insertions(+), 44 deletions(-) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index eb1548c..0d23795 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -1577,15 +1577,6 @@ __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, return page; } -static inline int -is_allocation_high_priority(struct task_struct *p, gfp_t gfp_mask) -{ - if (((p->flags & PF_MEMALLOC) || unlikely(test_thread_flag(TIF_MEMDIE))) - && !in_interrupt()) - return 1; - return 0; -} - /* * This is called in the allocator slow-path if the allocation request is of * sufficient urgency to ignore watermarks and take other desperate measures @@ -1621,6 +1612,42 @@ void wake_all_kswapd(unsigned int order, struct zonelist *zonelist, wakeup_kswapd(zone, order); } +static inline int +gfp_to_alloc_flags(gfp_t gfp_mask) +{ + struct task_struct *p = current; + int alloc_flags = ALLOC_WMARK_MIN | ALLOC_CPUSET; + const gfp_t wait = gfp_mask & __GFP_WAIT; + + /* + * The caller may dip into page reserves a bit more if the caller + * cannot run direct reclaim, or if the caller has realtime scheduling + * policy or is asking for __GFP_HIGH memory. GFP_ATOMIC requests will + * set both ALLOC_HARDER (!wait) and ALLOC_HIGH (__GFP_HIGH). + */ + if (gfp_mask & __GFP_HIGH) + alloc_flags |= ALLOC_HIGH; + + if (!wait) { + alloc_flags |= ALLOC_HARDER; + /* + * Ignore cpuset if GFP_ATOMIC (!wait) rather than fail alloc. + * See also cpuset_zone_allowed() comment in kernel/cpuset.c. + */ + alloc_flags &= ~ALLOC_CPUSET; + } else if (unlikely(rt_task(p))) + alloc_flags |= ALLOC_HARDER; + + if (likely(!(gfp_mask & __GFP_NOMEMALLOC))) { + if (!in_interrupt() && + ((p->flags & PF_MEMALLOC) || + unlikely(test_thread_flag(TIF_MEMDIE)))) + alloc_flags |= ALLOC_NO_WATERMARKS; + } + + return alloc_flags; +} + static inline struct page * __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, struct zonelist *zonelist, enum zone_type high_zoneidx, @@ -1651,56 +1678,35 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, * OK, we're below the kswapd watermark and have kicked background * reclaim. Now things get more complex, so set up alloc_flags according * to how we want to proceed. - * - * The caller may dip into page reserves a bit more if the caller - * cannot run direct reclaim, or if the caller has realtime scheduling - * policy or is asking for __GFP_HIGH memory. GFP_ATOMIC requests will - * set both ALLOC_HARDER (!wait) and ALLOC_HIGH (__GFP_HIGH). */ - alloc_flags = ALLOC_WMARK_MIN; - if ((unlikely(rt_task(p)) && !in_interrupt()) || !wait) - alloc_flags |= ALLOC_HARDER; - if (gfp_mask & __GFP_HIGH) - alloc_flags |= ALLOC_HIGH; - if (wait) - alloc_flags |= ALLOC_CPUSET; + alloc_flags = gfp_to_alloc_flags(gfp_mask); restart: - /* - * Go through the zonelist again. Let __GFP_HIGH and allocations - * coming from realtime tasks go deeper into reserves. - * - * This is the last chance, in general, before the goto nopage. - * Ignore cpuset if GFP_ATOMIC (!wait) rather than fail alloc. - * See also cpuset_zone_allowed() comment in kernel/cpuset.c. - */ + /* This is the last chance, in general, before the goto nopage. */ page = get_page_from_freelist(gfp_mask, nodemask, order, zonelist, - high_zoneidx, alloc_flags, - preferred_zone, - migratetype); + high_zoneidx, alloc_flags & ~ALLOC_NO_WATERMARKS, + preferred_zone, migratetype); if (page) goto got_pg; rebalance: /* Allocate without watermarks if the context allows */ - if (is_allocation_high_priority(p, gfp_mask)) { - /* Do not dip into emergency reserves if specified */ - if (!(gfp_mask & __GFP_NOMEMALLOC)) { - page = __alloc_pages_high_priority(gfp_mask, order, - zonelist, high_zoneidx, nodemask, preferred_zone, - migratetype); - if (page) - goto got_pg; - } - - /* Ensure no recursion into the allocator */ - goto nopage; + if (alloc_flags & ALLOC_NO_WATERMARKS) { + page = __alloc_pages_high_priority(gfp_mask, order, + zonelist, high_zoneidx, nodemask, + preferred_zone, migratetype); + if (page) + goto got_pg; } /* Atomic allocations - we can't balance anything */ if (!wait) goto nopage; + /* Avoid recursion of direct reclaim */ + if (p->flags & PF_MEMALLOC) + goto nopage; + /* Try direct reclaim and then allocating */ page = __alloc_pages_direct_reclaim(gfp_mask, order, zonelist, high_zoneidx, -- 1.5.6.5 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2009-04-22 13:57 UTC|newest] Thread overview: 186+ messages / expand[flat|nested] mbox.gz Atom feed top 2009-04-22 13:53 [PATCH 00/22] Cleanup and optimise the page allocator V7 Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 01/22] Replace __alloc_pages_internal() with __alloc_pages_nodemask() Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 02/22] Do not sanity check order in the fast path Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 16:13 ` Dave Hansen 2009-04-22 16:13 ` Dave Hansen 2009-04-22 17:11 ` Mel Gorman 2009-04-22 17:11 ` Mel Gorman 2009-04-22 17:30 ` Dave Hansen 2009-04-22 17:30 ` Dave Hansen 2009-04-23 0:13 ` Mel Gorman 2009-04-23 0:13 ` Mel Gorman 2009-04-23 1:34 ` Dave Hansen 2009-04-23 1:34 ` Dave Hansen 2009-04-23 9:58 ` Mel Gorman 2009-04-23 9:58 ` Mel Gorman 2009-04-23 17:36 ` Dave Hansen 2009-04-23 17:36 ` Dave Hansen 2009-04-24 2:57 ` KOSAKI Motohiro 2009-04-24 2:57 ` KOSAKI Motohiro 2009-04-24 10:34 ` Mel Gorman 2009-04-24 10:34 ` Mel Gorman 2009-04-24 14:16 ` Dave Hansen 2009-04-24 14:16 ` Dave Hansen 2009-04-23 19:26 ` Dave Hansen 2009-04-23 19:26 ` Dave Hansen 2009-04-23 19:45 ` Dave Hansen 2009-04-23 19:45 ` Dave Hansen 2009-04-24 9:21 ` Mel Gorman 2009-04-24 9:21 ` Mel Gorman 2009-04-24 14:25 ` Dave Hansen 2009-04-24 14:25 ` Dave Hansen 2009-04-22 20:11 ` David Rientjes 2009-04-22 20:11 ` David Rientjes 2009-04-22 20:20 ` Christoph Lameter 2009-04-22 20:20 ` Christoph Lameter 2009-04-23 7:44 ` Pekka Enberg 2009-04-23 7:44 ` Pekka Enberg 2009-04-23 22:44 ` Andrew Morton 2009-04-23 22:44 ` Andrew Morton 2009-04-22 13:53 ` [PATCH 03/22] Do not check NUMA node ID when the caller knows the node is valid Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 04/22] Check only once if the zonelist is suitable for the allocation Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 05/22] Break up the allocator entry point into fast and slow paths Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 06/22] Move check for disabled anti-fragmentation out of fastpath Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 07/22] Calculate the preferred zone for allocation only once Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-23 22:48 ` Andrew Morton 2009-04-23 22:48 ` Andrew Morton 2009-04-22 13:53 ` [PATCH 08/22] Calculate the migratetype " Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` Mel Gorman [this message] 2009-04-22 13:53 ` [PATCH 09/22] Calculate the alloc_flags " Mel Gorman 2009-04-23 22:52 ` Andrew Morton 2009-04-23 22:52 ` Andrew Morton 2009-04-24 10:47 ` Mel Gorman 2009-04-24 10:47 ` Mel Gorman 2009-04-24 17:51 ` Andrew Morton 2009-04-24 17:51 ` Andrew Morton 2009-04-22 13:53 ` [PATCH 10/22] Remove a branch by assuming __GFP_HIGH == ALLOC_HIGH Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 11/22] Inline __rmqueue_smallest() Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 12/22] Inline buffered_rmqueue() Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 13/22] Inline __rmqueue_fallback() Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 14/22] Do not call get_pageblock_migratetype() more than necessary Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 15/22] Do not disable interrupts in free_page_mlock() Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-23 22:59 ` Andrew Morton 2009-04-23 22:59 ` Andrew Morton 2009-04-24 0:07 ` KOSAKI Motohiro 2009-04-24 0:07 ` KOSAKI Motohiro 2009-04-24 0:33 ` KOSAKI Motohiro 2009-04-24 0:33 ` KOSAKI Motohiro 2009-04-24 11:33 ` Mel Gorman 2009-04-24 11:33 ` Mel Gorman 2009-04-24 11:52 ` Lee Schermerhorn 2009-04-24 11:52 ` Lee Schermerhorn 2009-04-24 11:18 ` Mel Gorman 2009-04-24 11:18 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 16/22] Do not setup zonelist cache when there is only one node Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 20:24 ` David Rientjes 2009-04-22 20:24 ` David Rientjes 2009-04-22 20:32 ` Lee Schermerhorn 2009-04-22 20:32 ` Lee Schermerhorn 2009-04-22 20:34 ` David Rientjes 2009-04-22 20:34 ` David Rientjes 2009-04-23 0:11 ` KOSAKI Motohiro 2009-04-23 0:11 ` KOSAKI Motohiro 2009-04-23 0:19 ` Mel Gorman 2009-04-23 0:19 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 17/22] Do not check for compound pages during the page allocator sanity checks Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 18/22] Use allocation flags as an index to the zone watermark Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 17:11 ` Dave Hansen 2009-04-22 17:11 ` Dave Hansen 2009-04-22 17:14 ` Mel Gorman 2009-04-22 17:14 ` Mel Gorman 2009-04-22 17:47 ` Dave Hansen 2009-04-22 17:47 ` Dave Hansen 2009-04-23 0:27 ` KOSAKI Motohiro 2009-04-23 0:27 ` KOSAKI Motohiro 2009-04-23 10:03 ` Mel Gorman 2009-04-23 10:03 ` Mel Gorman 2009-04-24 6:41 ` KOSAKI Motohiro 2009-04-24 6:41 ` KOSAKI Motohiro 2009-04-22 20:06 ` David Rientjes 2009-04-22 20:06 ` David Rientjes 2009-04-23 0:29 ` Mel Gorman 2009-04-23 0:29 ` Mel Gorman 2009-04-27 17:00 ` [RFC] Replace the watermark-related union in struct zone with a watermark[] array Mel Gorman 2009-04-27 17:00 ` Mel Gorman 2009-04-27 20:48 ` David Rientjes 2009-04-27 20:48 ` David Rientjes 2009-04-27 20:54 ` Mel Gorman 2009-04-27 20:54 ` Mel Gorman 2009-04-27 20:51 ` Christoph Lameter 2009-04-27 20:51 ` Christoph Lameter 2009-04-27 21:04 ` David Rientjes 2009-04-27 21:04 ` David Rientjes 2009-04-30 13:35 ` Mel Gorman 2009-04-30 13:35 ` Mel Gorman 2009-04-30 13:48 ` Dave Hansen 2009-04-30 13:48 ` Dave Hansen 2009-05-12 14:13 ` [RFC] Replace the watermark-related union in struct zone with a watermark[] array V2 Mel Gorman 2009-05-12 14:13 ` Mel Gorman 2009-05-12 15:05 ` [RFC] Replace the watermark-related union in struct zone with awatermark[] " Dave Hansen 2009-05-12 15:05 ` Dave Hansen 2009-05-13 8:31 ` [RFC] Replace the watermark-related union in struct zone with a watermark[] " KOSAKI Motohiro 2009-05-13 8:31 ` KOSAKI Motohiro 2009-04-22 13:53 ` [PATCH 19/22] Update NR_FREE_PAGES only as necessary Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-23 23:06 ` Andrew Morton 2009-04-23 23:06 ` Andrew Morton 2009-04-23 23:04 ` Christoph Lameter 2009-04-23 23:04 ` Christoph Lameter 2009-04-24 13:06 ` Mel Gorman 2009-04-24 13:06 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 20/22] Get the pageblock migratetype without disabling interrupts Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 21/22] Use a pre-calculated value instead of num_online_nodes() in fast paths Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 23:04 ` David Rientjes 2009-04-22 23:04 ` David Rientjes 2009-04-23 0:44 ` Mel Gorman 2009-04-23 0:44 ` Mel Gorman 2009-04-23 19:29 ` David Rientjes 2009-04-23 19:29 ` David Rientjes 2009-04-24 13:31 ` [PATCH] Do not override definition of node_set_online() with macro Mel Gorman 2009-04-24 13:31 ` Mel Gorman 2009-04-22 13:53 ` [PATCH 22/22] slab: Use nr_online_nodes to check for a NUMA platform Mel Gorman 2009-04-22 13:53 ` Mel Gorman 2009-04-22 14:37 ` Pekka Enberg 2009-04-22 14:37 ` Pekka Enberg 2009-04-27 7:58 ` [PATCH 00/22] Cleanup and optimise the page allocator V7 Zhang, Yanmin 2009-04-27 7:58 ` Zhang, Yanmin 2009-04-27 14:38 ` Mel Gorman 2009-04-27 14:38 ` Mel Gorman 2009-04-28 1:59 ` Zhang, Yanmin 2009-04-28 1:59 ` Zhang, Yanmin 2009-04-28 10:27 ` Mel Gorman 2009-04-28 10:27 ` Mel Gorman 2009-04-28 10:31 ` [PATCH] Properly account for freed pages in free_pages_bulk() and when allocating high-order pages in buffered_rmqueue() Mel Gorman 2009-04-28 10:31 ` Mel Gorman 2009-04-28 16:37 ` Christoph Lameter 2009-04-28 16:37 ` Christoph Lameter 2009-04-28 16:51 ` Mel Gorman 2009-04-28 16:51 ` Mel Gorman 2009-04-28 17:15 ` Hugh Dickins 2009-04-28 17:15 ` Hugh Dickins 2009-04-28 18:07 ` [PATCH] Properly account for freed pages in free_pages_bulk() and when allocating high-order pages in buffered_rmqueue() V2 Mel Gorman 2009-04-28 18:07 ` Mel Gorman 2009-04-28 18:25 ` Hugh Dickins 2009-04-28 18:25 ` Hugh Dickins 2009-04-28 18:36 ` [PATCH] Properly account for freed pages in free_pages_bulk() and when allocating high-order pages in buffered_rmqueue() Mel Gorman 2009-04-28 18:36 ` Mel Gorman
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=1240408407-21848-10-git-send-email-mel@csn.ul.ie \ --to=mel@csn.ul.ie \ --cc=akpm@linux-foundation.org \ --cc=cl@linux-foundation.org \ --cc=kosaki.motohiro@jp.fujitsu.com \ --cc=linux-kernel@vger.kernel.org \ --cc=linux-mm@kvack.org \ --cc=ming.m.lin@intel.com \ --cc=npiggin@suse.de \ --cc=penberg@cs.helsinki.fi \ --cc=peterz@infradead.org \ --cc=yanmin_zhang@linux.intel.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.