From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail191.messagelabs.com (mail191.messagelabs.com [216.82.242.19]) by kanga.kvack.org (Postfix) with SMTP id 7D1526B0078 for ; Thu, 16 Sep 2010 22:34:56 -0400 (EDT) Date: Fri, 17 Sep 2010 10:34:57 +0800 From: Shaohua Li Subject: Re: [RFC]pagealloc: compensate a task for direct page reclaim Message-ID: <20100917023457.GA26307@sli10-conroe.sh.intel.com> References: <1284636396.1726.5.camel@shli-laptop> <20100916150009.GD16115@barrios-desktop> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20100916150009.GD16115@barrios-desktop> Sender: owner-linux-mm@kvack.org To: Minchan Kim Cc: "linux-mm@kvack.org" , Andrew Morton , Mel Gorman List-ID: On Thu, Sep 16, 2010 at 11:00:10PM +0800, Minchan Kim wrote: > On Thu, Sep 16, 2010 at 07:26:36PM +0800, Shaohua Li wrote: > > A task enters into direct page reclaim, free some memory. But sometimes > > the task can't get a free page after direct page reclaim because > > other tasks take them (this is quite common in a multi-task workload > > in my test). This behavior will bring extra latency to the task and is > > unfair. Since the task already gets penalty, we'd better give it a compensation. > > If a task frees some pages from direct page reclaim, we cache one freed page, > > and the task will get it soon. We only consider order 0 allocation, because > > it's hard to cache order > 0 page. > > > > Below is a trace output when a task frees some pages in try_to_free_pages(), but > > get_page_from_freelist() can't get a page in direct page reclaim. > > > > <...>-809 [004] 730.218991: __alloc_pages_nodemask: progress 147, order 0, pid 809, comm mmap_test > > <...>-806 [001] 730.237969: __alloc_pages_nodemask: progress 147, order 0, pid 806, comm mmap_test > > <...>-810 [005] 730.237971: __alloc_pages_nodemask: progress 147, order 0, pid 810, comm mmap_test > > <...>-809 [004] 730.237972: __alloc_pages_nodemask: progress 147, order 0, pid 809, comm mmap_test > > <...>-811 [006] 730.241409: __alloc_pages_nodemask: progress 147, order 0, pid 811, comm mmap_test > > <...>-809 [004] 730.241412: __alloc_pages_nodemask: progress 147, order 0, pid 809, comm mmap_test > > <...>-812 [007] 730.241435: __alloc_pages_nodemask: progress 147, order 0, pid 812, comm mmap_test > > <...>-809 [004] 730.245036: __alloc_pages_nodemask: progress 147, order 0, pid 809, comm mmap_test > > <...>-809 [004] 730.260360: __alloc_pages_nodemask: progress 147, order 0, pid 809, comm mmap_test > > <...>-805 [000] 730.260362: __alloc_pages_nodemask: progress 147, order 0, pid 805, comm mmap_test > > <...>-811 [006] 730.263877: __alloc_pages_nodemask: progress 147, order 0, pid 811, comm mmap_test > > > > The idea is good. > > I think we need to reserve at least one page for direct reclaimer who make the effort so that > it can reduce latency of stalled process. > > But I don't like this implementation. > > 1. It selects random page of reclaimed pages as cached page. > This doesn't consider requestor's migratetype so that it causes fragment problem in future. maybe we can limit the migratetype to MIGRATE_MOVABLE, which is the most common case. > 2. It skips buddy allocator. It means we lost coalescence chance so that fragement problem > would be severe than old. we only cache order 0 allocation, which doesn't enter lumpy reclaim, so this sounds not an issue to me. > In addition, I think this patch needs some number about enhancing of latency > and fragmentation if you are going with this approach. ok, sure. Thanks, Shaohua -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org