From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756696Ab1EPXur (ORCPT ); Mon, 16 May 2011 19:50:47 -0400 Received: from mail-qy0-f174.google.com ([209.85.216.174]:54397 "EHLO mail-qy0-f174.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755862Ab1EPXup convert rfc822-to-8bit (ORCPT ); Mon, 16 May 2011 19:50:45 -0400 DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :cc:content-type:content-transfer-encoding; b=Yf/eJRNxEF84f4kl4ogMWp0witqC7WghI6xeyIX60MDCDlwRfoe8KceCJmWZzP5Bkj PXjuZA5D8XmIJc8q1fokqxcwhfv72meEfJmOl1wvtKJM6AsPNhhEO9gb7/3TONmEMj2D 1EQlaFmZQanF/JffTUYWrKO511O11pyD0oFy4= MIME-Version: 1.0 In-Reply-To: <20110516102753.GF5279@suse.de> References: <1305295404-12129-1-git-send-email-mgorman@suse.de> <1305295404-12129-5-git-send-email-mgorman@suse.de> <4DCFAA80.7040109@jp.fujitsu.com> <1305519711.4806.7.camel@mulgrave.site> <20110516084558.GE5279@suse.de> <20110516102753.GF5279@suse.de> Date: Tue, 17 May 2011 08:50:44 +0900 Message-ID: Subject: Re: [PATCH 4/4] mm: vmscan: If kswapd has been running too long, allow it to sleep From: Minchan Kim To: Mel Gorman Cc: James Bottomley , KOSAKI Motohiro , akpm@linux-foundation.org, colin.king@canonical.com, raghu.prabhu13@gmail.com, jack@suse.cz, chris.mason@oracle.com, cl@linux.com, penberg@kernel.org, riel@redhat.com, hannes@cmpxchg.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-ext4@vger.kernel.org Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8BIT Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, May 16, 2011 at 7:27 PM, Mel Gorman wrote: > On Mon, May 16, 2011 at 05:58:59PM +0900, Minchan Kim wrote: >> On Mon, May 16, 2011 at 5:45 PM, Mel Gorman wrote: >> > On Mon, May 16, 2011 at 02:04:00PM +0900, Minchan Kim wrote: >> >> On Mon, May 16, 2011 at 1:21 PM, James Bottomley >> >> wrote: >> >> > On Sun, 2011-05-15 at 19:27 +0900, KOSAKI Motohiro wrote: >> >> >> (2011/05/13 23:03), Mel Gorman wrote: >> >> >> > Under constant allocation pressure, kswapd can be in the situation where >> >> >> > sleeping_prematurely() will always return true even if kswapd has been >> >> >> > running a long time. Check if kswapd needs to be scheduled. >> >> >> > >> >> >> > Signed-off-by: Mel Gorman >> >> >> > --- >> >> >> >   mm/vmscan.c |    4 ++++ >> >> >> >   1 files changed, 4 insertions(+), 0 deletions(-) >> >> >> > >> >> >> > diff --git a/mm/vmscan.c b/mm/vmscan.c >> >> >> > index af24d1e..4d24828 100644 >> >> >> > --- a/mm/vmscan.c >> >> >> > +++ b/mm/vmscan.c >> >> >> > @@ -2251,6 +2251,10 @@ static bool sleeping_prematurely(pg_data_t *pgdat, int order, long remaining, >> >> >> >     unsigned long balanced = 0; >> >> >> >     bool all_zones_ok = true; >> >> >> > >> >> >> > +   /* If kswapd has been running too long, just sleep */ >> >> >> > +   if (need_resched()) >> >> >> > +           return false; >> >> >> > + >> >> >> >> >> >> Hmm... I don't like this patch so much. because this code does >> >> >> >> >> >> - don't sleep if kswapd got context switch at shrink_inactive_list >> >> > >> >> > This isn't entirely true:  need_resched() will be false, so we'll follow >> >> > the normal path for determining whether to sleep or not, in effect >> >> > leaving the current behaviour unchanged. >> >> > >> >> >> - sleep if kswapd didn't >> >> > >> >> > This also isn't entirely true: whether need_resched() is true at this >> >> > point depends on a whole lot more that whether we did a context switch >> >> > in shrink_inactive. It mostly depends on how long we've been running >> >> > without giving up the CPU.  Generally that will mean we've been round >> >> > the shrinker loop hundreds to thousands of times without sleeping. >> >> > >> >> >> It seems to be semi random behavior. >> >> > >> >> > Well, we have to do something.  Chris Mason first suspected the hang was >> >> > a kswapd rescheduling problem a while ago.  We tried putting >> >> > cond_rescheds() in several places in the vmscan code, but to no avail. >> >> >> >> Is it a result of  test with patch of Hannes(ie, !pgdat_balanced)? >> >> >> >> If it isn't, it would be nop regardless of putting cond_reshed at vmscan.c. >> >> Because, although we complete zone balancing, kswapd doesn't sleep as >> >> pgdat_balance returns wrong result. And at last VM calls >> >> balance_pgdat. In this case, balance_pgdat returns without any work as >> >> kswap couldn't find zones which have not enough free pages and goto >> >> out. kswapd could repeat this work infinitely. So you don't have a >> >> chance to call cond_resched. >> >> >> >> But if your test was with Hanne's patch, I am very curious how come >> >> kswapd consumes CPU a lot. >> >> >> >> > The need_resched() in sleeping_prematurely() seems to be about the best >> >> > option.  The other option might be just to put a cond_resched() in >> >> > kswapd_try_to_sleep(), but that will really have about the same effect. >> >> >> >> I don't oppose it but before that, I think we have to know why kswapd >> >> consumes CPU a lot although we applied Hannes' patch. >> >> >> > >> > Because it's still possible for processes to allocate pages at the same >> > rate kswapd is freeing them leading to a situation where kswapd does not >> > consider the zone balanced for prolonged periods of time. >> >> We have cond_resched in shrink_page_list, shrink_slab and balance_pgdat. >> So I think kswapd can be scheduled out although it's scheduled in >> after a short time as task scheduled also need page reclaim. Although >> all task in system need reclaim, kswapd cpu 99% consumption is a >> natural result, I think. >> Do I miss something? >> > > Lets see; > > shrink_page_list() only applies if inactive pages were isolated >        which in turn may not happen if all_unreclaimable is set in >        shrink_zones(). If for whatver reason, all_unreclaimable is >        set on all zones, we can miss calling cond_resched(). > > shrink_slab only applies if we are reclaiming slab pages. If the first >        shrinker returns -1, we do not call cond_resched(). If that >        first shrinker is dcache and __GFP_FS is not set, direct >        reclaimers will not shrink at all. However, if there are >        enough of them running or if one of the other shrinkers >        is running for a very long time, kswapd could be starved >        acquiring the shrinker_rwsem and never reaching the >        cond_resched(). Don't we have to move cond_resched? diff --git a/mm/vmscan.c b/mm/vmscan.c index 292582c..633e761 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -231,8 +231,10 @@ unsigned long shrink_slab(struct shrink_control *shrink, if (scanned == 0) scanned = SWAP_CLUSTER_MAX; - if (!down_read_trylock(&shrinker_rwsem)) - return 1; /* Assume we'll be able to shrink next time */ + if (!down_read_trylock(&shrinker_rwsem)) { + ret = 1; + goto out; /* Assume we'll be able to shrink next time */ + } list_for_each_entry(shrinker, &shrinker_list, list) { unsigned long long delta; @@ -280,12 +282,14 @@ unsigned long shrink_slab(struct shrink_control *shrink, count_vm_events(SLABS_SCANNED, this_scan); total_scan -= this_scan; - cond_resched(); } shrinker->nr += total_scan; + cond_resched(); } up_read(&shrinker_rwsem); +out: + cond_resched(); return ret; } > > balance_pgdat() only calls cond_resched if the zones are not >        balanced. For a high-order allocation that is balanced, it >        checks order-0 again. During that window, order-0 might have >        become unbalanced so it loops again for order-0 and returns >        that was reclaiming for order-0 to kswapd(). It can then find >        that a caller has rewoken kswapd for a high-order and re-enters >        balance_pgdat() without ever have called cond_resched(). If kswapd reclaims order-o followed by high order, it would have a chance to call cond_resched in shrink_page_list. But if all zones are all_unreclaimable is set, balance_pgdat could return any work. Okay. It does make sense. By your scenario, someone wakes up kswapd with higher order, again. So re-enters balance_pgdat without ever have called cond_resched. But if someone wakes up higher order again, we can't have a chance to call kswapd_try_to_sleep. So your patch effect would be nop, too. It would be better to put cond_resched after balance_pgdat? diff --git a/mm/vmscan.c b/mm/vmscan.c index 292582c..61c45d0 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -2753,6 +2753,7 @@ static int kswapd(void *p) if (!ret) { trace_mm_vmscan_kswapd_wake(pgdat->node_id, order); order = balance_pgdat(pgdat, order, &classzone_idx); + cond_resched(); } } return 0; > > While it appears unlikely, there are bad conditions which can result > in cond_resched() being avoided. > > -- > Mel Gorman > SUSE Labs > -- Kind regards, Minchan Kim From mboxrd@z Thu Jan 1 00:00:00 1970 From: Minchan Kim Subject: Re: [PATCH 4/4] mm: vmscan: If kswapd has been running too long, allow it to sleep Date: Tue, 17 May 2011 08:50:44 +0900 Message-ID: References: <1305295404-12129-1-git-send-email-mgorman@suse.de> <1305295404-12129-5-git-send-email-mgorman@suse.de> <4DCFAA80.7040109@jp.fujitsu.com> <1305519711.4806.7.camel@mulgrave.site> <20110516084558.GE5279@suse.de> <20110516102753.GF5279@suse.de> Mime-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: quoted-printable Cc: James Bottomley , KOSAKI Motohiro , akpm@linux-foundation.org, colin.king@canonical.com, raghu.prabhu13@gmail.com, jack@suse.cz, chris.mason@oracle.com, cl@linux.com, penberg@kernel.org, riel@redhat.com, hannes@cmpxchg.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-ext4@vger.kernel.org To: Mel Gorman Return-path: In-Reply-To: <20110516102753.GF5279@suse.de> Sender: owner-linux-mm@kvack.org List-Id: linux-fsdevel.vger.kernel.org On Mon, May 16, 2011 at 7:27 PM, Mel Gorman wrote: > On Mon, May 16, 2011 at 05:58:59PM +0900, Minchan Kim wrote: >> On Mon, May 16, 2011 at 5:45 PM, Mel Gorman wrote: >> > On Mon, May 16, 2011 at 02:04:00PM +0900, Minchan Kim wrote: >> >> On Mon, May 16, 2011 at 1:21 PM, James Bottomley >> >> wrote: >> >> > On Sun, 2011-05-15 at 19:27 +0900, KOSAKI Motohiro wrote: >> >> >> (2011/05/13 23:03), Mel Gorman wrote: >> >> >> > Under constant allocation pressure, kswapd can be in the situati= on where >> >> >> > sleeping_prematurely() will always return true even if kswapd ha= s been >> >> >> > running a long time. Check if kswapd needs to be scheduled. >> >> >> > >> >> >> > Signed-off-by: Mel Gorman >> >> >> > --- >> >> >> > =C2=A0 mm/vmscan.c | =C2=A0 =C2=A04 ++++ >> >> >> > =C2=A0 1 files changed, 4 insertions(+), 0 deletions(-) >> >> >> > >> >> >> > diff --git a/mm/vmscan.c b/mm/vmscan.c >> >> >> > index af24d1e..4d24828 100644 >> >> >> > --- a/mm/vmscan.c >> >> >> > +++ b/mm/vmscan.c >> >> >> > @@ -2251,6 +2251,10 @@ static bool sleeping_prematurely(pg_data_= t *pgdat, int order, long remaining, >> >> >> > =C2=A0 =C2=A0 unsigned long balanced =3D 0; >> >> >> > =C2=A0 =C2=A0 bool all_zones_ok =3D true; >> >> >> > >> >> >> > + =C2=A0 /* If kswapd has been running too long, just sleep */ >> >> >> > + =C2=A0 if (need_resched()) >> >> >> > + =C2=A0 =C2=A0 =C2=A0 =C2=A0 =C2=A0 return false; >> >> >> > + >> >> >> >> >> >> Hmm... I don't like this patch so much. because this code does >> >> >> >> >> >> - don't sleep if kswapd got context switch at shrink_inactive_list >> >> > >> >> > This isn't entirely true: =C2=A0need_resched() will be false, so we= 'll follow >> >> > the normal path for determining whether to sleep or not, in effect >> >> > leaving the current behaviour unchanged. >> >> > >> >> >> - sleep if kswapd didn't >> >> > >> >> > This also isn't entirely true: whether need_resched() is true at th= is >> >> > point depends on a whole lot more that whether we did a context swi= tch >> >> > in shrink_inactive. It mostly depends on how long we've been runnin= g >> >> > without giving up the CPU. =C2=A0Generally that will mean we've bee= n round >> >> > the shrinker loop hundreds to thousands of times without sleeping. >> >> > >> >> >> It seems to be semi random behavior. >> >> > >> >> > Well, we have to do something. =C2=A0Chris Mason first suspected th= e hang was >> >> > a kswapd rescheduling problem a while ago. =C2=A0We tried putting >> >> > cond_rescheds() in several places in the vmscan code, but to no ava= il. >> >> >> >> Is it a result of =C2=A0test with patch of Hannes(ie, !pgdat_balanced= )? >> >> >> >> If it isn't, it would be nop regardless of putting cond_reshed at vms= can.c. >> >> Because, although we complete zone balancing, kswapd doesn't sleep as >> >> pgdat_balance returns wrong result. And at last VM calls >> >> balance_pgdat. In this case, balance_pgdat returns without any work a= s >> >> kswap couldn't find zones which have not enough free pages and goto >> >> out. kswapd could repeat this work infinitely. So you don't have a >> >> chance to call cond_resched. >> >> >> >> But if your test was with Hanne's patch, I am very curious how come >> >> kswapd consumes CPU a lot. >> >> >> >> > The need_resched() in sleeping_prematurely() seems to be about the = best >> >> > option. =C2=A0The other option might be just to put a cond_resched(= ) in >> >> > kswapd_try_to_sleep(), but that will really have about the same eff= ect. >> >> >> >> I don't oppose it but before that, I think we have to know why kswapd >> >> consumes CPU a lot although we applied Hannes' patch. >> >> >> > >> > Because it's still possible for processes to allocate pages at the sam= e >> > rate kswapd is freeing them leading to a situation where kswapd does n= ot >> > consider the zone balanced for prolonged periods of time. >> >> We have cond_resched in shrink_page_list, shrink_slab and balance_pgdat. >> So I think kswapd can be scheduled out although it's scheduled in >> after a short time as task scheduled also need page reclaim. Although >> all task in system need reclaim, kswapd cpu 99% consumption is a >> natural result, I think. >> Do I miss something? >> > > Lets see; > > shrink_page_list() only applies if inactive pages were isolated > =C2=A0 =C2=A0 =C2=A0 =C2=A0which in turn may not happen if all_unreclaima= ble is set in > =C2=A0 =C2=A0 =C2=A0 =C2=A0shrink_zones(). If for whatver reason, all_unr= eclaimable is > =C2=A0 =C2=A0 =C2=A0 =C2=A0set on all zones, we can miss calling cond_res= ched(). > > shrink_slab only applies if we are reclaiming slab pages. If the first > =C2=A0 =C2=A0 =C2=A0 =C2=A0shrinker returns -1, we do not call cond_resch= ed(). If that > =C2=A0 =C2=A0 =C2=A0 =C2=A0first shrinker is dcache and __GFP_FS is not s= et, direct > =C2=A0 =C2=A0 =C2=A0 =C2=A0reclaimers will not shrink at all. However, if= there are > =C2=A0 =C2=A0 =C2=A0 =C2=A0enough of them running or if one of the other = shrinkers > =C2=A0 =C2=A0 =C2=A0 =C2=A0is running for a very long time, kswapd could = be starved > =C2=A0 =C2=A0 =C2=A0 =C2=A0acquiring the shrinker_rwsem and never reachin= g the > =C2=A0 =C2=A0 =C2=A0 =C2=A0cond_resched(). Don't we have to move cond_resched? diff --git a/mm/vmscan.c b/mm/vmscan.c index 292582c..633e761 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -231,8 +231,10 @@ unsigned long shrink_slab(struct shrink_control *shrin= k, if (scanned =3D=3D 0) scanned =3D SWAP_CLUSTER_MAX; - if (!down_read_trylock(&shrinker_rwsem)) - return 1; /* Assume we'll be able to shrink next time= */ + if (!down_read_trylock(&shrinker_rwsem)) { + ret =3D 1; + goto out; /* Assume we'll be able to shrink next time */ + } list_for_each_entry(shrinker, &shrinker_list, list) { unsigned long long delta; @@ -280,12 +282,14 @@ unsigned long shrink_slab(struct shrink_control *shri= nk, count_vm_events(SLABS_SCANNED, this_scan); total_scan -=3D this_scan; - cond_resched(); } shrinker->nr +=3D total_scan; + cond_resched(); } up_read(&shrinker_rwsem); +out: + cond_resched(); return ret; } > > balance_pgdat() only calls cond_resched if the zones are not > =C2=A0 =C2=A0 =C2=A0 =C2=A0balanced. For a high-order allocation that is = balanced, it > =C2=A0 =C2=A0 =C2=A0 =C2=A0checks order-0 again. During that window, orde= r-0 might have > =C2=A0 =C2=A0 =C2=A0 =C2=A0become unbalanced so it loops again for order-= 0 and returns > =C2=A0 =C2=A0 =C2=A0 =C2=A0that was reclaiming for order-0 to kswapd(). I= t can then find > =C2=A0 =C2=A0 =C2=A0 =C2=A0that a caller has rewoken kswapd for a high-or= der and re-enters > =C2=A0 =C2=A0 =C2=A0 =C2=A0balance_pgdat() without ever have called cond_= resched(). If kswapd reclaims order-o followed by high order, it would have a chance to call cond_resched in shrink_page_list. But if all zones are all_unreclaimable is set, balance_pgdat could return any work. Okay. It does make sense. By your scenario, someone wakes up kswapd with higher order, again. So re-enters balance_pgdat without ever have called cond_resched. But if someone wakes up higher order again, we can't have a chance to call kswapd_try_to_sleep. So your patch effect would be nop, too. It would be better to put cond_resched after balance_pgdat? diff --git a/mm/vmscan.c b/mm/vmscan.c index 292582c..61c45d0 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -2753,6 +2753,7 @@ static int kswapd(void *p) if (!ret) { trace_mm_vmscan_kswapd_wake(pgdat->node_id, order); order =3D balance_pgdat(pgdat, order, &classzone_id= x); + cond_resched(); } } return 0; > > While it appears unlikely, there are bad conditions which can result > in cond_resched() being avoided. > > -- > Mel Gorman > SUSE Labs > --=20 Kind regards, Minchan Kim -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: email@kvack.org