From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1754940Ab1A1IMP (ORCPT ); Fri, 28 Jan 2011 03:12:15 -0500 Received: from fgwmail5.fujitsu.co.jp ([192.51.44.35]:56771 "EHLO fgwmail5.fujitsu.co.jp" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754574Ab1A1IMO (ORCPT ); Fri, 28 Jan 2011 03:12:14 -0500 X-SecurityPolicyCheck-FJ: OK by FujitsuOutboundMailChecker v1.3.1 Date: Fri, 28 Jan 2011 17:06:10 +0900 From: KAMEZAWA Hiroyuki To: Johannes Weiner Cc: "linux-mm@kvack.org" , "linux-kernel@vger.kernel.org" , "nishimura@mxp.nes.nec.co.jp" , "balbir@linux.vnet.ibm.com" Subject: Re: [BUGFIX][PATCH 1/4] memcg: fix limit estimation at reclaim for hugepage Message-Id: <20110128170610.9ecb2ffa.kamezawa.hiroyu@jp.fujitsu.com> In-Reply-To: <20110128075215.GA2213@cmpxchg.org> References: <20110128122229.6a4c74a2.kamezawa.hiroyu@jp.fujitsu.com> <20110128122449.e4bb0e5f.kamezawa.hiroyu@jp.fujitsu.com> <20110128075215.GA2213@cmpxchg.org> Organization: FUJITSU Co. LTD. X-Mailer: Sylpheed 3.0.3 (GTK+ 2.10.14; i686-pc-mingw32) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, 28 Jan 2011 08:52:15 +0100 Johannes Weiner wrote: > > @@ -1853,7 +1869,14 @@ static int __mem_cgroup_do_charge(struct > > * Check the limit again to see if the reclaim reduced the > > * current usage of the cgroup before giving up > > */ > > - if (ret || mem_cgroup_check_under_limit(mem_over_limit)) > > + if (mem_cgroup_check_margin(mem_over_limit) >= csize) > > + return CHARGE_RETRY; > > + > > + /* > > + * If the charge size is a PAGE_SIZE, it's not hopeless while > > + * we can reclaim a page. > > + */ > > + if (csize == PAGE_SIZE && ret) > > return CHARGE_RETRY; > > That makes sense. > here. It's okay to add your Signed-off because this one is very similar to yours. == From: KAMEZAWA Hiroyuki Current memory cgroup's code tends to assume page_size == PAGE_SIZE and arrangement for THP is not enough yet. This is one of fixes for supporing THP. This adds mem_cgroup_get_margin() and checks whether there are required amount of free resource after memory reclaim. By this, THP page allocation can know whether it really succeeded or not and avoid infinite-loop and hangup. Total fixes for do_charge()/reclaim memory will follow this patch. Changelog: - rename check_margin -> get_margin. - style fix. Signed-off-by: KAMEZAWA Hiroyuki --- include/linux/res_counter.h | 11 +++++++++++ mm/memcontrol.c | 22 +++++++++++++++++++++- 2 files changed, 32 insertions(+), 1 deletion(-) Index: mmotm-0125/include/linux/res_counter.h =================================================================== --- mmotm-0125.orig/include/linux/res_counter.h +++ mmotm-0125/include/linux/res_counter.h @@ -182,6 +182,17 @@ static inline bool res_counter_check_und return ret; } +static inline s64 res_counter_get_margin(struct res_counter *cnt) +{ + s64 ret; + unsigned long flags; + + spin_lock_irqsave(&cnt->lock, flags); + ret = cnt->limit - cnt->usage; + spin_unlock_irqrestore(&cnt->lock, flags); + return ret; +} + static inline bool res_counter_check_under_soft_limit(struct res_counter *cnt) { bool ret; Index: mmotm-0125/mm/memcontrol.c =================================================================== --- mmotm-0125.orig/mm/memcontrol.c +++ mmotm-0125/mm/memcontrol.c @@ -1111,6 +1111,19 @@ static bool mem_cgroup_check_under_limit return false; } +static s64 mem_cgroup_get_margin(struct mem_cgroup *mem) +{ + s64 mem_margin = res_counter_get_margin(&mem->res); + s64 memsw_margin; + + if (do_swap_account) + memsw_margin = res_counter_get_margin(&mem->memsw); + else + memsw_margin = RESOURCE_MAX; + + return min(mem_margin, memsw_margin); +} + static unsigned int get_swappiness(struct mem_cgroup *memcg) { struct cgroup *cgrp = memcg->css.cgroup; @@ -1853,7 +1866,14 @@ static int __mem_cgroup_do_charge(struct * Check the limit again to see if the reclaim reduced the * current usage of the cgroup before giving up */ - if (ret || mem_cgroup_check_under_limit(mem_over_limit)) + if (mem_cgroup_get_margin(mem_over_limit) >= csize) + return CHARGE_RETRY; + + /* + * If the charge size is a PAGE_SIZE, it's not hopeless while + * we can reclaim a page. + */ + if (csize == PAGE_SIZE && ret) return CHARGE_RETRY; /* From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail143.messagelabs.com (mail143.messagelabs.com [216.82.254.35]) by kanga.kvack.org (Postfix) with SMTP id D80758D0039 for ; Fri, 28 Jan 2011 03:12:41 -0500 (EST) Received: from m4.gw.fujitsu.co.jp (unknown [10.0.50.74]) by fgwmail5.fujitsu.co.jp (Postfix) with ESMTP id 421B93EE0BD for ; Fri, 28 Jan 2011 17:12:12 +0900 (JST) Received: from smail (m4 [127.0.0.1]) by outgoing.m4.gw.fujitsu.co.jp (Postfix) with ESMTP id 239FF45DE4E for ; Fri, 28 Jan 2011 17:12:12 +0900 (JST) Received: from s4.gw.fujitsu.co.jp (s4.gw.fujitsu.co.jp [10.0.50.94]) by m4.gw.fujitsu.co.jp (Postfix) with ESMTP id 030B845DE4D for ; Fri, 28 Jan 2011 17:12:12 +0900 (JST) Received: from s4.gw.fujitsu.co.jp (localhost.localdomain [127.0.0.1]) by s4.gw.fujitsu.co.jp (Postfix) with ESMTP id E947F1DB803A for ; Fri, 28 Jan 2011 17:12:11 +0900 (JST) Received: from ml13.s.css.fujitsu.com (ml13.s.css.fujitsu.com [10.249.87.103]) by s4.gw.fujitsu.co.jp (Postfix) with ESMTP id A85A01DB8037 for ; Fri, 28 Jan 2011 17:12:11 +0900 (JST) Date: Fri, 28 Jan 2011 17:06:10 +0900 From: KAMEZAWA Hiroyuki Subject: Re: [BUGFIX][PATCH 1/4] memcg: fix limit estimation at reclaim for hugepage Message-Id: <20110128170610.9ecb2ffa.kamezawa.hiroyu@jp.fujitsu.com> In-Reply-To: <20110128075215.GA2213@cmpxchg.org> References: <20110128122229.6a4c74a2.kamezawa.hiroyu@jp.fujitsu.com> <20110128122449.e4bb0e5f.kamezawa.hiroyu@jp.fujitsu.com> <20110128075215.GA2213@cmpxchg.org> Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: owner-linux-mm@kvack.org To: Johannes Weiner Cc: "linux-mm@kvack.org" , "linux-kernel@vger.kernel.org" , "nishimura@mxp.nes.nec.co.jp" , "balbir@linux.vnet.ibm.com" List-ID: On Fri, 28 Jan 2011 08:52:15 +0100 Johannes Weiner wrote: > > @@ -1853,7 +1869,14 @@ static int __mem_cgroup_do_charge(struct > > * Check the limit again to see if the reclaim reduced the > > * current usage of the cgroup before giving up > > */ > > - if (ret || mem_cgroup_check_under_limit(mem_over_limit)) > > + if (mem_cgroup_check_margin(mem_over_limit) >= csize) > > + return CHARGE_RETRY; > > + > > + /* > > + * If the charge size is a PAGE_SIZE, it's not hopeless while > > + * we can reclaim a page. > > + */ > > + if (csize == PAGE_SIZE && ret) > > return CHARGE_RETRY; > > That makes sense. > here. It's okay to add your Signed-off because this one is very similar to yours. == From: KAMEZAWA Hiroyuki Current memory cgroup's code tends to assume page_size == PAGE_SIZE and arrangement for THP is not enough yet. This is one of fixes for supporing THP. This adds mem_cgroup_get_margin() and checks whether there are required amount of free resource after memory reclaim. By this, THP page allocation can know whether it really succeeded or not and avoid infinite-loop and hangup. Total fixes for do_charge()/reclaim memory will follow this patch. Changelog: - rename check_margin -> get_margin. - style fix. Signed-off-by: KAMEZAWA Hiroyuki --- include/linux/res_counter.h | 11 +++++++++++ mm/memcontrol.c | 22 +++++++++++++++++++++- 2 files changed, 32 insertions(+), 1 deletion(-) Index: mmotm-0125/include/linux/res_counter.h =================================================================== --- mmotm-0125.orig/include/linux/res_counter.h +++ mmotm-0125/include/linux/res_counter.h @@ -182,6 +182,17 @@ static inline bool res_counter_check_und return ret; } +static inline s64 res_counter_get_margin(struct res_counter *cnt) +{ + s64 ret; + unsigned long flags; + + spin_lock_irqsave(&cnt->lock, flags); + ret = cnt->limit - cnt->usage; + spin_unlock_irqrestore(&cnt->lock, flags); + return ret; +} + static inline bool res_counter_check_under_soft_limit(struct res_counter *cnt) { bool ret; Index: mmotm-0125/mm/memcontrol.c =================================================================== --- mmotm-0125.orig/mm/memcontrol.c +++ mmotm-0125/mm/memcontrol.c @@ -1111,6 +1111,19 @@ static bool mem_cgroup_check_under_limit return false; } +static s64 mem_cgroup_get_margin(struct mem_cgroup *mem) +{ + s64 mem_margin = res_counter_get_margin(&mem->res); + s64 memsw_margin; + + if (do_swap_account) + memsw_margin = res_counter_get_margin(&mem->memsw); + else + memsw_margin = RESOURCE_MAX; + + return min(mem_margin, memsw_margin); +} + static unsigned int get_swappiness(struct mem_cgroup *memcg) { struct cgroup *cgrp = memcg->css.cgroup; @@ -1853,7 +1866,14 @@ static int __mem_cgroup_do_charge(struct * Check the limit again to see if the reclaim reduced the * current usage of the cgroup before giving up */ - if (ret || mem_cgroup_check_under_limit(mem_over_limit)) + if (mem_cgroup_get_margin(mem_over_limit) >= csize) + return CHARGE_RETRY; + + /* + * If the charge size is a PAGE_SIZE, it's not hopeless while + * we can reclaim a page. + */ + if (csize == PAGE_SIZE && ret) return CHARGE_RETRY; /* -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom policy in Canada: sign http://dissolvethecrtc.ca/ Don't email: email@kvack.org