From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1945943Ab2EKRuJ (ORCPT ); Fri, 11 May 2012 13:50:09 -0400 Received: from mailhub.sw.ru ([195.214.232.25]:28608 "EHLO relay.sw.ru" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1161018Ab2EKRuC (ORCPT ); Fri, 11 May 2012 13:50:02 -0400 From: Glauber Costa To: Cc: , , , Tejun Heo , Li Zefan , Greg Thelen , Suleiman Souhlal , Michal Hocko , Johannes Weiner , , Frederic Weisbecker , Li Zefan , Aditya Kali , Oleg Nesterov , Kay Sievers , Tim Hockin , Tejun Heo , Andrew Morton Subject: [PATCH v2 11/29] cgroups: ability to stop res charge propagation on bounded ancestor Date: Fri, 11 May 2012 14:44:13 -0300 Message-Id: <1336758272-24284-12-git-send-email-glommer@parallels.com> X-Mailer: git-send-email 1.7.7.6 In-Reply-To: <1336758272-24284-1-git-send-email-glommer@parallels.com> References: <1336758272-24284-1-git-send-email-glommer@parallels.com> Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Frederic Weisbecker Moving a task from a cgroup to another may require to substract its resource charge from the old cgroup and add it to the new one. For this to happen, the uncharge/charge propagation can just stop when we reach the common ancestor for the two cgroups. Further the performance reasons, we also want to avoid to temporarily overload the common ancestors with a non-accurate resource counter usage if we charge first the new cgroup and uncharge the old one thereafter. This is going to be a requirement for the coming max number of task subsystem. To solve this, provide a pair of new API that can charge/uncharge a resource counter until we reach a given ancestor. Signed-off-by: Frederic Weisbecker Acked-by: Paul Menage Acked-by: Glauber Costa Cc: Li Zefan Cc: Johannes Weiner Cc: Aditya Kali Cc: Oleg Nesterov Cc: Kay Sievers Cc: Tim Hockin Cc: Tejun Heo Acked-by: Kirill A. Shutemov Signed-off-by: Andrew Morton --- Documentation/cgroups/resource_counter.txt | 18 +++++++++++++++++- include/linux/res_counter.h | 21 ++++++++++++++++++--- kernel/res_counter.c | 13 ++++++++----- 3 files changed, 43 insertions(+), 9 deletions(-) diff --git a/Documentation/cgroups/resource_counter.txt b/Documentation/cgroups/resource_counter.txt index 95b24d7..a2cd05b 100644 --- a/Documentation/cgroups/resource_counter.txt +++ b/Documentation/cgroups/resource_counter.txt @@ -83,7 +83,15 @@ to work with it. res_counter->lock internally (it must be called with res_counter->lock held). - e. void res_counter_uncharge[_locked] + e. int res_counter_charge_until(struct res_counter *counter, + struct res_counter *limit, unsigned long val, + struct res_counter **limit_fail_at) + + The same as res_counter_charge(), but the charge propagation to + the hierarchy stops at the limit given in the "limit" parameter. + + + f. void res_counter_uncharge[_locked] (struct res_counter *rc, unsigned long val) When a resource is released (freed) it should be de-accounted @@ -92,6 +100,14 @@ to work with it. The _locked routines imply that the res_counter->lock is taken. + + g. void res_counter_uncharge_until(struct res_counter *counter, + struct res_counter *limit, + unsigned long val) + + The same as res_counter_charge, but the uncharge propagation to + the hierarchy stops at the limit given in the "limit" parameter. + 2.1 Other accounting routines There are more routines that may help you with common needs, like diff --git a/include/linux/res_counter.h b/include/linux/res_counter.h index f7621cf..c12143e 100644 --- a/include/linux/res_counter.h +++ b/include/linux/res_counter.h @@ -117,11 +117,20 @@ void res_counter_init(struct res_counter *counter, struct res_counter *parent); int __must_check res_counter_charge_locked(struct res_counter *counter, unsigned long val); -int __must_check res_counter_charge(struct res_counter *counter, - unsigned long val, struct res_counter **limit_fail_at); +int __must_check res_counter_charge_until(struct res_counter *counter, + struct res_counter *limit, + unsigned long val, + struct res_counter **limit_fail_at); int res_counter_charge_nofail(struct res_counter *counter, unsigned long val, struct res_counter **limit_fail_at); +static inline int __must_check +res_counter_charge(struct res_counter *counter, unsigned long val, + struct res_counter **limit_fail_at) +{ + return res_counter_charge_until(counter, NULL, val, limit_fail_at); +} + /* * uncharge - tell that some portion of the resource is released * @@ -133,7 +142,13 @@ int res_counter_charge_nofail(struct res_counter *counter, */ void res_counter_uncharge_locked(struct res_counter *counter, unsigned long val); -void res_counter_uncharge(struct res_counter *counter, unsigned long val); +void res_counter_uncharge_until(struct res_counter *counter, + struct res_counter *limit, unsigned long val); +static inline void res_counter_uncharge(struct res_counter *counter, + unsigned long val) +{ + res_counter_uncharge_until(counter, NULL, val); +} /** * res_counter_margin - calculate chargeable space of a counter diff --git a/kernel/res_counter.c b/kernel/res_counter.c index d508363..7aebf96 100644 --- a/kernel/res_counter.c +++ b/kernel/res_counter.c @@ -35,8 +35,9 @@ int res_counter_charge_locked(struct res_counter *counter, unsigned long val) return 0; } -int res_counter_charge(struct res_counter *counter, unsigned long val, - struct res_counter **limit_fail_at) +int res_counter_charge_until(struct res_counter *counter, + struct res_counter *limit, unsigned long val, + struct res_counter **limit_fail_at) { int ret; unsigned long flags; @@ -44,7 +45,7 @@ int res_counter_charge(struct res_counter *counter, unsigned long val, *limit_fail_at = NULL; local_irq_save(flags); - for (c = counter; c != NULL; c = c->parent) { + for (c = counter; c != limit; c = c->parent) { spin_lock(&c->lock); ret = res_counter_charge_locked(c, val); spin_unlock(&c->lock); @@ -99,13 +100,15 @@ void res_counter_uncharge_locked(struct res_counter *counter, unsigned long val) counter->usage -= val; } -void res_counter_uncharge(struct res_counter *counter, unsigned long val) +void res_counter_uncharge_until(struct res_counter *counter, + struct res_counter *limit, + unsigned long val) { unsigned long flags; struct res_counter *c; local_irq_save(flags); - for (c = counter; c != NULL; c = c->parent) { + for (c = counter; c != limit; c = c->parent) { spin_lock(&c->lock); res_counter_uncharge_locked(c, val); spin_unlock(&c->lock); -- 1.7.7.6 From mboxrd@z Thu Jan 1 00:00:00 1970 From: Glauber Costa Subject: [PATCH v2 11/29] cgroups: ability to stop res charge propagation on bounded ancestor Date: Fri, 11 May 2012 14:44:13 -0300 Message-ID: <1336758272-24284-12-git-send-email-glommer@parallels.com> References: <1336758272-24284-1-git-send-email-glommer@parallels.com> Return-path: In-Reply-To: <1336758272-24284-1-git-send-email-glommer@parallels.com> Sender: owner-linux-mm@kvack.org List-ID: MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit To: linux-kernel@vger.kernel.org Cc: cgroups@vger.kernel.org, linux-mm@kvack.org, kamezawa.hiroyu@jp.fujitsu.com, Tejun Heo , Li Zefan , Greg Thelen , Suleiman Souhlal , Michal Hocko , Johannes Weiner , devel@openvz.org, Frederic Weisbecker , Li Zefan , Aditya Kali , Oleg Nesterov , Kay Sievers , Tim Hockin , Tejun Heo , Andrew Morton From: Frederic Weisbecker Moving a task from a cgroup to another may require to substract its resource charge from the old cgroup and add it to the new one. For this to happen, the uncharge/charge propagation can just stop when we reach the common ancestor for the two cgroups. Further the performance reasons, we also want to avoid to temporarily overload the common ancestors with a non-accurate resource counter usage if we charge first the new cgroup and uncharge the old one thereafter. This is going to be a requirement for the coming max number of task subsystem. To solve this, provide a pair of new API that can charge/uncharge a resource counter until we reach a given ancestor. Signed-off-by: Frederic Weisbecker Acked-by: Paul Menage Acked-by: Glauber Costa Cc: Li Zefan Cc: Johannes Weiner Cc: Aditya Kali Cc: Oleg Nesterov Cc: Kay Sievers Cc: Tim Hockin Cc: Tejun Heo Acked-by: Kirill A. Shutemov Signed-off-by: Andrew Morton --- Documentation/cgroups/resource_counter.txt | 18 +++++++++++++++++- include/linux/res_counter.h | 21 ++++++++++++++++++--- kernel/res_counter.c | 13 ++++++++----- 3 files changed, 43 insertions(+), 9 deletions(-) diff --git a/Documentation/cgroups/resource_counter.txt b/Documentation/cgroups/resource_counter.txt index 95b24d7..a2cd05b 100644 --- a/Documentation/cgroups/resource_counter.txt +++ b/Documentation/cgroups/resource_counter.txt @@ -83,7 +83,15 @@ to work with it. res_counter->lock internally (it must be called with res_counter->lock held). - e. void res_counter_uncharge[_locked] + e. int res_counter_charge_until(struct res_counter *counter, + struct res_counter *limit, unsigned long val, + struct res_counter **limit_fail_at) + + The same as res_counter_charge(), but the charge propagation to + the hierarchy stops at the limit given in the "limit" parameter. + + + f. void res_counter_uncharge[_locked] (struct res_counter *rc, unsigned long val) When a resource is released (freed) it should be de-accounted @@ -92,6 +100,14 @@ to work with it. The _locked routines imply that the res_counter->lock is taken. + + g. void res_counter_uncharge_until(struct res_counter *counter, + struct res_counter *limit, + unsigned long val) + + The same as res_counter_charge, but the uncharge propagation to + the hierarchy stops at the limit given in the "limit" parameter. + 2.1 Other accounting routines There are more routines that may help you with common needs, like diff --git a/include/linux/res_counter.h b/include/linux/res_counter.h index f7621cf..c12143e 100644 --- a/include/linux/res_counter.h +++ b/include/linux/res_counter.h @@ -117,11 +117,20 @@ void res_counter_init(struct res_counter *counter, struct res_counter *parent); int __must_check res_counter_charge_locked(struct res_counter *counter, unsigned long val); -int __must_check res_counter_charge(struct res_counter *counter, - unsigned long val, struct res_counter **limit_fail_at); +int __must_check res_counter_charge_until(struct res_counter *counter, + struct res_counter *limit, + unsigned long val, + struct res_counter **limit_fail_at); int res_counter_charge_nofail(struct res_counter *counter, unsigned long val, struct res_counter **limit_fail_at); +static inline int __must_check +res_counter_charge(struct res_counter *counter, unsigned long val, + struct res_counter **limit_fail_at) +{ + return res_counter_charge_until(counter, NULL, val, limit_fail_at); +} + /* * uncharge - tell that some portion of the resource is released * @@ -133,7 +142,13 @@ int res_counter_charge_nofail(struct res_counter *counter, */ void res_counter_uncharge_locked(struct res_counter *counter, unsigned long val); -void res_counter_uncharge(struct res_counter *counter, unsigned long val); +void res_counter_uncharge_until(struct res_counter *counter, + struct res_counter *limit, unsigned long val); +static inline void res_counter_uncharge(struct res_counter *counter, + unsigned long val) +{ + res_counter_uncharge_until(counter, NULL, val); +} /** * res_counter_margin - calculate chargeable space of a counter diff --git a/kernel/res_counter.c b/kernel/res_counter.c index d508363..7aebf96 100644 --- a/kernel/res_counter.c +++ b/kernel/res_counter.c @@ -35,8 +35,9 @@ int res_counter_charge_locked(struct res_counter *counter, unsigned long val) return 0; } -int res_counter_charge(struct res_counter *counter, unsigned long val, - struct res_counter **limit_fail_at) +int res_counter_charge_until(struct res_counter *counter, + struct res_counter *limit, unsigned long val, + struct res_counter **limit_fail_at) { int ret; unsigned long flags; @@ -44,7 +45,7 @@ int res_counter_charge(struct res_counter *counter, unsigned long val, *limit_fail_at = NULL; local_irq_save(flags); - for (c = counter; c != NULL; c = c->parent) { + for (c = counter; c != limit; c = c->parent) { spin_lock(&c->lock); ret = res_counter_charge_locked(c, val); spin_unlock(&c->lock); @@ -99,13 +100,15 @@ void res_counter_uncharge_locked(struct res_counter *counter, unsigned long val) counter->usage -= val; } -void res_counter_uncharge(struct res_counter *counter, unsigned long val) +void res_counter_uncharge_until(struct res_counter *counter, + struct res_counter *limit, + unsigned long val) { unsigned long flags; struct res_counter *c; local_irq_save(flags); - for (c = counter; c != NULL; c = c->parent) { + for (c = counter; c != limit; c = c->parent) { spin_lock(&c->lock); res_counter_uncharge_locked(c, val); spin_unlock(&c->lock); -- 1.7.7.6 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: email@kvack.org