From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1757813AbZDUTkb (ORCPT ); Tue, 21 Apr 2009 15:40:31 -0400 Received: (majordomo@vger.kernel.org) by vger.kernel.org id S1755157AbZDUTkM (ORCPT ); Tue, 21 Apr 2009 15:40:12 -0400 Received: from mx3.mail.elte.hu ([157.181.1.138]:42175 "EHLO mx3.mail.elte.hu" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752484AbZDUTkL (ORCPT ); Tue, 21 Apr 2009 15:40:11 -0400 Date: Tue, 21 Apr 2009 21:39:24 +0200 From: Ingo Molnar To: Stephen Hemminger Cc: Linus Torvalds , Paul Mackerras , paulmck@linux.vnet.ibm.com, Eric Dumazet , Evgeniy Polyakov , David Miller , kaber@trash.net, jeff.chua.linux@gmail.com, laijs@cn.fujitsu.com, jengelh@medozas.de, r000n@r000n.net, linux-kernel@vger.kernel.org, netfilter-devel@vger.kernel.org, netdev@vger.kernel.org, benh@kernel.crashing.org, mathieu.desnoyers@polymtl.ca Subject: Re: [PATCH] netfilter: use per-cpu recursive lock (v11) Message-ID: <20090421193924.GA24404@elte.hu> References: <20090418094001.GA2369@ioremap.net> <20090418141455.GA7082@linux.vnet.ibm.com> <20090420103414.1b4c490f@nehalam> <49ECBE0A.7010303@cosmosbay.com> <18924.59347.375292.102385@cargo.ozlabs.ibm.com> <20090420215827.GK6822@linux.vnet.ibm.com> <18924.64032.103954.171918@cargo.ozlabs.ibm.com> <20090420160121.268a8226@nehalam> <20090421111541.228e977a@nehalam> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20090421111541.228e977a@nehalam> User-Agent: Mutt/1.5.18 (2008-05-17) X-ELTE-VirusStatus: clean X-ELTE-SpamScore: -1.5 X-ELTE-SpamLevel: X-ELTE-SpamCheck: no X-ELTE-SpamVersion: ELTE 2.0 X-ELTE-SpamCheck-Details: score=-1.5 required=5.9 tests=BAYES_00 autolearn=no SpamAssassin version=3.2.3 -1.5 BAYES_00 BODY: Bayesian spam probability is 0 to 1% [score: 0.0000] Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org * Stephen Hemminger wrote: > +void xt_info_wrunlock_bh(void) > + __releases(xt_info_lock) > +{ > + unsigned int i; > + > + for_each_possible_cpu(i) { > +#if NR_CPUS > (PREEMPT_MASK - 1) > + /* > + * Spin_unlock calls preempt_enable, but since we had > + * to adjust the count in xt_info_wrlock_bh, do it again > + */ > + preempt_disable(); > +#endif > + write_unlock(&per_cpu(xt_info_locks, i)); > + } > + local_bh_enable(); > +} In the global/local lock scheme i proposed this would become: global_write_unlock(void) { write_unlock(&global_lock); } As we dont hold the local locks during the write-locked critical section. No loop needed over CPUs, no preempt nesting complications, no lockdep complications, etc. Ingo