From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8652EC4332F for ; Wed, 21 Dec 2022 17:21:37 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232414AbiLURVf (ORCPT ); Wed, 21 Dec 2022 12:21:35 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:59892 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234953AbiLURU5 (ORCPT ); Wed, 21 Dec 2022 12:20:57 -0500 Received: from smtpout.efficios.com (unknown [IPv6:2607:5300:203:b2ee::31e5]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 87C5A26481; Wed, 21 Dec 2022 09:20:10 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=efficios.com; s=smtpout1; t=1671643209; bh=n/YGAUOunLlUSiLTzekbIHtWmgD98wLpsvCgDOv+sUA=; h=Date:Subject:To:Cc:References:From:In-Reply-To:From; b=m7UkjoEn+NoL7YN3Px0a2JFzA75foGkf4ODQZ68ThavqzywQ01kWUs16juIR1JM+R 17etM2pfHgilo6f4V+zTSVI0fRSPBSdSITfatkkrxEG8uPFVh3wFJMItqbZHesvZYC ubXuIroWh6VHAdnO6+4ILa+YgNmYLmvs4GSXney+/BgoGoNGhmBwF/yS83VbApTcVt cg6152LRMQnAhFmIRKfVGynsFnzdBQnkRuOYjWmttIZ9o9zaBSpP0/Je+UmEQzy1MX aFl4wLTEpTyY/btGjPv4c+ofgttKC5Z0Fuk57vAjolKQ79CDMSHBFQCoqeZxJYidY1 SV1d4D8XPu31g== Received: from [10.1.0.30] (192-222-188-97.qc.cable.ebox.net [192.222.188.97]) by smtpout.efficios.com (Postfix) with ESMTPSA id 4NcgDn12cLzbg1; Wed, 21 Dec 2022 12:20:09 -0500 (EST) Message-ID: Date: Wed, 21 Dec 2022 12:20:34 -0500 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Thunderbird/102.6.0 Subject: Re: [RFC 0/2] srcu: Remove pre-flip memory barrier Content-Language: en-US To: Frederic Weisbecker Cc: Joel Fernandes , linux-kernel@vger.kernel.org, Josh Triplett , Lai Jiangshan , "Paul E. McKenney" , rcu@vger.kernel.org, Steven Rostedt References: <20221220140714.GB22763@lothringen> <20221220224459.GA25175@lothringen> <20221221004957.GA29021@lothringen> <20221221005858.GA29316@lothringen> <20221221121118.GA35081@lothringen> From: Mathieu Desnoyers In-Reply-To: <20221221121118.GA35081@lothringen> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2022-12-21 07:11, Frederic Weisbecker wrote: > On Tue, Dec 20, 2022 at 10:43:25PM -0500, Mathieu Desnoyers wrote: >> On 2022-12-20 19:58, Frederic Weisbecker wrote: >>> On Wed, Dec 21, 2022 at 01:49:57AM +0100, Frederic Weisbecker wrote: >>>> On Tue, Dec 20, 2022 at 07:15:00PM -0500, Joel Fernandes wrote: >>>>> On Tue, Dec 20, 2022 at 5:45 PM Frederic Weisbecker wrote: >>>>> Agreed about (1). >>>>> >>>>>> _ In (2), E pairs with the address-dependency between idx and lock_count. >>>>> >>>>> But that is not the only reason. If that was the only reason for (2), >>>>> then there is an smp_mb() just before the next-scan post-flip before >>>>> the lock counts are read. >>>> >>>> The post-flip barrier makes sure the new idx is visible on the next READER's >>>> turn, but it doesn't protect against the fact that "READ idx then WRITE lock[idx]" >>>> may appear unordered from the update side POV if there is no barrier between the >>>> scan and the flip. >>>> >>>> If you remove the smp_mb() from the litmus test I sent, things explode. >>> >>> Or rather, look at it the other way, if there is no barrier between the lock >>> scan and the index flip (E), then the index flip can appear to be written before the >>> lock is read. Which means you may start activating the index before you finish >>> reading it (at least it appears that way from the readers pont of view). >> >> Considering that you can have pre-existing readers from arbitrary index >> appearing anywhere in the grace period (because a reader can fetch the >> index and be preempted for an arbitrary amount of time before incrementing >> the lock count), the grace period algorithm needs to deal with the fact that >> a newcoming reader can appear in a given index either before or after the >> flip. > > True but the number of preempted tasks is bound and there is a forward progress guarantee. > >> I don't see how flipping the index before or after loading the unlock/lock >> values would break anything (except for unlikely counter overflow situations >> as previously discussed). > > Forward progress guarantee. Considering a coherent cache, the store-buffer will ensure that the index flip eventually reaches all readers. This bounds the time during which readers can flood the current index, and therefore guarantees forward progress. AFAIK the Linux kernel does not support architectures with incoherent caches. So I still don't see how having the barrier before or after the index flip is useful for forward progress. Thanks, Mathieu > > Thanks. > >> >> Thanks, >> >> Mathieu >> >> -- >> Mathieu Desnoyers >> EfficiOS Inc. >> https://www.efficios.com >> -- Mathieu Desnoyers EfficiOS Inc. https://www.efficios.com