All of lore.kernel.org
 help / color / mirror / Atom feed
From: Andrea Parri <andrea.parri@amarulasolutions.com>
To: Will Deacon <will.deacon@arm.com>
Cc: Peter Zijlstra <peterz@infradead.org>,
	linux-kernel@vger.kernel.org,
	linux-arm-kernel@lists.infradead.org, mingo@kernel.org,
	boqun.feng@gmail.com, paulmck@linux.vnet.ibm.com,
	catalin.marinas@arm.com, Waiman Long <longman@redhat.com>
Subject: Re: [PATCH 10/10] locking/qspinlock: Elide back-to-back RELEASE operations with smp_wmb()
Date: Fri, 6 Apr 2018 17:49:44 +0200	[thread overview]
Message-ID: <20180406154944.GA14488@andrea> (raw)
In-Reply-To: <20180406152744.GE10528@arm.com>

On Fri, Apr 06, 2018 at 04:27:45PM +0100, Will Deacon wrote:
> Hi Andrea,
> 
> On Fri, Apr 06, 2018 at 03:05:12PM +0200, Andrea Parri wrote:
> > On Fri, Apr 06, 2018 at 12:34:36PM +0100, Will Deacon wrote:
> > > I could say something like:
> > > 
> > >   "Pairs with dependency ordering from both xchg_tail and explicit
> > >    dereferences of node->next"
> > > 
> > > but it's a bit cryptic :(
> > 
> > Agreed. ;)  It might be helpful to instead include a snippet to highlight
> > the interested memory accesses/dependencies; IIUC,
> > 
> > /*
> >  * Pairs with dependency ordering from both xchg_tail and explicit/?
> >  * dereferences of node->next:
> >  *
> >  *   CPU0
> >  *
> >  *   /* get node0, encode node0 in tail */
> >  *   pv_init_node(node0);
> >  *     ((struct pv_node *)node0)->cpu   = smp_processor_id();
> >  *     ((struct pv_node *)node0)->state = vcpu_running;
> 
> I'd probably ignore the PV case here and just focus on the native init
> of count/locked/next.
> 
> >  *   smp_wmb();
> >  *   old = xchg_tail(lock, tail);
> >  *
> >  *   CPU1:
> >  *
> >  *   /* get node1, encode tail from node1 */
> >  *   old = xchg_tail(lock, tail);   // = tail corresponding to node0
> >  *                                  // head an addr. dependency
> >  *   /* decode old in prev */
> >  *   pv_wait_node(node1, prev);
> 
> Similarly here -- the dependency is through decode_tail.
> 
> >  *     READ ((struct pv_node *)prev)->cpu   // addr. dependent read
> >  *     READ ((struct pv_node *)prev)->state // addr. dependend read
> >  *
> >  * [More details for the case "following our own ->next pointer" you
> >  *  mentioned dabove.]
> >  */
> > 
> > CPU1 would also have:
> > 
> >    WRITE_ONCE(prev->next, node1); // addr. dependent write
> > 
> > but I'm not sure how this pairs: does this belong to the the second
> > case above? can you elaborate on that?
> 
> This is dependent on the result of decode_tail, so it's still the first
> case. The second case is when we queued into an empty tail but somebody
> later queued behind us, so we don't find them until we're claiming the
> lock:
> 
>   if (!next)
>   	next = smp_cond_load_relaxed(&node->next, (VAL));
> 
>   arch_mcs_spin_unlock_contended(&next->locked);
> 
> here, this is all straightforward address dependencies rather than the
> arithmetic in decode_tail.

Got it. Thanks!

  Andrea


> 
> Will

WARNING: multiple messages have this Message-ID (diff)
From: andrea.parri@amarulasolutions.com (Andrea Parri)
To: linux-arm-kernel@lists.infradead.org
Subject: [PATCH 10/10] locking/qspinlock: Elide back-to-back RELEASE operations with smp_wmb()
Date: Fri, 6 Apr 2018 17:49:44 +0200	[thread overview]
Message-ID: <20180406154944.GA14488@andrea> (raw)
In-Reply-To: <20180406152744.GE10528@arm.com>

On Fri, Apr 06, 2018 at 04:27:45PM +0100, Will Deacon wrote:
> Hi Andrea,
> 
> On Fri, Apr 06, 2018 at 03:05:12PM +0200, Andrea Parri wrote:
> > On Fri, Apr 06, 2018 at 12:34:36PM +0100, Will Deacon wrote:
> > > I could say something like:
> > > 
> > >   "Pairs with dependency ordering from both xchg_tail and explicit
> > >    dereferences of node->next"
> > > 
> > > but it's a bit cryptic :(
> > 
> > Agreed. ;)  It might be helpful to instead include a snippet to highlight
> > the interested memory accesses/dependencies; IIUC,
> > 
> > /*
> >  * Pairs with dependency ordering from both xchg_tail and explicit/?
> >  * dereferences of node->next:
> >  *
> >  *   CPU0
> >  *
> >  *   /* get node0, encode node0 in tail */
> >  *   pv_init_node(node0);
> >  *     ((struct pv_node *)node0)->cpu   = smp_processor_id();
> >  *     ((struct pv_node *)node0)->state = vcpu_running;
> 
> I'd probably ignore the PV case here and just focus on the native init
> of count/locked/next.
> 
> >  *   smp_wmb();
> >  *   old = xchg_tail(lock, tail);
> >  *
> >  *   CPU1:
> >  *
> >  *   /* get node1, encode tail from node1 */
> >  *   old = xchg_tail(lock, tail);   // = tail corresponding to node0
> >  *                                  // head an addr. dependency
> >  *   /* decode old in prev */
> >  *   pv_wait_node(node1, prev);
> 
> Similarly here -- the dependency is through decode_tail.
> 
> >  *     READ ((struct pv_node *)prev)->cpu   // addr. dependent read
> >  *     READ ((struct pv_node *)prev)->state // addr. dependend read
> >  *
> >  * [More details for the case "following our own ->next pointer" you
> >  *  mentioned dabove.]
> >  */
> > 
> > CPU1 would also have:
> > 
> >    WRITE_ONCE(prev->next, node1); // addr. dependent write
> > 
> > but I'm not sure how this pairs: does this belong to the the second
> > case above? can you elaborate on that?
> 
> This is dependent on the result of decode_tail, so it's still the first
> case. The second case is when we queued into an empty tail but somebody
> later queued behind us, so we don't find them until we're claiming the
> lock:
> 
>   if (!next)
>   	next = smp_cond_load_relaxed(&node->next, (VAL));
> 
>   arch_mcs_spin_unlock_contended(&next->locked);
> 
> here, this is all straightforward address dependencies rather than the
> arithmetic in decode_tail.

Got it. Thanks!

  Andrea


> 
> Will

  reply	other threads:[~2018-04-06 15:49 UTC|newest]

Thread overview: 94+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-04-05 16:58 [PATCH 00/10] kernel/locking: qspinlock improvements Will Deacon
2018-04-05 16:58 ` Will Deacon
2018-04-05 16:58 ` [PATCH 01/10] locking/qspinlock: Don't spin on pending->locked transition in slowpath Will Deacon
2018-04-05 16:58   ` Will Deacon
2018-04-05 16:58 ` [PATCH 02/10] locking/qspinlock: Remove unbounded cmpxchg loop from locking slowpath Will Deacon
2018-04-05 16:58   ` Will Deacon
2018-04-05 17:07   ` Peter Zijlstra
2018-04-05 17:07     ` Peter Zijlstra
2018-04-06 15:08     ` Will Deacon
2018-04-06 15:08       ` Will Deacon
2018-04-05 17:13   ` Peter Zijlstra
2018-04-05 17:13     ` Peter Zijlstra
2018-04-05 21:16   ` Waiman Long
2018-04-05 21:16     ` Waiman Long
2018-04-06 15:08     ` Will Deacon
2018-04-06 15:08       ` Will Deacon
2018-04-06 20:50   ` Waiman Long
2018-04-06 20:50     ` Waiman Long
2018-04-06 21:09     ` Paul E. McKenney
2018-04-06 21:09       ` Paul E. McKenney
2018-04-07  8:47       ` Peter Zijlstra
2018-04-07  8:47         ` Peter Zijlstra
2018-04-07 23:37         ` Paul E. McKenney
2018-04-07 23:37           ` Paul E. McKenney
2018-04-09 10:58         ` Will Deacon
2018-04-09 10:58           ` Will Deacon
2018-04-07  9:07     ` Peter Zijlstra
2018-04-07  9:07       ` Peter Zijlstra
2018-04-09 10:58     ` Will Deacon
2018-04-09 10:58       ` Will Deacon
2018-04-09 14:54       ` Will Deacon
2018-04-09 14:54         ` Will Deacon
2018-04-09 15:54         ` Peter Zijlstra
2018-04-09 15:54           ` Peter Zijlstra
2018-04-09 17:19           ` Will Deacon
2018-04-09 17:19             ` Will Deacon
2018-04-10  9:35             ` Peter Zijlstra
2018-04-10  9:35               ` Peter Zijlstra
2018-09-20 16:08             ` Peter Zijlstra
2018-09-20 16:08               ` Peter Zijlstra
2018-09-20 16:22               ` Peter Zijlstra
2018-09-20 16:22                 ` Peter Zijlstra
2018-04-09 19:33         ` Waiman Long
2018-04-09 19:33           ` Waiman Long
2018-04-09 17:55       ` Waiman Long
2018-04-09 17:55         ` Waiman Long
2018-04-10 13:49   ` Sasha Levin
2018-04-10 13:49     ` Sasha Levin
2018-04-05 16:59 ` [PATCH 03/10] locking/qspinlock: Kill cmpxchg loop when claiming lock from head of queue Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 17:19   ` Peter Zijlstra
2018-04-05 17:19     ` Peter Zijlstra
2018-04-06 10:54     ` Will Deacon
2018-04-06 10:54       ` Will Deacon
2018-04-05 16:59 ` [PATCH 04/10] locking/qspinlock: Use atomic_cond_read_acquire Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 16:59 ` [PATCH 05/10] locking/mcs: Use smp_cond_load_acquire() in mcs spin loop Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 16:59 ` [PATCH 06/10] barriers: Introduce smp_cond_load_relaxed and atomic_cond_read_relaxed Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 17:22   ` Peter Zijlstra
2018-04-05 17:22     ` Peter Zijlstra
2018-04-06 10:55     ` Will Deacon
2018-04-06 10:55       ` Will Deacon
2018-04-05 16:59 ` [PATCH 07/10] locking/qspinlock: Use smp_cond_load_relaxed to wait for next node Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 16:59 ` [PATCH 08/10] locking/qspinlock: Merge struct __qspinlock into struct qspinlock Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-07  5:23   ` Boqun Feng
2018-04-07  5:23     ` Boqun Feng
2018-04-05 16:59 ` [PATCH 09/10] locking/qspinlock: Make queued_spin_unlock use smp_store_release Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 16:59 ` [PATCH 10/10] locking/qspinlock: Elide back-to-back RELEASE operations with smp_wmb() Will Deacon
2018-04-05 16:59   ` Will Deacon
2018-04-05 17:28   ` Peter Zijlstra
2018-04-05 17:28     ` Peter Zijlstra
2018-04-06 11:34     ` Will Deacon
2018-04-06 11:34       ` Will Deacon
2018-04-06 13:05       ` Andrea Parri
2018-04-06 13:05         ` Andrea Parri
2018-04-06 15:27         ` Will Deacon
2018-04-06 15:27           ` Will Deacon
2018-04-06 15:49           ` Andrea Parri [this message]
2018-04-06 15:49             ` Andrea Parri
2018-04-07  5:47   ` Boqun Feng
2018-04-07  5:47     ` Boqun Feng
2018-04-09 10:47     ` Will Deacon
2018-04-09 10:47       ` Will Deacon
2018-04-06 13:22 ` [PATCH 00/10] kernel/locking: qspinlock improvements Andrea Parri
2018-04-06 13:22   ` Andrea Parri
2018-04-11 10:20   ` Catalin Marinas
2018-04-11 10:20     ` Catalin Marinas
2018-04-11 15:39     ` Andrea Parri
2018-04-11 15:39       ` Andrea Parri

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20180406154944.GA14488@andrea \
    --to=andrea.parri@amarulasolutions.com \
    --cc=boqun.feng@gmail.com \
    --cc=catalin.marinas@arm.com \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=longman@redhat.com \
    --cc=mingo@kernel.org \
    --cc=paulmck@linux.vnet.ibm.com \
    --cc=peterz@infradead.org \
    --cc=will.deacon@arm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.