All of lore.kernel.org
 help / color / mirror / Atom feed
From: Muchun Song <songmuchun@bytedance.com>
To: Waiman Long <longman@redhat.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>,
	mhocko@kernel.org, roman.gushchin@linux.dev, shakeelb@google.com,
	cgroups@vger.kernel.org, linux-mm@kvack.org,
	linux-kernel@vger.kernel.org, duanxiongchun@bytedance.com
Subject: Re: [PATCH v4 03/11] mm: memcontrol: make lruvec lock safe when LRU pages are reparented
Date: Fri, 27 May 2022 10:55:30 +0800	[thread overview]
Message-ID: <YpA9omBZ8O4+szbY@FVFYT0MHHV2J.usts.net> (raw)
In-Reply-To: <9fe57cf7-9d21-3f91-ef27-e046b426c219@redhat.com>

On Thu, May 26, 2022 at 04:17:27PM -0400, Waiman Long wrote:
> On 5/25/22 11:38, Muchun Song wrote:
> > On Wed, May 25, 2022 at 10:48:54AM -0400, Johannes Weiner wrote:
> > > On Wed, May 25, 2022 at 09:03:59PM +0800, Muchun Song wrote:
> > > > On Wed, May 25, 2022 at 08:30:15AM -0400, Johannes Weiner wrote:
> > > > > On Wed, May 25, 2022 at 05:53:30PM +0800, Muchun Song wrote:
> > > > > > On Tue, May 24, 2022 at 03:27:20PM -0400, Johannes Weiner wrote:
> > > > > > > On Tue, May 24, 2022 at 02:05:43PM +0800, Muchun Song wrote:
> > > > > > > > @@ -1230,10 +1213,23 @@ void lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio)
> > > > > > > >    */
> > > > > > > >   struct lruvec *folio_lruvec_lock(struct folio *folio)
> > > > > > > >   {
> > > > > > > > -	struct lruvec *lruvec = folio_lruvec(folio);
> > > > > > > > +	struct lruvec *lruvec;
> > > > > > > > +	rcu_read_lock();
> > > > > > > > +retry:
> > > > > > > > +	lruvec = folio_lruvec(folio);
> > > > > > > >   	spin_lock(&lruvec->lru_lock);
> > > > > > > > -	lruvec_memcg_debug(lruvec, folio);
> > > > > > > > +
> > > > > > > > +	if (unlikely(lruvec_memcg(lruvec) != folio_memcg(folio))) {
> > > > > > > > +		spin_unlock(&lruvec->lru_lock);
> > > > > > > > +		goto retry;
> > > > > > > > +	}
> > > > > > > > +
> > > > > > > > +	/*
> > > > > > > > +	 * Preemption is disabled in the internal of spin_lock, which can serve
> > > > > > > > +	 * as RCU read-side critical sections.
> > > > > > > > +	 */
> > > > > > > > +	rcu_read_unlock();
> > > > > > > The code looks right to me, but I don't understand the comment: why do
> > > > > > > we care that the rcu read-side continues? With the lru_lock held,
> > > > > > > reparenting is on hold and the lruvec cannot be rcu-freed anyway, no?
> > > > > > > 
> > > > > > Right. We could hold rcu read lock until end of reparting.  So you mean
> > > > > > we do rcu_read_unlock in folio_lruvec_lock()?
> > > > > The comment seems to suggest that disabling preemption is what keeps
> > > > > the lruvec alive. But it's the lru_lock that keeps it alive. The
> > > > > cgroup destruction path tries to take the lru_lock long before it even
> > > > > gets to synchronize_rcu(). Once you hold the lru_lock, having an
> > > > > implied read-side critical section as well doesn't seem to matter.
> > > > > 
> > > > Well, I thought that spinlocks have implicit read-side critical sections
> > > > because it disables preemption (I learned from the comments above
> > > > synchronize_rcu() that says interrupts, preemption, or softirqs have been
> > > > disabled also serve as RCU read-side critical sections).  So I have a
> > > > question: is it still true in a PREEMPT_RT kernel (I am not familiar with
> > > > this)?
> > > Yes, but you're missing my point.
> > > 
> > > > > Should the comment be deleted?
> > > > I think we could remove the comments. If the above question is false, seems
> > > > like we should continue holding rcu read lock.
> > > It's true.
> > > 
> > Thanks for your answer.
> > 
> > > But assume it's false for a second. Why would you need to continue
> > > holding it? What would it protect? The lruvec would be pinned by the
> > > spinlock even if it DIDN'T imply an RCU lock, right?
> > > 
> > > So I don't understand the point of the comment. If the implied RCU
> > > lock is protecting something not covered by the bare spinlock itself,
> > > it should be added to the comment. Otherwise, the comment should go.
> > > 
> > Got it. Thanks for your nice explanation. I'll remove
> > the comment here.
> 
> Note that there is a similar comment in patch 6 which may have to be removed
> as well.
>

I have noticed that. Thank you for remindering me as well.
 

WARNING: multiple messages have this Message-ID (diff)
From: Muchun Song <songmuchun-EC8Uxl6Npydl57MIdRCFDg@public.gmane.org>
To: Waiman Long <longman-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
Cc: Johannes Weiner <hannes-druUgvl0LCNAfugRpC6u6w@public.gmane.org>,
	mhocko-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org,
	roman.gushchin-fxUVXftIFDnyG1zEObXtfA@public.gmane.org,
	shakeelb-hpIqsD4AKlfQT0dZR+AlfA@public.gmane.org,
	cgroups-u79uwXL29TY76Z2rM5mHXA@public.gmane.org,
	linux-mm-Bw31MaZKKs3YtjvyW6yDsg@public.gmane.org,
	linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org,
	duanxiongchun-EC8Uxl6Npydl57MIdRCFDg@public.gmane.org
Subject: Re: [PATCH v4 03/11] mm: memcontrol: make lruvec lock safe when LRU pages are reparented
Date: Fri, 27 May 2022 10:55:30 +0800	[thread overview]
Message-ID: <YpA9omBZ8O4+szbY@FVFYT0MHHV2J.usts.net> (raw)
In-Reply-To: <9fe57cf7-9d21-3f91-ef27-e046b426c219-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>

On Thu, May 26, 2022 at 04:17:27PM -0400, Waiman Long wrote:
> On 5/25/22 11:38, Muchun Song wrote:
> > On Wed, May 25, 2022 at 10:48:54AM -0400, Johannes Weiner wrote:
> > > On Wed, May 25, 2022 at 09:03:59PM +0800, Muchun Song wrote:
> > > > On Wed, May 25, 2022 at 08:30:15AM -0400, Johannes Weiner wrote:
> > > > > On Wed, May 25, 2022 at 05:53:30PM +0800, Muchun Song wrote:
> > > > > > On Tue, May 24, 2022 at 03:27:20PM -0400, Johannes Weiner wrote:
> > > > > > > On Tue, May 24, 2022 at 02:05:43PM +0800, Muchun Song wrote:
> > > > > > > > @@ -1230,10 +1213,23 @@ void lruvec_memcg_debug(struct lruvec *lruvec, struct folio *folio)
> > > > > > > >    */
> > > > > > > >   struct lruvec *folio_lruvec_lock(struct folio *folio)
> > > > > > > >   {
> > > > > > > > -	struct lruvec *lruvec = folio_lruvec(folio);
> > > > > > > > +	struct lruvec *lruvec;
> > > > > > > > +	rcu_read_lock();
> > > > > > > > +retry:
> > > > > > > > +	lruvec = folio_lruvec(folio);
> > > > > > > >   	spin_lock(&lruvec->lru_lock);
> > > > > > > > -	lruvec_memcg_debug(lruvec, folio);
> > > > > > > > +
> > > > > > > > +	if (unlikely(lruvec_memcg(lruvec) != folio_memcg(folio))) {
> > > > > > > > +		spin_unlock(&lruvec->lru_lock);
> > > > > > > > +		goto retry;
> > > > > > > > +	}
> > > > > > > > +
> > > > > > > > +	/*
> > > > > > > > +	 * Preemption is disabled in the internal of spin_lock, which can serve
> > > > > > > > +	 * as RCU read-side critical sections.
> > > > > > > > +	 */
> > > > > > > > +	rcu_read_unlock();
> > > > > > > The code looks right to me, but I don't understand the comment: why do
> > > > > > > we care that the rcu read-side continues? With the lru_lock held,
> > > > > > > reparenting is on hold and the lruvec cannot be rcu-freed anyway, no?
> > > > > > > 
> > > > > > Right. We could hold rcu read lock until end of reparting.  So you mean
> > > > > > we do rcu_read_unlock in folio_lruvec_lock()?
> > > > > The comment seems to suggest that disabling preemption is what keeps
> > > > > the lruvec alive. But it's the lru_lock that keeps it alive. The
> > > > > cgroup destruction path tries to take the lru_lock long before it even
> > > > > gets to synchronize_rcu(). Once you hold the lru_lock, having an
> > > > > implied read-side critical section as well doesn't seem to matter.
> > > > > 
> > > > Well, I thought that spinlocks have implicit read-side critical sections
> > > > because it disables preemption (I learned from the comments above
> > > > synchronize_rcu() that says interrupts, preemption, or softirqs have been
> > > > disabled also serve as RCU read-side critical sections).  So I have a
> > > > question: is it still true in a PREEMPT_RT kernel (I am not familiar with
> > > > this)?
> > > Yes, but you're missing my point.
> > > 
> > > > > Should the comment be deleted?
> > > > I think we could remove the comments. If the above question is false, seems
> > > > like we should continue holding rcu read lock.
> > > It's true.
> > > 
> > Thanks for your answer.
> > 
> > > But assume it's false for a second. Why would you need to continue
> > > holding it? What would it protect? The lruvec would be pinned by the
> > > spinlock even if it DIDN'T imply an RCU lock, right?
> > > 
> > > So I don't understand the point of the comment. If the implied RCU
> > > lock is protecting something not covered by the bare spinlock itself,
> > > it should be added to the comment. Otherwise, the comment should go.
> > > 
> > Got it. Thanks for your nice explanation. I'll remove
> > the comment here.
> 
> Note that there is a similar comment in patch 6 which may have to be removed
> as well.
>

I have noticed that. Thank you for remindering me as well.
 

  reply	other threads:[~2022-05-27  2:55 UTC|newest]

Thread overview: 93+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-05-24  6:05 [PATCH v4 00/11] Use obj_cgroup APIs to charge the LRU pages Muchun Song
2022-05-24  6:05 ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 01/11] mm: memcontrol: prepare objcg API for non-kmem usage Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:01   ` Johannes Weiner
2022-05-24 19:01     ` Johannes Weiner
2022-05-25  8:46     ` Muchun Song
2022-05-25  8:46       ` Muchun Song
2022-05-25  2:36   ` Roman Gushchin
2022-05-25  2:36     ` Roman Gushchin
2022-05-25  7:57     ` Muchun Song
2022-05-25  7:57       ` Muchun Song
2022-05-25 12:37       ` Johannes Weiner
2022-05-25 12:37         ` Johannes Weiner
2022-05-25 13:08         ` Muchun Song
2022-05-25 13:08           ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 02/11] mm: memcontrol: introduce compact_folio_lruvec_lock_irqsave Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:22   ` Johannes Weiner
2022-05-24 19:22     ` Johannes Weiner
2022-05-25  9:38     ` Muchun Song
2022-05-25  9:38       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 03/11] mm: memcontrol: make lruvec lock safe when LRU pages are reparented Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:23   ` Waiman Long
2022-05-25 10:20     ` Muchun Song
2022-05-25 10:20       ` Muchun Song
2022-05-25 14:59       ` Waiman Long
2022-05-25 14:59         ` Waiman Long
2022-05-24 19:27   ` Johannes Weiner
2022-05-24 19:27     ` Johannes Weiner
2022-05-25  9:53     ` Muchun Song
2022-05-25  9:53       ` Muchun Song
2022-05-25 12:30       ` Johannes Weiner
2022-05-25 12:30         ` Johannes Weiner
2022-05-25 13:03         ` Muchun Song
2022-05-25 13:03           ` Muchun Song
2022-05-25 14:48           ` Johannes Weiner
2022-05-25 14:48             ` Johannes Weiner
2022-05-25 15:38             ` Muchun Song
2022-05-25 15:38               ` Muchun Song
2022-05-26 20:17               ` Waiman Long
2022-05-26 20:17                 ` Waiman Long
2022-05-27  2:55                 ` Muchun Song [this message]
2022-05-27  2:55                   ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 04/11] mm: vmscan: rework move_pages_to_lru() Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:38   ` Johannes Weiner
2022-05-24 19:38     ` Johannes Weiner
2022-05-25 11:38     ` Muchun Song
2022-05-25 11:38       ` Muchun Song
2022-05-24 19:52   ` Waiman Long
2022-05-24 19:52     ` Waiman Long
2022-05-25 11:43     ` Muchun Song
2022-05-25 11:43       ` Muchun Song
2022-05-25  2:43   ` Roman Gushchin
2022-05-25  2:43     ` Roman Gushchin
2022-05-25 11:41     ` Muchun Song
2022-05-25 11:41       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 05/11] mm: thp: introduce folio_split_queue_lock{_irqsave}() Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 06/11] mm: thp: make split queue lock safe when LRU pages are reparented Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-25  2:54   ` Roman Gushchin
2022-05-25  2:54     ` Roman Gushchin
2022-05-25 11:44     ` Muchun Song
2022-05-25 11:44       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 07/11] mm: memcontrol: make all the callers of {folio,page}_memcg() safe Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-25  3:03   ` Roman Gushchin
2022-05-25  3:03     ` Roman Gushchin
2022-05-25 11:51     ` Muchun Song
2022-05-25 11:51       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 08/11] mm: memcontrol: introduce memcg_reparent_ops Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 09/11] mm: memcontrol: use obj_cgroup APIs to charge the LRU pages Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 12:29   ` kernel test robot
2022-05-24 18:16   ` kernel test robot
2022-05-24 18:16     ` kernel test robot
2022-05-25  7:14   ` [mm] bec0ae1210: WARNING:possible_recursive_locking_detected kernel test robot
2022-05-25  7:14     ` kernel test robot
2022-05-24  6:05 ` [PATCH v4 10/11] mm: lru: add VM_BUG_ON_FOLIO to lru maintenance function Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:44   ` Johannes Weiner
2022-05-24 19:44     ` Johannes Weiner
2022-05-25 11:59     ` Muchun Song
2022-05-25 11:59       ` Muchun Song
2022-05-25  2:40   ` Roman Gushchin
2022-05-25  2:40     ` Roman Gushchin
2022-05-25 11:58     ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 11/11] mm: lru: use lruvec lock to serialize memcg changes Muchun Song
2022-05-24  6:05   ` Muchun Song

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=YpA9omBZ8O4+szbY@FVFYT0MHHV2J.usts.net \
    --to=songmuchun@bytedance.com \
    --cc=cgroups@vger.kernel.org \
    --cc=duanxiongchun@bytedance.com \
    --cc=hannes@cmpxchg.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=longman@redhat.com \
    --cc=mhocko@kernel.org \
    --cc=roman.gushchin@linux.dev \
    --cc=shakeelb@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.