All of lore.kernel.org
 help / color / mirror / Atom feed
From: chengming.zhou@linux.dev
To: vbabka@suse.cz, cl@linux.com, penberg@kernel.org, willy@infradead.org
Cc: rientjes@google.com, iamjoonsoo.kim@lge.com,
	akpm@linux-foundation.org, roman.gushchin@linux.dev,
	42.hyeyoo@gmail.com, linux-mm@kvack.org,
	linux-kernel@vger.kernel.org, chengming.zhou@linux.dev,
	Chengming Zhou <zhouchengming@bytedance.com>
Subject: [RFC PATCH v4 3/9] slub: Keep track of whether slub is on the per-node partial list
Date: Tue, 31 Oct 2023 14:07:35 +0000	[thread overview]
Message-ID: <20231031140741.79387-4-chengming.zhou@linux.dev> (raw)
In-Reply-To: <20231031140741.79387-1-chengming.zhou@linux.dev>

From: Chengming Zhou <zhouchengming@bytedance.com>

Now we rely on the "frozen" bit to see if we should manipulate the
slab->slab_list, which will be changed in the following patch.

Instead we introduce another way to keep track of whether slub is on
the per-node partial list, here we reuse the PG_workingset bit.

We use __set_bit and __clear_bit directly instead of the atomic version
for better performance and it's safe since it's protected by the slub
node list_lock.

Suggested-by: Matthew Wilcox <willy@infradead.org>
Signed-off-by: Chengming Zhou <zhouchengming@bytedance.com>
---
 mm/slub.c | 22 ++++++++++++++++++++++
 1 file changed, 22 insertions(+)

diff --git a/mm/slub.c b/mm/slub.c
index 03384cd965c5..eed8ae0dbaf9 100644
--- a/mm/slub.c
+++ b/mm/slub.c
@@ -2116,6 +2116,25 @@ static void discard_slab(struct kmem_cache *s, struct slab *slab)
 	free_slab(s, slab);
 }
 
+/*
+ * SLUB reuses PG_workingset bit to keep track of whether it's on
+ * the per-node partial list.
+ */
+static inline bool slab_test_node_partial(const struct slab *slab)
+{
+	return folio_test_workingset((struct folio *)slab_folio(slab));
+}
+
+static inline void slab_set_node_partial(struct slab *slab)
+{
+	__set_bit(PG_workingset, folio_flags(slab_folio(slab), 0));
+}
+
+static inline void slab_clear_node_partial(struct slab *slab)
+{
+	__clear_bit(PG_workingset, folio_flags(slab_folio(slab), 0));
+}
+
 /*
  * Management of partially allocated slabs.
  */
@@ -2127,6 +2146,7 @@ __add_partial(struct kmem_cache_node *n, struct slab *slab, int tail)
 		list_add_tail(&slab->slab_list, &n->partial);
 	else
 		list_add(&slab->slab_list, &n->partial);
+	slab_set_node_partial(slab);
 }
 
 static inline void add_partial(struct kmem_cache_node *n,
@@ -2141,6 +2161,7 @@ static inline void remove_partial(struct kmem_cache_node *n,
 {
 	lockdep_assert_held(&n->list_lock);
 	list_del(&slab->slab_list);
+	slab_clear_node_partial(slab);
 	n->nr_partial--;
 }
 
@@ -4833,6 +4854,7 @@ static int __kmem_cache_do_shrink(struct kmem_cache *s)
 
 			if (free == slab->objects) {
 				list_move(&slab->slab_list, &discard);
+				slab_clear_node_partial(slab);
 				n->nr_partial--;
 				dec_slabs_node(s, node, slab->objects);
 			} else if (free <= SHRINK_PROMOTE_MAX)
-- 
2.20.1


  parent reply	other threads:[~2023-10-31 14:09 UTC|newest]

Thread overview: 23+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-10-31 14:07 [RFC PATCH v4 0/9] slub: Delay freezing of CPU partial slabs chengming.zhou
2023-10-31 14:07 ` [RFC PATCH v4 1/9] slub: Reflow ___slab_alloc() chengming.zhou
2023-10-31 14:07 ` [RFC PATCH v4 2/9] slub: Change get_partial() interfaces to return slab chengming.zhou
2023-10-31 14:07 ` chengming.zhou [this message]
2023-11-01 12:23   ` [RFC PATCH v4 3/9] slub: Keep track of whether slub is on the per-node partial list Vlastimil Babka
2023-10-31 14:07 ` [RFC PATCH v4 4/9] slub: Prepare __slab_free() for unfrozen partial slab out of node " chengming.zhou
2023-11-01 12:26   ` Vlastimil Babka
2023-10-31 14:07 ` [RFC PATCH v4 5/9] slub: Introduce freeze_slab() chengming.zhou
2023-12-03  6:08   ` Hyeonggon Yoo
2023-10-31 14:07 ` [RFC PATCH v4 6/9] slub: Delay freezing of partial slabs chengming.zhou
2023-10-31 14:07 ` [RFC PATCH v4 7/9] slub: Optimize deactivate_slab() chengming.zhou
2023-11-01 13:21   ` Vlastimil Babka
2023-11-02  2:10     ` Chengming Zhou
2023-10-31 14:07 ` [RFC PATCH v4 8/9] slub: Rename all *unfreeze_partials* functions to *put_partials* chengming.zhou
2023-11-01 13:40   ` Vlastimil Babka
2023-11-02  2:12     ` Chengming Zhou
2023-10-31 14:07 ` [RFC PATCH v4 9/9] slub: Update frozen slabs documentations in the source chengming.zhou
2023-11-01 13:51   ` Vlastimil Babka
2023-11-02  2:48     ` Chengming Zhou
2023-11-01 13:33 ` [RFC PATCH v4 0/9] slub: Delay freezing of CPU partial slabs Hyeonggon Yoo
2023-11-02  2:17   ` Chengming Zhou
2023-11-01 13:59 ` Vlastimil Babka
2023-11-02  2:19   ` Chengming Zhou

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20231031140741.79387-4-chengming.zhou@linux.dev \
    --to=chengming.zhou@linux.dev \
    --cc=42.hyeyoo@gmail.com \
    --cc=akpm@linux-foundation.org \
    --cc=cl@linux.com \
    --cc=iamjoonsoo.kim@lge.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=penberg@kernel.org \
    --cc=rientjes@google.com \
    --cc=roman.gushchin@linux.dev \
    --cc=vbabka@suse.cz \
    --cc=willy@infradead.org \
    --cc=zhouchengming@bytedance.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.