All of lore.kernel.org
 help / color / mirror / Atom feed
From: Roman Gushchin <guro@fb.com>
To: Naresh Kamboju <naresh.kamboju@linaro.org>
Cc: <linux-mips@vger.kernel.org>,
	Andrew Morton <akpm@linux-foundation.org>,
	Christoph Lameter <cl@linux.com>,
	Johannes Weiner <hannes@cmpxchg.org>,
	Michal Hocko <mhocko@kernel.org>,
	Shakeel Butt <shakeelb@google.com>, linux-mm <linux-mm@kvack.org>,
	Vlastimil Babka <vbabka@suse.cz>,
	Kernel Team <kernel-team@fb.com>,
	open list <linux-kernel@vger.kernel.org>,
	<lkft-triage@lists.linaro.org>
Subject: Re: [PATCH v7 08/19] mm: memcg/slab: save obj_cgroup for non-root slab objects
Date: Thu, 16 Jul 2020 13:07:44 -0700	[thread overview]
Message-ID: <20200716200744.GB13387@carbon.dhcp.thefacebook.com> (raw)
In-Reply-To: <CA+G9fYs0vDPAL_84oDEVdGdbFEDjAR1RFoVeFTpjN6b2yS+ZPg@mail.gmail.com>

On Thu, Jul 16, 2020 at 10:25:01PM +0530, Naresh Kamboju wrote:
> On Tue, 23 Jun 2020 at 23:11, Roman Gushchin <guro@fb.com> wrote:
> >
> > Store the obj_cgroup pointer in the corresponding place of
> > page->obj_cgroups for each allocated non-root slab object.  Make sure that
> > each allocated object holds a reference to obj_cgroup.
> >
> > Objcg pointer is obtained from the memcg->objcg dereferencing in
> > memcg_kmem_get_cache() and passed from pre_alloc_hook to post_alloc_hook.
> > Then in case of successful allocation(s) it's getting stored in the
> > page->obj_cgroups vector.
> >
> > The objcg obtaining part look a bit bulky now, but it will be simplified
> > by next commits in the series.
> >
> > Signed-off-by: Roman Gushchin <guro@fb.com>
> > Reviewed-by: Vlastimil Babka <vbabka@suse.cz>
> > Reviewed-by: Shakeel Butt <shakeelb@google.com>
> > ---
> >  include/linux/memcontrol.h |  3 +-
> >  mm/memcontrol.c            | 14 +++++++--
> >  mm/slab.c                  | 18 +++++++-----
> >  mm/slab.h                  | 60 ++++++++++++++++++++++++++++++++++----
> >  mm/slub.c                  | 14 +++++----
> >  5 files changed, 88 insertions(+), 21 deletions(-)
> >
> > diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h
> > index f2f9d5d6b7d1..b845e908e76e 100644
> > --- a/include/linux/memcontrol.h
> > +++ b/include/linux/memcontrol.h
> > @@ -1404,7 +1404,8 @@ static inline void memcg_set_shrinker_bit(struct mem_cgroup *memcg,
> >  }
> >  #endif
> >
> > -struct kmem_cache *memcg_kmem_get_cache(struct kmem_cache *cachep);
> > +struct kmem_cache *memcg_kmem_get_cache(struct kmem_cache *cachep,
> > +                                       struct obj_cgroup **objcgp);
> >  void memcg_kmem_put_cache(struct kmem_cache *cachep);
> >
> >  #ifdef CONFIG_MEMCG_KMEM
> > diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> > index 60e3f3ca75ca..5e55c7afc18c 100644
> > --- a/mm/memcontrol.c
> > +++ b/mm/memcontrol.c
> > @@ -2973,7 +2973,8 @@ static inline bool memcg_kmem_bypass(void)
> >   * done with it, memcg_kmem_put_cache() must be called to release the
> >   * reference.
> >   */
> > -struct kmem_cache *memcg_kmem_get_cache(struct kmem_cache *cachep)
> > +struct kmem_cache *memcg_kmem_get_cache(struct kmem_cache *cachep,
> > +                                       struct obj_cgroup **objcgp)
> >  {
> >         struct mem_cgroup *memcg;
> >         struct kmem_cache *memcg_cachep;
> > @@ -3029,8 +3030,17 @@ struct kmem_cache *memcg_kmem_get_cache(struct kmem_cache *cachep)
> >          */
> >         if (unlikely(!memcg_cachep))
> >                 memcg_schedule_kmem_cache_create(memcg, cachep);
> > -       else if (percpu_ref_tryget(&memcg_cachep->memcg_params.refcnt))
> > +       else if (percpu_ref_tryget(&memcg_cachep->memcg_params.refcnt)) {
> > +               struct obj_cgroup *objcg = rcu_dereference(memcg->objcg);
> > +
> > +               if (!objcg || !obj_cgroup_tryget(objcg)) {
> > +                       percpu_ref_put(&memcg_cachep->memcg_params.refcnt);
> > +                       goto out_unlock;
> > +               }
> > +
> > +               *objcgp = objcg;
> >                 cachep = memcg_cachep;
> > +       }
> >  out_unlock:
> >         rcu_read_unlock();
> >         return cachep;
> > diff --git a/mm/slab.c b/mm/slab.c
> > index 4c7013eeacd9..a7cc1336221f 100644
> > --- a/mm/slab.c
> > +++ b/mm/slab.c
> > @@ -3222,9 +3222,10 @@ slab_alloc_node(struct kmem_cache *cachep, gfp_t flags, int nodeid,
> >         unsigned long save_flags;
> >         void *ptr;
> >         int slab_node = numa_mem_id();
> > +       struct obj_cgroup *objcg = NULL;
> >
> >         flags &= gfp_allowed_mask;
> > -       cachep = slab_pre_alloc_hook(cachep, flags);
> > +       cachep = slab_pre_alloc_hook(cachep, &objcg, 1, flags);
> >         if (unlikely(!cachep))
> >                 return NULL;
> >
> > @@ -3260,7 +3261,7 @@ slab_alloc_node(struct kmem_cache *cachep, gfp_t flags, int nodeid,
> >         if (unlikely(slab_want_init_on_alloc(flags, cachep)) && ptr)
> >                 memset(ptr, 0, cachep->object_size);
> >
> > -       slab_post_alloc_hook(cachep, flags, 1, &ptr);
> > +       slab_post_alloc_hook(cachep, objcg, flags, 1, &ptr);
> >         return ptr;
> >  }
> >
> > @@ -3301,9 +3302,10 @@ slab_alloc(struct kmem_cache *cachep, gfp_t flags, unsigned long caller)
> >  {
> >         unsigned long save_flags;
> >         void *objp;
> > +       struct obj_cgroup *objcg = NULL;
> >
> >         flags &= gfp_allowed_mask;
> > -       cachep = slab_pre_alloc_hook(cachep, flags);
> > +       cachep = slab_pre_alloc_hook(cachep, &objcg, 1, flags);
> >         if (unlikely(!cachep))
> >                 return NULL;
> >
> > @@ -3317,7 +3319,7 @@ slab_alloc(struct kmem_cache *cachep, gfp_t flags, unsigned long caller)
> >         if (unlikely(slab_want_init_on_alloc(flags, cachep)) && objp)
> >                 memset(objp, 0, cachep->object_size);
> >
> > -       slab_post_alloc_hook(cachep, flags, 1, &objp);
> > +       slab_post_alloc_hook(cachep, objcg, flags, 1, &objp);
> >         return objp;
> >  }
> >
> > @@ -3443,6 +3445,7 @@ void ___cache_free(struct kmem_cache *cachep, void *objp,
> >                 memset(objp, 0, cachep->object_size);
> >         kmemleak_free_recursive(objp, cachep->flags);
> >         objp = cache_free_debugcheck(cachep, objp, caller);
> > +       memcg_slab_free_hook(cachep, virt_to_head_page(objp), objp);
> >
> >         /*
> >          * Skip calling cache_free_alien() when the platform is not numa.
> > @@ -3508,8 +3511,9 @@ int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> >                           void **p)
> >  {
> >         size_t i;
> > +       struct obj_cgroup *objcg = NULL;
> >
> > -       s = slab_pre_alloc_hook(s, flags);
> > +       s = slab_pre_alloc_hook(s, &objcg, size, flags);
> >         if (!s)
> >                 return 0;
> >
> > @@ -3532,13 +3536,13 @@ int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> >                 for (i = 0; i < size; i++)
> >                         memset(p[i], 0, s->object_size);
> >
> > -       slab_post_alloc_hook(s, flags, size, p);
> > +       slab_post_alloc_hook(s, objcg, flags, size, p);
> >         /* FIXME: Trace call missing. Christoph would like a bulk variant */
> >         return size;
> >  error:
> >         local_irq_enable();
> >         cache_alloc_debugcheck_after_bulk(s, flags, i, p, _RET_IP_);
> > -       slab_post_alloc_hook(s, flags, i, p);
> > +       slab_post_alloc_hook(s, objcg, flags, i, p);
> >         __kmem_cache_free_bulk(s, i, p);
> >         return 0;
> >  }
> > diff --git a/mm/slab.h b/mm/slab.h
> > index 7d175c2f1a61..c37a50f26e41 100644
> > --- a/mm/slab.h
> > +++ b/mm/slab.h
> > @@ -469,6 +469,41 @@ static inline void memcg_free_page_obj_cgroups(struct page *page)
> >         page->obj_cgroups = NULL;
> >  }
> >
> > +static inline void memcg_slab_post_alloc_hook(struct kmem_cache *s,
> > +                                             struct obj_cgroup *objcg,
> > +                                             size_t size, void **p)
> > +{
> > +       struct page *page;
> > +       unsigned long off;
> > +       size_t i;
> > +
> > +       for (i = 0; i < size; i++) {
> > +               if (likely(p[i])) {
> > +                       page = virt_to_head_page(p[i]);
> > +                       off = obj_to_index(s, page, p[i]);
> > +                       obj_cgroup_get(objcg);
> > +                       page_obj_cgroups(page)[off] = objcg;
> > +               }
> > +       }
> > +       obj_cgroup_put(objcg);
> > +       memcg_kmem_put_cache(s);
> > +}
> > +
> > +static inline void memcg_slab_free_hook(struct kmem_cache *s, struct page *page,
> > +                                       void *p)
> > +{
> > +       struct obj_cgroup *objcg;
> > +       unsigned int off;
> > +
> > +       if (!memcg_kmem_enabled() || is_root_cache(s))
> > +               return;
> > +
> > +       off = obj_to_index(s, page, p);
> > +       objcg = page_obj_cgroups(page)[off];
> > +       page_obj_cgroups(page)[off] = NULL;
> > +       obj_cgroup_put(objcg);
> > +}
> > +
> >  extern void slab_init_memcg_params(struct kmem_cache *);
> >  extern void memcg_link_cache(struct kmem_cache *s, struct mem_cgroup *memcg);
> >
> > @@ -528,6 +563,17 @@ static inline void memcg_free_page_obj_cgroups(struct page *page)
> >  {
> >  }
> >
> > +static inline void memcg_slab_post_alloc_hook(struct kmem_cache *s,
> > +                                             struct obj_cgroup *objcg,
> > +                                             size_t size, void **p)
> > +{
> > +}
> > +
> > +static inline void memcg_slab_free_hook(struct kmem_cache *s, struct page *page,
> > +                                       void *p)
> > +{
> > +}
> > +
> >  static inline void slab_init_memcg_params(struct kmem_cache *s)
> >  {
> >  }
> > @@ -630,7 +676,8 @@ static inline size_t slab_ksize(const struct kmem_cache *s)
> >  }
> >
> >  static inline struct kmem_cache *slab_pre_alloc_hook(struct kmem_cache *s,
> > -                                                    gfp_t flags)
> > +                                                    struct obj_cgroup **objcgp,
> > +                                                    size_t size, gfp_t flags)
> >  {
> >         flags &= gfp_allowed_mask;
> >
> > @@ -644,13 +691,14 @@ static inline struct kmem_cache *slab_pre_alloc_hook(struct kmem_cache *s,
> >
> >         if (memcg_kmem_enabled() &&
> >             ((flags & __GFP_ACCOUNT) || (s->flags & SLAB_ACCOUNT)))
> > -               return memcg_kmem_get_cache(s);
> > +               return memcg_kmem_get_cache(s, objcgp);
> >
> >         return s;
> >  }
> >
> > -static inline void slab_post_alloc_hook(struct kmem_cache *s, gfp_t flags,
> > -                                       size_t size, void **p)
> > +static inline void slab_post_alloc_hook(struct kmem_cache *s,
> > +                                       struct obj_cgroup *objcg,
> > +                                       gfp_t flags, size_t size, void **p)
> >  {
> >         size_t i;
> >
> > @@ -662,8 +710,8 @@ static inline void slab_post_alloc_hook(struct kmem_cache *s, gfp_t flags,
> >                                          s->flags, flags);
> >         }
> >
> > -       if (memcg_kmem_enabled())
> > -               memcg_kmem_put_cache(s);
> > +       if (memcg_kmem_enabled() && !is_root_cache(s))
> > +               memcg_slab_post_alloc_hook(s, objcg, size, p);
> >  }
> >
> >  #ifndef CONFIG_SLOB
> > diff --git a/mm/slub.c b/mm/slub.c
> > index aa8d18824e62..25810980a26c 100644
> > --- a/mm/slub.c
> > +++ b/mm/slub.c
> > @@ -2821,8 +2821,9 @@ static __always_inline void *slab_alloc_node(struct kmem_cache *s,
> >         struct kmem_cache_cpu *c;
> >         struct page *page;
> >         unsigned long tid;
> > +       struct obj_cgroup *objcg = NULL;
> >
> > -       s = slab_pre_alloc_hook(s, gfpflags);
> > +       s = slab_pre_alloc_hook(s, &objcg, 1, gfpflags);
> >         if (!s)
> >                 return NULL;
> >  redo:
> > @@ -2898,7 +2899,7 @@ static __always_inline void *slab_alloc_node(struct kmem_cache *s,
> >         if (unlikely(slab_want_init_on_alloc(gfpflags, s)) && object)
> >                 memset(object, 0, s->object_size);
> >
> > -       slab_post_alloc_hook(s, gfpflags, 1, &object);
> > +       slab_post_alloc_hook(s, objcg, gfpflags, 1, &object);
> >
> >         return object;
> >  }
> > @@ -3103,6 +3104,8 @@ static __always_inline void do_slab_free(struct kmem_cache *s,
> >         void *tail_obj = tail ? : head;
> >         struct kmem_cache_cpu *c;
> >         unsigned long tid;
> > +
> > +       memcg_slab_free_hook(s, page, head);
> >  redo:
> >         /*
> >          * Determine the currently cpus per cpu slab.
> > @@ -3282,9 +3285,10 @@ int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> >  {
> >         struct kmem_cache_cpu *c;
> >         int i;
> > +       struct obj_cgroup *objcg = NULL;
> >
> >         /* memcg and kmem_cache debug support */
> > -       s = slab_pre_alloc_hook(s, flags);
> > +       s = slab_pre_alloc_hook(s, &objcg, size, flags);
> >         if (unlikely(!s))
> >                 return false;
> >         /*
> > @@ -3338,11 +3342,11 @@ int kmem_cache_alloc_bulk(struct kmem_cache *s, gfp_t flags, size_t size,
> >         }
> >
> >         /* memcg and kmem_cache debug support */
> > -       slab_post_alloc_hook(s, flags, size, p);
> > +       slab_post_alloc_hook(s, objcg, flags, size, p);
> >         return i;
> >  error:
> >         local_irq_enable();
> > -       slab_post_alloc_hook(s, flags, i, p);
> > +       slab_post_alloc_hook(s, objcg, flags, i, p);
> >         __kmem_cache_free_bulk(s, i, p);
> >         return 0;
> >  }
> >
> 
> I am not sure if this is the related patch or not that is causing
> mips architecture build failure on linux -next.

Hello, Naresh!

Thank you for the report, interesting...
There is nothing arch-specific in the code, so there must be something
compiler-dependent. My wild guess is that the problem is caused by a memory
allocation from the memcg_slab_post_alloc_hook(), but it's added by a later
patch in the series. So if it really fails at this patch, there must be something
different. I'll try to reproduce it, but I have to install the MIPS toolchain first,
so it might take some time. If it creates some troubles, can you, please, check
if s/__always_inline/inline helps?

Thanks!

> 
> make -sk KBUILD_BUILD_USER=TuxBuild -C/linux -j16 ARCH=mips
> CROSS_COMPILE=mips-linux-gnu- HOSTCC=gcc CC="sccache
> mips-linux-gnu-gcc" O=build
> #
> ../mm/slub.c: In function ‘slab_alloc.constprop’:
> ../mm/slub.c:2897:30: error: inlining failed in call to always_inline
> ‘slab_alloc.constprop’: recursive inlining
>  2897 | static __always_inline void *slab_alloc(struct kmem_cache *s,
>       |                              ^~~~~~~~~~
> ../mm/slub.c:2905:14: note: called from here
>  2905 |  void *ret = slab_alloc(s, gfpflags, _RET_IP_);
>       |              ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> ../mm/slub.c: In function ‘sysfs_slab_alias’:
> ../mm/slub.c:2897:30: error: inlining failed in call to always_inline
> ‘slab_alloc.constprop’: recursive inlining
>  2897 | static __always_inline void *slab_alloc(struct kmem_cache *s,
>       |                              ^~~~~~~~~~
> ../mm/slub.c:2905:14: note: called from here
>  2905 |  void *ret = slab_alloc(s, gfpflags, _RET_IP_);
>       |              ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> ../mm/slub.c: In function ‘sysfs_slab_add’:
> ../mm/slub.c:2897:30: error: inlining failed in call to always_inline
> ‘slab_alloc.constprop’: recursive inlining
>  2897 | static __always_inline void *slab_alloc(struct kmem_cache *s,
>       |                              ^~~~~~~~~~
> ../mm/slub.c:2905:14: note: called from here
>  2905 |  void *ret = slab_alloc(s, gfpflags, _RET_IP_);
>       |              ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> 
> 
> Full build log link,
> https://urldefense.proofpoint.com/v2/url?u=https-3A__builds.tuxbuild.com_jBgeEp1SD-2DbUldWES782yQ_build.log&d=DwIFaQ&c=5VD0RTtNlTh3ycd41b3MUw&r=jJYgtDM7QT-W-Fz_d29HYQ&m=XO4W3q0FbuSIPvvA7_av_jFzLvJw7jfRpBhtUGJfpKE&s=tlrwIrPSvEYT8QV4wnWxOnXQQMA8EYeYS-4O-BxVLOA&e= 
> 
> -- 
> Linaro LKFT
> https://urldefense.proofpoint.com/v2/url?u=https-3A__lkft.linaro.org&d=DwIFaQ&c=5VD0RTtNlTh3ycd41b3MUw&r=jJYgtDM7QT-W-Fz_d29HYQ&m=XO4W3q0FbuSIPvvA7_av_jFzLvJw7jfRpBhtUGJfpKE&s=WFK6dW0eWT7ZY1HBBJeLSUgOxE2J11TOhbqdnCIdvCE&e= 

  reply	other threads:[~2020-07-16 20:08 UTC|newest]

Thread overview: 33+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-06-23 17:40 [PATCH v7 00/19] The new cgroup slab memory controller Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 01/19] mm: memcg: factor out memcg- and lruvec-level changes out of __mod_lruvec_state() Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 02/19] mm: memcg: prepare for byte-sized vmstat items Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 03/19] mm: memcg: convert vmstat slab counters to bytes Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 04/19] mm: slub: implement SLUB version of obj_to_index() Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 05/19] mm: memcontrol: decouple reference counting from page accounting Roman Gushchin
2020-08-03  9:00   ` Michal Hocko
2020-08-03 15:03     ` Johannes Weiner
2020-08-03 15:08       ` Michal Hocko
2020-06-23 17:40 ` [PATCH v7 06/19] mm: memcg/slab: obj_cgroup API Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 07/19] mm: memcg/slab: allocate obj_cgroups for non-root slab pages Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 08/19] mm: memcg/slab: save obj_cgroup for non-root slab objects Roman Gushchin
2020-07-16 16:55   ` Naresh Kamboju
2020-07-16 16:55     ` Naresh Kamboju
2020-07-16 20:07     ` Roman Gushchin [this message]
2020-07-17  5:34       ` Naresh Kamboju
2020-07-17  5:34         ` Naresh Kamboju
2020-06-23 17:40 ` [PATCH v7 09/19] mm: memcg/slab: charge individual slab objects instead of pages Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 10/19] mm: memcg/slab: deprecate memory.kmem.slabinfo Roman Gushchin
2020-06-24  1:43   ` Shakeel Butt
2020-06-24  1:43     ` Shakeel Butt
2020-06-24  1:53     ` Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 11/19] mm: memcg/slab: move memcg_kmem_bypass() to memcontrol.h Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 12/19] mm: memcg/slab: use a single set of kmem_caches for all accounted allocations Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 13/19] mm: memcg/slab: simplify memcg cache creation Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 14/19] mm: memcg/slab: remove memcg_kmem_get_cache() Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 15/19] mm: memcg/slab: deprecate slab_root_caches Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 16/19] mm: memcg/slab: remove redundant check in memcg_accumulate_slabinfo() Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 17/19] mm: memcg/slab: use a single set of kmem_caches for all allocations Roman Gushchin
2020-07-18 17:24   ` Guenter Roeck
2020-07-18 23:03     ` Roman Gushchin
2020-06-23 17:40 ` [PATCH v7 18/19] kselftests: cgroup: add kernel memory accounting tests Roman Gushchin
  -- strict thread matches above, loose matches on Subject: below --
2020-06-23  1:58 [PATCH v7 00/19] The new cgroup slab memory controller Roman Gushchin
2020-06-23  1:58 ` [PATCH v7 08/19] mm: memcg/slab: save obj_cgroup for non-root slab objects Roman Gushchin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200716200744.GB13387@carbon.dhcp.thefacebook.com \
    --to=guro@fb.com \
    --cc=akpm@linux-foundation.org \
    --cc=cl@linux.com \
    --cc=hannes@cmpxchg.org \
    --cc=kernel-team@fb.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mips@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=lkft-triage@lists.linaro.org \
    --cc=mhocko@kernel.org \
    --cc=naresh.kamboju@linaro.org \
    --cc=shakeelb@google.com \
    --cc=vbabka@suse.cz \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.