From: "Paul E. McKenney" <paulmck@kernel.org>
To: "Uladzislau Rezki (Sony)" <urezki@gmail.com>
Cc: LKML <linux-kernel@vger.kernel.org>,
linux-mm@kvack.org, Andrew Morton <akpm@linux-foundation.org>,
"Theodore Y . Ts'o" <tytso@mit.edu>,
Matthew Wilcox <willy@infradead.org>,
Joel Fernandes <joel@joelfernandes.org>,
RCU <rcu@vger.kernel.org>,
Oleksiy Avramchenko <oleksiy.avramchenko@sonymobile.com>
Subject: Re: [PATCH 11/24] rcu/tree: Maintain separate array for vmalloc ptrs
Date: Fri, 1 May 2020 14:37:53 -0700 [thread overview]
Message-ID: <20200501213753.GE7560@paulmck-ThinkPad-P72> (raw)
In-Reply-To: <20200428205903.61704-12-urezki@gmail.com>
On Tue, Apr 28, 2020 at 10:58:50PM +0200, Uladzislau Rezki (Sony) wrote:
> To do so we use an array of common kvfree_rcu_bulk_data
> structure. It consists of two elements, index number 0
> corresponds to SLAB ptrs., whereas vmalloc pointers can
> be accessed by using index number 1.
>
> The reason of not mixing pointers is to have an easy way
> to to distinguish them.
>
> It is also the preparation patch for head-less objects
> support. When an object is head-less we can not queue
> it into any list, instead a pointer is placed directly
> into an array.
>
> Signed-off-by: Uladzislau Rezki (Sony) <urezki@gmail.com>
> Signed-off-by: Joel Fernandes (Google) <joel@joelfernandes.org>
> Reviewed-by: Joel Fernandes (Google) <joel@joelfernandes.org>
> Co-developed-by: Joel Fernandes (Google) <joel@joelfernandes.org>
> ---
> kernel/rcu/tree.c | 172 +++++++++++++++++++++++++++++-----------------
> 1 file changed, 109 insertions(+), 63 deletions(-)
>
> diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c
> index d8975819b1c9..7983926af95b 100644
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -57,6 +57,7 @@
> #include <linux/slab.h>
> #include <linux/sched/isolation.h>
> #include <linux/sched/clock.h>
> +#include <linux/mm.h>
> #include "../time/tick-internal.h"
>
> #include "tree.h"
> @@ -2857,44 +2858,44 @@ EXPORT_SYMBOL_GPL(call_rcu);
> #define KFREE_N_BATCHES 2
>
> /**
> - * struct kfree_rcu_bulk_data - single block to store kfree_rcu() pointers
> + * struct kvfree_rcu_bulk_data - single block to store kvfree_rcu() pointers
> * @nr_records: Number of active pointers in the array
> - * @records: Array of the kfree_rcu() pointers
> * @next: Next bulk object in the block chain
> + * @records: Array of the kvfree_rcu() pointers
> */
> -struct kfree_rcu_bulk_data {
> +struct kvfree_rcu_bulk_data {
> unsigned long nr_records;
> - struct kfree_rcu_bulk_data *next;
> + struct kvfree_rcu_bulk_data *next;
> void *records[];
> };
>
> /*
> * This macro defines how many entries the "records" array
> * will contain. It is based on the fact that the size of
> - * kfree_rcu_bulk_data structure becomes exactly one page.
> + * kvfree_rcu_bulk_data structure becomes exactly one page.
> */
> -#define KFREE_BULK_MAX_ENTR \
> - ((PAGE_SIZE - sizeof(struct kfree_rcu_bulk_data)) / sizeof(void *))
> +#define KVFREE_BULK_MAX_ENTR \
> + ((PAGE_SIZE - sizeof(struct kvfree_rcu_bulk_data)) / sizeof(void *))
>
> /**
> * struct kfree_rcu_cpu_work - single batch of kfree_rcu() requests
> * @rcu_work: Let queue_rcu_work() invoke workqueue handler after grace period
> * @head_free: List of kfree_rcu() objects waiting for a grace period
> - * @bhead_free: Bulk-List of kfree_rcu() objects waiting for a grace period
> + * @bkvhead_free: Bulk-List of kvfree_rcu() objects waiting for a grace period
> * @krcp: Pointer to @kfree_rcu_cpu structure
> */
>
> struct kfree_rcu_cpu_work {
> struct rcu_work rcu_work;
> struct rcu_head *head_free;
> - struct kfree_rcu_bulk_data *bhead_free;
> + struct kvfree_rcu_bulk_data *bkvhead_free[2];
> struct kfree_rcu_cpu *krcp;
> };
>
> /**
> * struct kfree_rcu_cpu - batch up kfree_rcu() requests for RCU grace period
> * @head: List of kfree_rcu() objects not yet waiting for a grace period
> - * @bhead: Bulk-List of kfree_rcu() objects not yet waiting for a grace period
> + * @bkvhead: Bulk-List of kvfree_rcu() objects not yet waiting for a grace period
> * @krw_arr: Array of batches of kfree_rcu() objects waiting for a grace period
> * @lock: Synchronize access to this structure
> * @monitor_work: Promote @head to @head_free after KFREE_DRAIN_JIFFIES
> @@ -2908,7 +2909,7 @@ struct kfree_rcu_cpu_work {
> */
> struct kfree_rcu_cpu {
> struct rcu_head *head;
> - struct kfree_rcu_bulk_data *bhead;
> + struct kvfree_rcu_bulk_data *bkvhead[2];
> struct kfree_rcu_cpu_work krw_arr[KFREE_N_BATCHES];
> raw_spinlock_t lock;
> struct delayed_work monitor_work;
> @@ -2932,7 +2933,7 @@ static DEFINE_PER_CPU(struct kfree_rcu_cpu, krc) = {
> };
>
> static __always_inline void
> -debug_rcu_bhead_unqueue(struct kfree_rcu_bulk_data *bhead)
> +debug_rcu_bhead_unqueue(struct kvfree_rcu_bulk_data *bhead)
> {
> #ifdef CONFIG_DEBUG_OBJECTS_RCU_HEAD
> int i;
> @@ -2961,20 +2962,20 @@ krc_this_cpu_unlock(struct kfree_rcu_cpu *krcp, unsigned long flags)
> local_irq_restore(flags);
> }
>
> -static inline struct kfree_rcu_bulk_data *
> +static inline struct kvfree_rcu_bulk_data *
> get_cached_bnode(struct kfree_rcu_cpu *krcp)
> {
> if (!krcp->nr_bkv_objs)
> return NULL;
>
> krcp->nr_bkv_objs--;
> - return (struct kfree_rcu_bulk_data *)
> + return (struct kvfree_rcu_bulk_data *)
> llist_del_first(&krcp->bkvcache);
> }
>
> static inline bool
> put_cached_bnode(struct kfree_rcu_cpu *krcp,
> - struct kfree_rcu_bulk_data *bnode)
> + struct kvfree_rcu_bulk_data *bnode)
> {
> /* Check the limit. */
> if (krcp->nr_bkv_objs >= rcu_min_cached_objs)
> @@ -2993,41 +2994,73 @@ put_cached_bnode(struct kfree_rcu_cpu *krcp,
> static void kfree_rcu_work(struct work_struct *work)
> {
> unsigned long flags;
> + struct kvfree_rcu_bulk_data *bkhead, *bvhead, *bnext;
> struct rcu_head *head, *next;
> - struct kfree_rcu_bulk_data *bhead, *bnext;
> struct kfree_rcu_cpu *krcp;
> struct kfree_rcu_cpu_work *krwp;
> + int i;
>
> krwp = container_of(to_rcu_work(work),
> struct kfree_rcu_cpu_work, rcu_work);
> krcp = krwp->krcp;
> +
> raw_spin_lock_irqsave(&krcp->lock, flags);
> + /* Channel 1. */
> + bkhead = krwp->bkvhead_free[0];
> + krwp->bkvhead_free[0] = NULL;
> +
> + /* Channel 2. */
> + bvhead = krwp->bkvhead_free[1];
> + krwp->bkvhead_free[1] = NULL;
> +
> + /* Channel 3. */
> head = krwp->head_free;
> krwp->head_free = NULL;
> - bhead = krwp->bhead_free;
> - krwp->bhead_free = NULL;
> raw_spin_unlock_irqrestore(&krcp->lock, flags);
>
> - /* "bhead" is now private, so traverse locklessly. */
> - for (; bhead; bhead = bnext) {
> - bnext = bhead->next;
> -
> - debug_rcu_bhead_unqueue(bhead);
> + /* kmalloc()/kfree() channel. */
> + for (; bkhead; bkhead = bnext) {
> + bnext = bkhead->next;
> + debug_rcu_bhead_unqueue(bkhead);
>
> rcu_lock_acquire(&rcu_callback_map);
Given that rcu_lock_acquire() only affects lockdep, I have to ask exactly
what concurrency design you are using here...
> trace_rcu_invoke_kfree_bulk_callback(rcu_state.name,
> - bhead->nr_records, bhead->records);
> + bkhead->nr_records, bkhead->records);
> +
> + kfree_bulk(bkhead->nr_records, bkhead->records);
> + rcu_lock_release(&rcu_callback_map);
> +
> + krcp = krc_this_cpu_lock(&flags);
> + if (put_cached_bnode(krcp, bkhead))
> + bkhead = NULL;
> + krc_this_cpu_unlock(krcp, flags);
> +
> + if (bkhead)
> + free_page((unsigned long) bkhead);
> +
> + cond_resched_tasks_rcu_qs();
> + }
> +
> + /* vmalloc()/vfree() channel. */
> + for (; bvhead; bvhead = bnext) {
> + bnext = bvhead->next;
> + debug_rcu_bhead_unqueue(bvhead);
>
> - kfree_bulk(bhead->nr_records, bhead->records);
> + rcu_lock_acquire(&rcu_callback_map);
And the same here.
> + for (i = 0; i < bvhead->nr_records; i++) {
> + trace_rcu_invoke_kfree_callback(rcu_state.name,
> + (struct rcu_head *) bvhead->records[i], 0);
> + vfree(bvhead->records[i]);
> + }
> rcu_lock_release(&rcu_callback_map);
>
> krcp = krc_this_cpu_lock(&flags);
> - if (put_cached_bnode(krcp, bhead))
> - bhead = NULL;
> + if (put_cached_bnode(krcp, bvhead))
> + bvhead = NULL;
> krc_this_cpu_unlock(krcp, flags);
>
> - if (bhead)
> - free_page((unsigned long) bhead);
> + if (bvhead)
> + free_page((unsigned long) bvhead);
>
> cond_resched_tasks_rcu_qs();
> }
> @@ -3047,7 +3080,7 @@ static void kfree_rcu_work(struct work_struct *work)
> trace_rcu_invoke_kfree_callback(rcu_state.name, head, offset);
>
> if (!WARN_ON_ONCE(!__is_kfree_rcu_offset(offset)))
> - kfree(ptr);
> + kvfree(ptr);
>
> rcu_lock_release(&rcu_callback_map);
> cond_resched_tasks_rcu_qs();
> @@ -3072,21 +3105,34 @@ static inline bool queue_kfree_rcu_work(struct kfree_rcu_cpu *krcp)
> krwp = &(krcp->krw_arr[i]);
>
> /*
> - * Try to detach bhead or head and attach it over any
> + * Try to detach bkvhead or head and attach it over any
> * available corresponding free channel. It can be that
> * a previous RCU batch is in progress, it means that
> * immediately to queue another one is not possible so
> * return false to tell caller to retry.
> */
> - if ((krcp->bhead && !krwp->bhead_free) ||
> + if ((krcp->bkvhead[0] && !krwp->bkvhead_free[0]) ||
> + (krcp->bkvhead[1] && !krwp->bkvhead_free[1]) ||
> (krcp->head && !krwp->head_free)) {
> - /* Channel 1. */
> - if (!krwp->bhead_free) {
> - krwp->bhead_free = krcp->bhead;
> - krcp->bhead = NULL;
> + /*
> + * Channel 1 corresponds to SLAB ptrs.
> + */
> + if (!krwp->bkvhead_free[0]) {
> + krwp->bkvhead_free[0] = krcp->bkvhead[0];
> + krcp->bkvhead[0] = NULL;
> }
>
> - /* Channel 2. */
> + /*
> + * Channel 2 corresponds to vmalloc ptrs.
> + */
> + if (!krwp->bkvhead_free[1]) {
> + krwp->bkvhead_free[1] = krcp->bkvhead[1];
> + krcp->bkvhead[1] = NULL;
> + }
Why not a "for" loop here? Duplicate code is most certainly not what
we want, as it can cause all sorts of trouble down the road.
Thanx, Paul
> + /*
> + * Channel 3 corresponds to emergency path.
> + */
> if (!krwp->head_free) {
> krwp->head_free = krcp->head;
> krcp->head = NULL;
> @@ -3095,16 +3141,17 @@ static inline bool queue_kfree_rcu_work(struct kfree_rcu_cpu *krcp)
> WRITE_ONCE(krcp->count, 0);
>
> /*
> - * One work is per one batch, so there are two "free channels",
> - * "bhead_free" and "head_free" the batch can handle. It can be
> - * that the work is in the pending state when two channels have
> - * been detached following each other, one by one.
> + * One work is per one batch, so there are three
> + * "free channels", the batch can handle. It can
> + * be that the work is in the pending state when
> + * channels have been detached following by each
> + * other.
> */
> queue_rcu_work(system_wq, &krwp->rcu_work);
> }
>
> /* Repeat if any "free" corresponding channel is still busy. */
> - if (krcp->bhead || krcp->head)
> + if (krcp->bkvhead[0] || krcp->bkvhead[1] || krcp->head)
> repeat = true;
> }
>
> @@ -3146,23 +3193,22 @@ static void kfree_rcu_monitor(struct work_struct *work)
> }
>
> static inline bool
> -kfree_call_rcu_add_ptr_to_bulk(struct kfree_rcu_cpu *krcp,
> - struct rcu_head *head, rcu_callback_t func)
> +kvfree_call_rcu_add_ptr_to_bulk(struct kfree_rcu_cpu *krcp, void *ptr)
> {
> - struct kfree_rcu_bulk_data *bnode;
> + struct kvfree_rcu_bulk_data *bnode;
> + int idx;
>
> if (unlikely(!krcp->initialized))
> return false;
>
> lockdep_assert_held(&krcp->lock);
> + idx = !!is_vmalloc_addr(ptr);
>
> /* Check if a new block is required. */
> - if (!krcp->bhead ||
> - krcp->bhead->nr_records == KFREE_BULK_MAX_ENTR) {
> + if (!krcp->bkvhead[idx] ||
> + krcp->bkvhead[idx]->nr_records == KVFREE_BULK_MAX_ENTR) {
> bnode = get_cached_bnode(krcp);
> if (!bnode) {
> - WARN_ON_ONCE(sizeof(struct kfree_rcu_bulk_data) > PAGE_SIZE);
> -
> /*
> * To keep this path working on raw non-preemptible
> * sections, prevent the optional entry into the
> @@ -3175,7 +3221,7 @@ kfree_call_rcu_add_ptr_to_bulk(struct kfree_rcu_cpu *krcp,
> if (IS_ENABLED(CONFIG_PREEMPT_RT))
> return false;
>
> - bnode = (struct kfree_rcu_bulk_data *)
> + bnode = (struct kvfree_rcu_bulk_data *)
> __get_free_page(GFP_NOWAIT | __GFP_NOWARN);
> }
>
> @@ -3185,30 +3231,30 @@ kfree_call_rcu_add_ptr_to_bulk(struct kfree_rcu_cpu *krcp,
>
> /* Initialize the new block. */
> bnode->nr_records = 0;
> - bnode->next = krcp->bhead;
> + bnode->next = krcp->bkvhead[idx];
>
> /* Attach it to the head. */
> - krcp->bhead = bnode;
> + krcp->bkvhead[idx] = bnode;
> }
>
> /* Finally insert. */
> - krcp->bhead->records[krcp->bhead->nr_records++] =
> - (void *) head - (unsigned long) func;
> + krcp->bkvhead[idx]->records
> + [krcp->bkvhead[idx]->nr_records++] = ptr;
>
> return true;
> }
>
> /*
> - * Queue a request for lazy invocation of kfree_bulk()/kfree() after a grace
> - * period. Please note there are two paths are maintained, one is the main one
> - * that uses kfree_bulk() interface and second one is emergency one, that is
> - * used only when the main path can not be maintained temporary, due to memory
> - * pressure.
> + * Queue a request for lazy invocation of appropriate free routine after a
> + * grace period. Please note there are three paths are maintained, two are the
> + * main ones that use array of pointers interface and third one is emergency
> + * one, that is used only when the main path can not be maintained temporary,
> + * due to memory pressure.
> *
> * Each kfree_call_rcu() request is added to a batch. The batch will be drained
> * every KFREE_DRAIN_JIFFIES number of jiffies. All the objects in the batch will
> * be free'd in workqueue context. This allows us to: batch requests together to
> - * reduce the number of grace periods during heavy kfree_rcu() load.
> + * reduce the number of grace periods during heavy kfree_rcu()/kvfree_rcu() load.
> */
> void kfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
> {
> @@ -3231,7 +3277,7 @@ void kfree_call_rcu(struct rcu_head *head, rcu_callback_t func)
> * Under high memory pressure GFP_NOWAIT can fail,
> * in that case the emergency path is maintained.
> */
> - if (unlikely(!kfree_call_rcu_add_ptr_to_bulk(krcp, head, func))) {
> + if (unlikely(!kvfree_call_rcu_add_ptr_to_bulk(krcp, ptr))) {
> head->func = func;
> head->next = krcp->head;
> krcp->head = head;
> @@ -4212,7 +4258,7 @@ static void __init kfree_rcu_batch_init(void)
>
> for_each_possible_cpu(cpu) {
> struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu);
> - struct kfree_rcu_bulk_data *bnode;
> + struct kvfree_rcu_bulk_data *bnode;
>
> for (i = 0; i < KFREE_N_BATCHES; i++) {
> INIT_RCU_WORK(&krcp->krw_arr[i].rcu_work, kfree_rcu_work);
> @@ -4220,7 +4266,7 @@ static void __init kfree_rcu_batch_init(void)
> }
>
> for (i = 0; i < rcu_min_cached_objs; i++) {
> - bnode = (struct kfree_rcu_bulk_data *)
> + bnode = (struct kvfree_rcu_bulk_data *)
> __get_free_page(GFP_NOWAIT | __GFP_NOWARN);
>
> if (bnode)
> --
> 2.20.1
>
next prev parent reply other threads:[~2020-05-01 21:37 UTC|newest]
Thread overview: 78+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-04-28 20:58 [PATCH 00/24] Introduce kvfree_rcu(1 or 2 arguments) Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 01/24] rcu/tree: Keep kfree_rcu() awake during lock contention Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 02/24] rcu/tree: Skip entry into the page allocator for PREEMPT_RT Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 03/24] rcu/tree: Use consistent style for comments Uladzislau Rezki (Sony)
2020-05-01 19:05 ` Paul E. McKenney
2020-05-01 20:52 ` Joe Perches
2020-05-03 23:44 ` Joel Fernandes
2020-05-04 0:23 ` Paul E. McKenney
2020-05-04 0:34 ` Joe Perches
2020-05-04 0:41 ` Joel Fernandes
2020-05-03 23:52 ` Joel Fernandes
2020-05-04 0:26 ` Paul E. McKenney
2020-05-04 0:39 ` Joel Fernandes
2020-04-28 20:58 ` [PATCH 04/24] rcu/tree: Repeat the monitor if any free channel is busy Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 05/24] rcu/tree: Simplify debug_objects handling Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 06/24] rcu/tree: Simplify KFREE_BULK_MAX_ENTR macro Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 07/24] rcu/tree: move locking/unlocking to separate functions Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 08/24] rcu/tree: Use static initializer for krc.lock Uladzislau Rezki (Sony)
2020-05-01 21:17 ` Paul E. McKenney
2020-05-04 12:10 ` Uladzislau Rezki
2020-04-28 20:58 ` [PATCH 09/24] rcu/tree: cache specified number of objects Uladzislau Rezki (Sony)
2020-05-01 21:27 ` Paul E. McKenney
2020-05-04 12:43 ` Uladzislau Rezki
2020-05-04 15:24 ` Paul E. McKenney
2020-05-04 17:48 ` Uladzislau Rezki
2020-05-04 18:07 ` Paul E. McKenney
2020-05-04 18:08 ` Joel Fernandes
2020-05-04 19:01 ` Paul E. McKenney
2020-05-04 19:37 ` Joel Fernandes
2020-05-04 19:51 ` Uladzislau Rezki
2020-05-04 20:15 ` joel
2020-05-04 20:16 ` Paul E. McKenney
2020-05-05 11:03 ` Uladzislau Rezki
2020-04-28 20:58 ` [PATCH 10/24] rcu/tree: add rcutree.rcu_min_cached_objs description Uladzislau Rezki (Sony)
2020-05-01 22:25 ` Paul E. McKenney
2020-05-04 12:44 ` Uladzislau Rezki
2020-04-28 20:58 ` [PATCH 11/24] rcu/tree: Maintain separate array for vmalloc ptrs Uladzislau Rezki (Sony)
2020-05-01 21:37 ` Paul E. McKenney [this message]
2020-05-03 23:42 ` Joel Fernandes
2020-05-04 0:20 ` Paul E. McKenney
2020-05-04 0:58 ` Joel Fernandes
2020-05-04 2:20 ` Paul E. McKenney
2020-05-04 14:25 ` Uladzislau Rezki
2020-04-28 20:58 ` [PATCH 12/24] rcu/tiny: support vmalloc in tiny-RCU Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 13/24] rcu: Rename rcu_invoke_kfree_callback/rcu_kfree_callback Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 14/24] rcu: Rename __is_kfree_rcu_offset() macro Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 15/24] rcu: Rename kfree_call_rcu() to the kvfree_call_rcu() Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 16/24] mm/list_lru.c: Rename kvfree_rcu() to local variant Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 17/24] rcu: Introduce 2 arg kvfree_rcu() interface Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 18/24] mm/list_lru.c: Remove kvfree_rcu_local() function Uladzislau Rezki (Sony)
2020-04-28 20:58 ` [PATCH 19/24] rcu/tree: Support reclaim for head-less object Uladzislau Rezki (Sony)
2020-05-01 22:39 ` Paul E. McKenney
2020-05-04 0:12 ` Joel Fernandes
2020-05-04 0:28 ` Paul E. McKenney
2020-05-04 0:32 ` Joel Fernandes
2020-05-04 14:21 ` Uladzislau Rezki
2020-05-04 15:31 ` Paul E. McKenney
2020-05-04 16:56 ` Uladzislau Rezki
2020-05-04 17:08 ` Paul E. McKenney
2020-05-04 12:57 ` Uladzislau Rezki
2020-04-28 20:58 ` [PATCH 20/24] rcu/tree: Make kvfree_rcu() tolerate any alignment Uladzislau Rezki (Sony)
2020-05-01 23:00 ` Paul E. McKenney
2020-05-04 0:24 ` Joel Fernandes
2020-05-04 0:29 ` Paul E. McKenney
2020-05-04 0:31 ` Joel Fernandes
2020-05-04 12:56 ` Uladzislau Rezki
2020-04-28 20:59 ` [PATCH 21/24] rcu/tiny: move kvfree_call_rcu() out of header Uladzislau Rezki (Sony)
2020-05-01 23:03 ` Paul E. McKenney
2020-05-04 12:45 ` Uladzislau Rezki
2020-05-06 18:29 ` Uladzislau Rezki
2020-05-06 18:45 ` Paul E. McKenney
2020-05-07 17:34 ` Uladzislau Rezki
2020-04-28 20:59 ` [PATCH 22/24] rcu/tiny: support reclaim for head-less object Uladzislau Rezki (Sony)
2020-05-01 23:06 ` Paul E. McKenney
2020-05-04 0:27 ` Joel Fernandes
2020-05-04 12:45 ` Uladzislau Rezki
2020-04-28 20:59 ` [PATCH 23/24] rcu: Introduce 1 arg kvfree_rcu() interface Uladzislau Rezki (Sony)
2020-04-28 20:59 ` [PATCH 24/24] lib/test_vmalloc.c: Add test cases for kvfree_rcu() Uladzislau Rezki (Sony)
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200501213753.GE7560@paulmck-ThinkPad-P72 \
--to=paulmck@kernel.org \
--cc=akpm@linux-foundation.org \
--cc=joel@joelfernandes.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=oleksiy.avramchenko@sonymobile.com \
--cc=rcu@vger.kernel.org \
--cc=tytso@mit.edu \
--cc=urezki@gmail.com \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).