All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH 0/2] generic smp helpers vs kmalloc
@ 2009-02-12 22:32 Peter Zijlstra
  2009-02-12 22:32 ` [PATCH 1/2] generic-smp: remove single ipi fallback for smp_call_function_many() Peter Zijlstra
  2009-02-12 22:32 ` [PATCH 2/2] generic-smp: remove kmalloc usage Peter Zijlstra
  0 siblings, 2 replies; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-12 22:32 UTC (permalink / raw)
  To: Linus Torvalds, Nick Piggin, Jens Axboe, Paul E. McKenney,
	Ingo Molnar, Rusty Russell
  Cc: linux-kernel, Oleg Nesterov, Peter Zijlstra

Steve's recent patches brought up the need (or rather, lack thereof) of the
kmalloc() usage in smp_call_function_single().

The first patch, removes the need for kmalloc() for smp_call_function() too.
The second patch then fully removes the kmalloc() usage.

Patches have been compile and boot tested on a dual core -- so no real
stress testing.
-- 


^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH 1/2] generic-smp: remove single ipi fallback for smp_call_function_many()
  2009-02-12 22:32 [PATCH 0/2] generic smp helpers vs kmalloc Peter Zijlstra
@ 2009-02-12 22:32 ` Peter Zijlstra
  2009-02-14 14:41   ` [PATCH 1.5/2] generic-smp: fix initial quiesent count Peter Zijlstra
  2009-02-12 22:32 ` [PATCH 2/2] generic-smp: remove kmalloc usage Peter Zijlstra
  1 sibling, 1 reply; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-12 22:32 UTC (permalink / raw)
  To: Linus Torvalds, Nick Piggin, Jens Axboe, Paul E. McKenney,
	Ingo Molnar, Rusty Russell
  Cc: linux-kernel, Oleg Nesterov, Peter Zijlstra

[-- Attachment #1: smp-many-fallback.patch --]
[-- Type: text/plain, Size: 7674 bytes --]

In preparation of removing the kmalloc() calls from the generic-ipi code
get rid of the single ipi fallback for smp_call_function_many().

Because we cannot get around carrying the cpumask in the data -- imagine
2 such calls with different but overlapping masks -- put in a full mask.

Also, since we cannot simply remove an item from the global queue (another
cpu might be observing it), a quiesence of sorts needs to be observed. The
current code uses regular RCU for that purpose.

However, since we'll be wanting to quickly reuse an item, we need something
with a much faster turn-around. We do this by simply observing the global
queue quiesence. Since there are a limited number of elements, it will auto
force a quiecent state if we wait for it.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
---
 kernel/smp.c |  118 +++++++++++++++++++++++++++++++++++++++--------------------
 1 file changed, 78 insertions(+), 40 deletions(-)

Index: linux-2.6/kernel/smp.c
===================================================================
--- linux-2.6.orig/kernel/smp.c
+++ linux-2.6/kernel/smp.c
@@ -12,8 +12,22 @@
 #include <linux/smp.h>
 
 static DEFINE_PER_CPU(struct call_single_queue, call_single_queue);
-static LIST_HEAD(call_function_queue);
-__cacheline_aligned_in_smp DEFINE_SPINLOCK(call_function_lock);
+
+static struct {
+	struct list_head	queue;
+	spinlock_t		lock;
+	/*
+	 * mini-RCU state machine
+	 *
+	 * @counter counts the number of entries on the queue, and @quiesent
+	 * is a sequence count incremented every time the queue is idle.
+	 */
+	unsigned int		counter;
+	unsigned int		quiesent;
+} call_function __cacheline_aligned_in_smp = {
+	.queue = LIST_HEAD_INIT(call_function.queue),
+	.lock  = __SPIN_LOCK_UNLOCKED(call_function.lock),
+};
 
 enum {
 	CSD_FLAG_WAIT		= 0x01,
@@ -25,8 +39,11 @@ struct call_function_data {
 	struct call_single_data csd;
 	spinlock_t lock;
 	unsigned int refs;
-	struct rcu_head rcu_head;
-	unsigned long cpumask_bits[];
+	union {
+		struct rcu_head rcu_head;
+		unsigned int stamp;
+	};
+	struct cpumask cpumask;
 };
 
 struct call_single_queue {
@@ -107,17 +124,16 @@ void generic_smp_call_function_interrupt
 	 * It's ok to use list_for_each_rcu() here even though we may delete
 	 * 'pos', since list_del_rcu() doesn't clear ->next
 	 */
-	rcu_read_lock();
-	list_for_each_entry_rcu(data, &call_function_queue, csd.list) {
-		int refs;
+	list_for_each_entry_rcu(data, &call_function.queue, csd.list) {
+		int refs, busy = 0;
 
-		if (!cpumask_test_cpu(cpu, to_cpumask(data->cpumask_bits)))
+		if (!cpumask_test_cpu(cpu, &data->cpumask))
 			continue;
 
 		data->csd.func(data->csd.info);
 
 		spin_lock(&data->lock);
-		cpumask_clear_cpu(cpu, to_cpumask(data->cpumask_bits));
+		cpumask_clear_cpu(cpu, &data->cpumask);
 		WARN_ON(data->refs == 0);
 		data->refs--;
 		refs = data->refs;
@@ -126,9 +142,13 @@ void generic_smp_call_function_interrupt
 		if (refs)
 			continue;
 
-		spin_lock(&call_function_lock);
+		spin_lock(&call_function.lock);
 		list_del_rcu(&data->csd.list);
-		spin_unlock(&call_function_lock);
+		if (!--call_function.counter)
+			call_function.quiesent++;
+		else
+			busy = 1;
+		spin_unlock(&call_function.lock);
 
 		if (data->csd.flags & CSD_FLAG_WAIT) {
 			/*
@@ -138,10 +158,17 @@ void generic_smp_call_function_interrupt
 			smp_wmb();
 			data->csd.flags &= ~CSD_FLAG_WAIT;
 		}
-		if (data->csd.flags & CSD_FLAG_ALLOC)
-			call_rcu(&data->rcu_head, rcu_free_call_data);
+		if (data->csd.flags & CSD_FLAG_LOCK) {
+			smp_wmb();
+			data->csd.flags &= ~CSD_FLAG_LOCK;
+		}
+		if (data->csd.flags & CSD_FLAG_ALLOC) {
+			if (busy)
+				call_rcu(&data->rcu_head, rcu_free_call_data);
+			else
+				kfree(data);
+		}
 	}
-	rcu_read_unlock();
 
 	put_cpu();
 }
@@ -302,6 +329,8 @@ void __smp_call_function_single(int cpu,
 	arch_send_call_function_ipi(*(maskp))
 #endif
 
+static DEFINE_PER_CPU(struct call_function_data, cfd_data);
+
 /**
  * smp_call_function_many(): Run a function on a set of other CPUs.
  * @mask: The set of cpus to run on (only runs on online subset).
@@ -323,14 +352,14 @@ void smp_call_function_many(const struct
 {
 	struct call_function_data *data;
 	unsigned long flags;
-	int cpu, next_cpu;
+	int cpu, next_cpu, me = smp_processor_id();
 
 	/* Can deadlock when called with interrupts disabled */
 	WARN_ON(irqs_disabled());
 
 	/* So, what's a CPU they want?  Ignoring this one. */
 	cpu = cpumask_first_and(mask, cpu_online_mask);
-	if (cpu == smp_processor_id())
+	if (cpu == me)
 		cpu = cpumask_next_and(cpu, mask, cpu_online_mask);
 	/* No online cpus?  We're done. */
 	if (cpu >= nr_cpu_ids)
@@ -338,7 +367,7 @@ void smp_call_function_many(const struct
 
 	/* Do we have another CPU which isn't us? */
 	next_cpu = cpumask_next_and(cpu, mask, cpu_online_mask);
-	if (next_cpu == smp_processor_id())
+	if (next_cpu == me)
 		next_cpu = cpumask_next_and(next_cpu, mask, cpu_online_mask);
 
 	/* Fastpath: do that cpu by itself. */
@@ -347,31 +376,40 @@ void smp_call_function_many(const struct
 		return;
 	}
 
-	data = kmalloc(sizeof(*data) + cpumask_size(), GFP_ATOMIC);
-	if (unlikely(!data)) {
-		/* Slow path. */
-		for_each_online_cpu(cpu) {
-			if (cpu == smp_processor_id())
-				continue;
-			if (cpumask_test_cpu(cpu, mask))
-				smp_call_function_single(cpu, func, info, wait);
-		}
-		return;
+	data = kmalloc(sizeof(*data), GFP_ATOMIC);
+	if (data)
+		data->csd.flags = CSD_FLAG_ALLOC;
+	else {
+		data = &per_cpu(cfd_data, me);
+		/*
+		 * We need to wait for all previous users to go away.
+		 */
+		while (data->csd.flags & CSD_FLAG_LOCK)
+			cpu_relax();
+		/*
+		 * Then we need to wait for the queue to pass through a
+		 * quiesent state, so that no other cpus can observe the
+		 * element anymore.
+		 */
+		while (data->stamp == call_function.quiesent)
+			cpu_relax();
+		data->csd.flags = CSD_FLAG_LOCK;
 	}
 
 	spin_lock_init(&data->lock);
-	data->csd.flags = CSD_FLAG_ALLOC;
 	if (wait)
 		data->csd.flags |= CSD_FLAG_WAIT;
 	data->csd.func = func;
 	data->csd.info = info;
-	cpumask_and(to_cpumask(data->cpumask_bits), mask, cpu_online_mask);
-	cpumask_clear_cpu(smp_processor_id(), to_cpumask(data->cpumask_bits));
-	data->refs = cpumask_weight(to_cpumask(data->cpumask_bits));
-
-	spin_lock_irqsave(&call_function_lock, flags);
-	list_add_tail_rcu(&data->csd.list, &call_function_queue);
-	spin_unlock_irqrestore(&call_function_lock, flags);
+	cpumask_and(&data->cpumask, mask, cpu_online_mask);
+	cpumask_clear_cpu(smp_processor_id(), &data->cpumask);
+	data->refs = cpumask_weight(&data->cpumask);
+
+	spin_lock_irqsave(&call_function.lock, flags);
+	call_function.counter++;
+	data->stamp = call_function.quiesent;
+	list_add_tail_rcu(&data->csd.list, &call_function.queue);
+	spin_unlock_irqrestore(&call_function.lock, flags);
 
 	/*
 	 * Make the list addition visible before sending the ipi.
@@ -379,7 +417,7 @@ void smp_call_function_many(const struct
 	smp_mb();
 
 	/* Send a message to all CPUs in the map */
-	arch_send_call_function_ipi_mask(to_cpumask(data->cpumask_bits));
+	arch_send_call_function_ipi_mask(&data->cpumask);
 
 	/* optionally wait for the CPUs to complete */
 	if (wait)
@@ -413,20 +451,20 @@ EXPORT_SYMBOL(smp_call_function);
 
 void ipi_call_lock(void)
 {
-	spin_lock(&call_function_lock);
+	spin_lock(&call_function.lock);
 }
 
 void ipi_call_unlock(void)
 {
-	spin_unlock(&call_function_lock);
+	spin_unlock(&call_function.lock);
 }
 
 void ipi_call_lock_irq(void)
 {
-	spin_lock_irq(&call_function_lock);
+	spin_lock_irq(&call_function.lock);
 }
 
 void ipi_call_unlock_irq(void)
 {
-	spin_unlock_irq(&call_function_lock);
+	spin_unlock_irq(&call_function.lock);
 }

-- 


^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH 2/2] generic-smp: remove kmalloc usage
  2009-02-12 22:32 [PATCH 0/2] generic smp helpers vs kmalloc Peter Zijlstra
  2009-02-12 22:32 ` [PATCH 1/2] generic-smp: remove single ipi fallback for smp_call_function_many() Peter Zijlstra
@ 2009-02-12 22:32 ` Peter Zijlstra
  2009-02-14 14:46   ` [PATCH 2.5/2] generic-smp: remove cfd rcu_head Peter Zijlstra
  1 sibling, 1 reply; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-12 22:32 UTC (permalink / raw)
  To: Linus Torvalds, Nick Piggin, Jens Axboe, Paul E. McKenney,
	Ingo Molnar, Rusty Russell
  Cc: linux-kernel, Oleg Nesterov, Peter Zijlstra

[-- Attachment #1: smp-remove-kmalloc.patch --]
[-- Type: text/plain, Size: 4120 bytes --]

Now that there is no strict need for kmalloc anymore, and nobody seems to
rely on the queueing behaviour, remove it.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
---
 kernel/smp.c |   79 ++++++++++++++++++-----------------------------------------
 1 file changed, 25 insertions(+), 54 deletions(-)

Index: linux-2.6/kernel/smp.c
===================================================================
--- linux-2.6.orig/kernel/smp.c
+++ linux-2.6/kernel/smp.c
@@ -31,8 +31,7 @@ static struct {
 
 enum {
 	CSD_FLAG_WAIT		= 0x01,
-	CSD_FLAG_ALLOC		= 0x02,
-	CSD_FLAG_LOCK		= 0x04,
+	CSD_FLAG_LOCK		= 0x02,
 };
 
 struct call_function_data {
@@ -102,15 +101,6 @@ static void generic_exec_single(int cpu,
 		csd_flag_wait(data);
 }
 
-static void rcu_free_call_data(struct rcu_head *head)
-{
-	struct call_function_data *data;
-
-	data = container_of(head, struct call_function_data, rcu_head);
-
-	kfree(data);
-}
-
 /*
  * Invoked by arch to handle an IPI for call function. Must be called with
  * interrupts disabled.
@@ -162,12 +152,6 @@ void generic_smp_call_function_interrupt
 			smp_wmb();
 			data->csd.flags &= ~CSD_FLAG_LOCK;
 		}
-		if (data->csd.flags & CSD_FLAG_ALLOC) {
-			if (busy)
-				call_rcu(&data->rcu_head, rcu_free_call_data);
-			else
-				kfree(data);
-		}
 	}
 
 	put_cpu();
@@ -217,8 +201,7 @@ void generic_smp_call_function_single_in
 			} else if (data_flags & CSD_FLAG_LOCK) {
 				smp_wmb();
 				data->flags &= ~CSD_FLAG_LOCK;
-			} else if (data_flags & CSD_FLAG_ALLOC)
-				kfree(data);
+			}
 		}
 		/*
 		 * See comment on outer loop
@@ -263,13 +246,11 @@ int smp_call_function_single(int cpu, vo
 			/*
 			 * We are calling a function on a single CPU
 			 * and we are not going to wait for it to finish.
-			 * We first try to allocate the data, but if we
-			 * fail, we fall back to use a per cpu data to pass
-			 * the information to that CPU. Since all callers
-			 * of this code will use the same data, we must
-			 * synchronize the callers to prevent a new caller
-			 * from corrupting the data before the callee
-			 * can access it.
+			 * We use a per cpu data to pass the information to
+			 * that CPU. Since all callers of this code will
+			 * use the same data, we must synchronize the
+			 * callers to prevent a new caller from corrupting
+			 * the data before the callee can access it.
 			 *
 			 * The CSD_FLAG_LOCK is used to let us know when
 			 * the IPI handler is done with the data.
@@ -279,15 +260,10 @@ int smp_call_function_single(int cpu, vo
 			 * will make sure the callee is done with the
 			 * data before a new caller will use it.
 			 */
-			data = kmalloc(sizeof(*data), GFP_ATOMIC);
-			if (data)
-				data->flags = CSD_FLAG_ALLOC;
-			else {
-				data = &per_cpu(csd_data, me);
-				while (data->flags & CSD_FLAG_LOCK)
-					cpu_relax();
-				data->flags = CSD_FLAG_LOCK;
-			}
+			data = &per_cpu(csd_data, me);
+			while (data->flags & CSD_FLAG_LOCK)
+				cpu_relax();
+			data->flags = CSD_FLAG_LOCK;
 		} else {
 			data = &d;
 			data->flags = CSD_FLAG_WAIT;
@@ -376,25 +352,20 @@ void smp_call_function_many(const struct
 		return;
 	}
 
-	data = kmalloc(sizeof(*data), GFP_ATOMIC);
-	if (data)
-		data->csd.flags = CSD_FLAG_ALLOC;
-	else {
-		data = &per_cpu(cfd_data, me);
-		/*
-		 * We need to wait for all previous users to go away.
-		 */
-		while (data->csd.flags & CSD_FLAG_LOCK)
-			cpu_relax();
-		/*
-		 * Then we need to wait for the queue to pass through a
-		 * quiesent state, so that no other cpus can observe the
-		 * element anymore.
-		 */
-		while (data->stamp == call_function.quiesent)
-			cpu_relax();
-		data->csd.flags = CSD_FLAG_LOCK;
-	}
+	data = &per_cpu(cfd_data, me);
+	/*
+	 * We need to wait for all previous users to go away.
+	 */
+	while (data->csd.flags & CSD_FLAG_LOCK)
+		cpu_relax();
+	/*
+	 * Then we need to wait for the queue to pass through a
+	 * quiesent state, so that no other cpus can observe the
+	 * element anymore.
+	 */
+	while (data->stamp == call_function.quiesent)
+		cpu_relax();
+	data->csd.flags = CSD_FLAG_LOCK;
 
 	spin_lock_init(&data->lock);
 	if (wait)

-- 


^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH 1.5/2] generic-smp: fix initial quiesent count.
  2009-02-12 22:32 ` [PATCH 1/2] generic-smp: remove single ipi fallback for smp_call_function_many() Peter Zijlstra
@ 2009-02-14 14:41   ` Peter Zijlstra
  2009-02-14 14:46     ` Peter Zijlstra
  0 siblings, 1 reply; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-14 14:41 UTC (permalink / raw)
  To: Linus Torvalds
  Cc: Nick Piggin, Jens Axboe, Paul E. McKenney, Ingo Molnar,
	Rusty Russell, linux-kernel, Oleg Nesterov

Subject: generic-smp: fix initial quiesent count.
From: Peter Zijlstra <a.p.zijlstra@chello.nl>
Date: Sat Feb 14 15:36:07 CET 2009

If we start with a quiesent sequence count of 0, we'll match the initial stamp
values of the cfd_data, and never make any progress.

To avoid getting stuck in this situation, start out with an increased quiesent
sequence count.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
---
 kernel/smp.c |    1 +
 1 file changed, 1 insertion(+)

Index: linux-2.6/kernel/smp.c
===================================================================
--- linux-2.6.orig/kernel/smp.c
+++ linux-2.6/kernel/smp.c
@@ -27,6 +27,7 @@ static struct {
 } call_function __cacheline_aligned_in_smp = {
 	.queue = LIST_HEAD_INIT(call_function.queue),
 	.lock  = __SPIN_LOCK_UNLOCKED(call_function.lock),
+	.quiesent = 1;
 };
 
 enum {



^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH 1.5/2] generic-smp: fix initial quiesent count.
  2009-02-14 14:41   ` [PATCH 1.5/2] generic-smp: fix initial quiesent count Peter Zijlstra
@ 2009-02-14 14:46     ` Peter Zijlstra
  2009-02-14 21:14       ` Peter Zijlstra
  0 siblings, 1 reply; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-14 14:46 UTC (permalink / raw)
  To: Linus Torvalds
  Cc: Nick Piggin, Jens Axboe, Paul E. McKenney, Ingo Molnar,
	Rusty Russell, linux-kernel, Oleg Nesterov

first compile, then send out...

---
Subject: generic-smp: fix initial quiesent count.
From: Peter Zijlstra <a.p.zijlstra@chello.nl>
Date: Sat Feb 14 15:36:07 CET 2009

If we start with a quiesent sequence count of 0, we'll match the initial stamp
values of the cfd_data, and never make any progress.

To avoid getting stuck in this situation, start out with an increased quiesent
sequence count.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
---
 kernel/smp.c |    1 +
 1 file changed, 1 insertion(+)

Index: linux-2.6/kernel/smp.c
===================================================================
--- linux-2.6.orig/kernel/smp.c
+++ linux-2.6/kernel/smp.c
@@ -27,6 +27,7 @@ static struct {
 } call_function __cacheline_aligned_in_smp = {
 	.queue = LIST_HEAD_INIT(call_function.queue),
 	.lock  = __SPIN_LOCK_UNLOCKED(call_function.lock),
+	.quiesent = 1,
 };
 
 enum {



^ permalink raw reply	[flat|nested] 8+ messages in thread

* [PATCH 2.5/2] generic-smp: remove cfd rcu_head
  2009-02-12 22:32 ` [PATCH 2/2] generic-smp: remove kmalloc usage Peter Zijlstra
@ 2009-02-14 14:46   ` Peter Zijlstra
  0 siblings, 0 replies; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-14 14:46 UTC (permalink / raw)
  To: Linus Torvalds
  Cc: Nick Piggin, Jens Axboe, Paul E. McKenney, Ingo Molnar,
	Rusty Russell, linux-kernel, Oleg Nesterov

Subject: generic-smp: remove cfd rcu_head
From: Peter Zijlstra <a.p.zijlstra@chello.nl>
Date: Sat Feb 14 15:41:24 CET 2009

Since we no longer use call_rcu() on call_function_data, we can remove the
rcu_head structure from it.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
---
 kernel/smp.c |    5 +----
 1 file changed, 1 insertion(+), 4 deletions(-)

Index: linux-2.6/kernel/smp.c
===================================================================
--- linux-2.6.orig/kernel/smp.c
+++ linux-2.6/kernel/smp.c
@@ -39,10 +39,7 @@ struct call_function_data {
 	struct call_single_data csd;
 	spinlock_t lock;
 	unsigned int refs;
-	union {
-		struct rcu_head rcu_head;
-		unsigned int stamp;
-	};
+	unsigned int stamp;
 	struct cpumask cpumask;
 };
 



^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH 1.5/2] generic-smp: fix initial quiesent count.
  2009-02-14 14:46     ` Peter Zijlstra
@ 2009-02-14 21:14       ` Peter Zijlstra
  2009-02-14 23:29         ` Ingo Molnar
  0 siblings, 1 reply; 8+ messages in thread
From: Peter Zijlstra @ 2009-02-14 21:14 UTC (permalink / raw)
  To: Linus Torvalds
  Cc: Nick Piggin, Jens Axboe, Paul E. McKenney, Ingo Molnar,
	Rusty Russell, linux-kernel, Oleg Nesterov

On Sat, 2009-02-14 at 15:46 +0100, Peter Zijlstra wrote:
> first compile, then send out...
> 
> ---
> Subject: generic-smp: fix initial quiesent count.
> From: Peter Zijlstra <a.p.zijlstra@chello.nl>
> Date: Sat Feb 14 15:36:07 CET 2009
> 
> If we start with a quiesent sequence count of 0, we'll match the initial stamp
> values of the cfd_data, and never make any progress.
> 
> To avoid getting stuck in this situation, start out with an increased quiesent
> sequence count.

OK, so this is not going to fix the problem in generic. Its still
possible to end up with the original issue.

We'd need a callback list to free used entries, much like regular RCU,
this quiesent sequence count's wrapping just seems too brittle.



^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: [PATCH 1.5/2] generic-smp: fix initial quiesent count.
  2009-02-14 21:14       ` Peter Zijlstra
@ 2009-02-14 23:29         ` Ingo Molnar
  0 siblings, 0 replies; 8+ messages in thread
From: Ingo Molnar @ 2009-02-14 23:29 UTC (permalink / raw)
  To: Peter Zijlstra
  Cc: Linus Torvalds, Nick Piggin, Jens Axboe, Paul E. McKenney,
	Rusty Russell, linux-kernel, Oleg Nesterov


* Peter Zijlstra <peterz@infradead.org> wrote:

> On Sat, 2009-02-14 at 15:46 +0100, Peter Zijlstra wrote:
> > first compile, then send out...
> > 
> > ---
> > Subject: generic-smp: fix initial quiesent count.
> > From: Peter Zijlstra <a.p.zijlstra@chello.nl>
> > Date: Sat Feb 14 15:36:07 CET 2009
> > 
> > If we start with a quiesent sequence count of 0, we'll match the initial stamp
> > values of the cfd_data, and never make any progress.
> > 
> > To avoid getting stuck in this situation, start out with an increased quiesent
> > sequence count.
> 
> OK, so this is not going to fix the problem in generic. Its still
> possible to end up with the original issue.
> 
> We'd need a callback list to free used entries, much like regular RCU,
> this quiesent sequence count's wrapping just seems too brittle.

Ok, given that your series was hanging on a number of testboxes here,
lets go with my original simplification?

The broadcasting-to-a-lot-of-CPUs case is going to suck on any seriously
large system no matter what. It's single-IPI and few-IPI performance that
matters mostly.

	Ingo

^ permalink raw reply	[flat|nested] 8+ messages in thread

end of thread, other threads:[~2009-02-14 23:32 UTC | newest]

Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2009-02-12 22:32 [PATCH 0/2] generic smp helpers vs kmalloc Peter Zijlstra
2009-02-12 22:32 ` [PATCH 1/2] generic-smp: remove single ipi fallback for smp_call_function_many() Peter Zijlstra
2009-02-14 14:41   ` [PATCH 1.5/2] generic-smp: fix initial quiesent count Peter Zijlstra
2009-02-14 14:46     ` Peter Zijlstra
2009-02-14 21:14       ` Peter Zijlstra
2009-02-14 23:29         ` Ingo Molnar
2009-02-12 22:32 ` [PATCH 2/2] generic-smp: remove kmalloc usage Peter Zijlstra
2009-02-14 14:46   ` [PATCH 2.5/2] generic-smp: remove cfd rcu_head Peter Zijlstra

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.