linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH] per-cpu data preempt-safing
@ 2002-09-24 20:50 Robert Love
  0 siblings, 0 replies; only message in thread
From: Robert Love @ 2002-09-24 20:50 UTC (permalink / raw)
  To: torvalds; +Cc: linux-kernel

[-- Attachment #1: Type: text/plain, Size: 556 bytes --]

Linus,

Attached patch fixes unsafe access to per-CPU data via reordering of
instructions or use of "get_cpu()".

The following files were affected:

 include/linux/brlock.h     |    5 ++++-
 include/linux/netdevice.h  |   12 ++++++++----
 include/linux/page-flags.h |    6 +++---
 mm/highmem.c               |    2 ++
 4 files changed, 17 insertions(+), 8 deletions(-)

Before anyone balks at the brlock.h fix, note this was in the
alternative version of the code which is not used by default.

Patch is against current BK.  Please, apply.

	Robert Love


[-- Attachment #2: Type: text/x-patch, Size: 3371 bytes --]

diff -urN linux-2.5.38/include/linux/brlock.h linux/include/linux/brlock.h
--- linux-2.5.38/include/linux/brlock.h	Sun Sep 22 00:25:06 2002
+++ linux/include/linux/brlock.h	Tue Sep 24 16:24:23 2002
@@ -85,7 +85,8 @@
 	if (idx >= __BR_END)
 		__br_lock_usage_bug();
 
-	read_lock(&__brlock_array[smp_processor_id()][idx]);
+	preempt_disable();
+	_raw_read_lock(&__brlock_array[smp_processor_id()][idx]);
 }
 
 static inline void br_read_unlock (enum brlock_indices idx)
@@ -109,6 +110,7 @@
 	if (idx >= __BR_END)
 		__br_lock_usage_bug();
 
+	preempt_disable();
 	ctr = &__brlock_array[smp_processor_id()][idx];
 	lock = &__br_write_locks[idx].lock;
 again:
@@ -147,6 +149,7 @@
 
 	wmb();
 	(*ctr)--;
+	preempt_enable();
 }
 #endif /* __BRLOCK_USE_ATOMICS */
 
diff -urN linux-2.5.38/include/linux/netdevice.h linux/include/linux/netdevice.h
--- linux-2.5.38/include/linux/netdevice.h	Sun Sep 22 00:25:27 2002
+++ linux/include/linux/netdevice.h	Tue Sep 24 16:24:23 2002
@@ -514,9 +514,10 @@
 {
 	if (!test_and_set_bit(__LINK_STATE_SCHED, &dev->state)) {
 		unsigned long flags;
-		int cpu = smp_processor_id();
+		int cpu;
 
 		local_irq_save(flags);
+		cpu = smp_processor_id();
 		dev->next_sched = softnet_data[cpu].output_queue;
 		softnet_data[cpu].output_queue = dev;
 		cpu_raise_softirq(cpu, NET_TX_SOFTIRQ);
@@ -563,10 +564,11 @@
 static inline void dev_kfree_skb_irq(struct sk_buff *skb)
 {
 	if (atomic_dec_and_test(&skb->users)) {
-		int cpu =smp_processor_id();
+		int cpu;
 		unsigned long flags;
 
 		local_irq_save(flags);
+		cpu = smp_processor_id();
 		skb->next = softnet_data[cpu].completion_queue;
 		softnet_data[cpu].completion_queue = skb;
 		cpu_raise_softirq(cpu, NET_TX_SOFTIRQ);
@@ -726,9 +728,10 @@
 static inline void __netif_rx_schedule(struct net_device *dev)
 {
 	unsigned long flags;
-	int cpu = smp_processor_id();
+	int cpu;
 
 	local_irq_save(flags);
+	cpu = smp_processor_id();
 	dev_hold(dev);
 	list_add_tail(&dev->poll_list, &softnet_data[cpu].poll_list);
 	if (dev->quota < 0)
@@ -754,11 +757,12 @@
 {
 	if (netif_rx_schedule_prep(dev)) {
 		unsigned long flags;
-		int cpu = smp_processor_id();
+		int cpu;
 
 		dev->quota += undo;
 
 		local_irq_save(flags);
+		cpu = smp_processor_id();
 		list_add_tail(&dev->poll_list, &softnet_data[cpu].poll_list);
 		__cpu_raise_softirq(cpu, NET_RX_SOFTIRQ);
 		local_irq_restore(flags);
diff -urN linux-2.5.38/include/linux/page-flags.h linux/include/linux/page-flags.h
--- linux-2.5.38/include/linux/page-flags.h	Sun Sep 22 00:25:16 2002
+++ linux/include/linux/page-flags.h	Tue Sep 24 16:24:23 2002
@@ -86,9 +86,9 @@
 
 #define mod_page_state(member, delta)					\
 	do {								\
-		preempt_disable();					\
-		page_states[smp_processor_id()].member += (delta);	\
-		preempt_enable();					\
+		int cpu = get_cpu();					\
+		page_states[cpu].member += (delta);			\
+		put_cpu();						\
 	} while (0)
 
 #define inc_page_state(member)	mod_page_state(member, 1UL)
diff -urN linux-2.5.38/mm/highmem.c linux/mm/highmem.c
--- linux-2.5.38/mm/highmem.c	Sun Sep 22 00:25:12 2002
+++ linux/mm/highmem.c	Tue Sep 24 16:24:23 2002
@@ -472,6 +472,7 @@
 {
 	int idx, type;
 
+	preempt_disable();
 	for (type = 0; type < KM_TYPE_NR; type++) {
 		idx = type + KM_TYPE_NR*smp_processor_id();
 		if (!pte_none(*(kmap_pte-idx))) {
@@ -479,6 +480,7 @@
 			BUG();
 		}
 	}
+	preempt_enable();
 }
 #endif
 

^ permalink raw reply	[flat|nested] only message in thread

only message in thread, other threads:[~2002-09-24 20:47 UTC | newest]

Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2002-09-24 20:50 [PATCH] per-cpu data preempt-safing Robert Love

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).