All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH 1/1] sched/headers: remove double_lock function
@ 2022-03-29 13:05 Jianxing Wang
  0 siblings, 0 replies; only message in thread
From: Jianxing Wang @ 2022-03-29 13:05 UTC (permalink / raw)
  To: mingo, peterz, juri.lelli, vincent.guittot, dietmar.eggemann,
	rostedt, bsegall, mgorman, bristot
  Cc: linux-kernel, Jianxing Wang, Mike Galbraith

commit 8c8a743c5087b ("sched/numa: Use {cpu, pid} to create task groups
for shared faults") import double_lock but
commit 60e69eed85bb ("sched/numa: Fix task_numa_free() lockdep splat")
replace it with double_lock_irq, now nobody care it, just remove it.

Cc: Mike Galbraith <bitbucket@online.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Signed-off-by: Jianxing Wang <wangjianxing@loongson.cn>
---
 kernel/sched/sched.h | 9 ---------
 1 file changed, 9 deletions(-)

diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
index 58263f90c559..f8ca414177bd 100644
--- a/kernel/sched/sched.h
+++ b/kernel/sched/sched.h
@@ -2586,15 +2586,6 @@ static inline void double_unlock_balance(struct rq *this_rq, struct rq *busiest)
 	lock_set_subclass(&__rq_lockp(this_rq)->dep_map, 0, _RET_IP_);
 }
 
-static inline void double_lock(spinlock_t *l1, spinlock_t *l2)
-{
-	if (l1 > l2)
-		swap(l1, l2);
-
-	spin_lock(l1);
-	spin_lock_nested(l2, SINGLE_DEPTH_NESTING);
-}
-
 static inline void double_lock_irq(spinlock_t *l1, spinlock_t *l2)
 {
 	if (l1 > l2)
-- 
2.31.1


^ permalink raw reply related	[flat|nested] only message in thread

only message in thread, other threads:[~2022-03-29 13:06 UTC | newest]

Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-03-29 13:05 [PATCH 1/1] sched/headers: remove double_lock function Jianxing Wang

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.