linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [RFC] lockdep: Put graph lock/unlock under lock_recursion protection
@ 2020-11-13 11:05 Boqun Feng
  2020-11-13 11:54 ` Peter Zijlstra
  2020-11-19  9:55 ` [tip: locking/urgent] " tip-bot2 for Boqun Feng
  0 siblings, 2 replies; 3+ messages in thread
From: Boqun Feng @ 2020-11-13 11:05 UTC (permalink / raw)
  To: linux-kernel, linux-hyperv
  Cc: Boqun Feng, K. Y. Srinivasan, Haiyang Zhang, Stephen Hemminger,
	Wei Liu, Thomas Gleixner, Ingo Molnar, Borislav Petkov, x86,
	H. Peter Anvin, Peter Zijlstra, Will Deacon

A warning was hit when running xfstests/generic/068 in a Hyper-V guest:

[...] ------------[ cut here ]------------
[...] DEBUG_LOCKS_WARN_ON(lockdep_hardirqs_enabled())
[...] WARNING: CPU: 2 PID: 1350 at kernel/locking/lockdep.c:5280 check_flags.part.0+0x165/0x170
[...] ...
[...] Workqueue: events pwq_unbound_release_workfn
[...] RIP: 0010:check_flags.part.0+0x165/0x170
[...] ...
[...] Call Trace:
[...]  lock_is_held_type+0x72/0x150
[...]  ? lock_acquire+0x16e/0x4a0
[...]  rcu_read_lock_sched_held+0x3f/0x80
[...]  __send_ipi_one+0x14d/0x1b0
[...]  hv_send_ipi+0x12/0x30
[...]  __pv_queued_spin_unlock_slowpath+0xd1/0x110
[...]  __raw_callee_save___pv_queued_spin_unlock_slowpath+0x11/0x20
[...]  .slowpath+0x9/0xe
[...]  lockdep_unregister_key+0x128/0x180
[...]  pwq_unbound_release_workfn+0xbb/0xf0
[...]  process_one_work+0x227/0x5c0
[...]  worker_thread+0x55/0x3c0
[...]  ? process_one_work+0x5c0/0x5c0
[...]  kthread+0x153/0x170
[...]  ? __kthread_bind_mask+0x60/0x60
[...]  ret_from_fork+0x1f/0x30

The cause of the problem is we have call chain lockdep_unregister_key()
-> <irq disabled by raw_local_irq_save()> lockdep_unlock() ->
arch_spin_unlock() -> __pv_queued_spin_unlock_slowpath() -> pv_kick() ->
__send_ipi_one() -> trace_hyperv_send_ipi_one().

Although this particular warning is triggered because Hyper-V has a
trace point in ipi sending, but in general arch_spin_unlock() may call
another function having a trace point in it, so put the arch_spin_lock()
and arch_spin_unlock() after lock_recursion protection to fix this
problem and avoid similiar problems.

Signed-off-by: Boqun Feng <boqun.feng@gmail.com>
Cc: "K. Y. Srinivasan" <kys@microsoft.com>
Cc: Haiyang Zhang <haiyangz@microsoft.com>
Cc: Stephen Hemminger <sthemmin@microsoft.com>
Cc: Wei Liu <wei.liu@kernel.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: x86@kernel.org
Cc: "H. Peter Anvin" <hpa@zytor.com>
---
 kernel/locking/lockdep.c | 6 ++++--
 1 file changed, 4 insertions(+), 2 deletions(-)

diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c
index b71ad8d9f1c9..b98e44f88c6a 100644
--- a/kernel/locking/lockdep.c
+++ b/kernel/locking/lockdep.c
@@ -108,19 +108,21 @@ static inline void lockdep_lock(void)
 {
 	DEBUG_LOCKS_WARN_ON(!irqs_disabled());
 
+	__this_cpu_inc(lockdep_recursion);
 	arch_spin_lock(&__lock);
 	__owner = current;
-	__this_cpu_inc(lockdep_recursion);
 }
 
 static inline void lockdep_unlock(void)
 {
+	DEBUG_LOCKS_WARN_ON(!irqs_disabled());
+
 	if (debug_locks && DEBUG_LOCKS_WARN_ON(__owner != current))
 		return;
 
-	__this_cpu_dec(lockdep_recursion);
 	__owner = NULL;
 	arch_spin_unlock(&__lock);
+	__this_cpu_dec(lockdep_recursion);
 }
 
 static inline bool lockdep_assert_locked(void)
-- 
2.29.2


^ permalink raw reply related	[flat|nested] 3+ messages in thread

* Re: [RFC] lockdep: Put graph lock/unlock under lock_recursion protection
  2020-11-13 11:05 [RFC] lockdep: Put graph lock/unlock under lock_recursion protection Boqun Feng
@ 2020-11-13 11:54 ` Peter Zijlstra
  2020-11-19  9:55 ` [tip: locking/urgent] " tip-bot2 for Boqun Feng
  1 sibling, 0 replies; 3+ messages in thread
From: Peter Zijlstra @ 2020-11-13 11:54 UTC (permalink / raw)
  To: Boqun Feng
  Cc: linux-kernel, linux-hyperv, K. Y. Srinivasan, Haiyang Zhang,
	Stephen Hemminger, Wei Liu, Thomas Gleixner, Ingo Molnar,
	Borislav Petkov, x86, H. Peter Anvin, Will Deacon

On Fri, Nov 13, 2020 at 07:05:03PM +0800, Boqun Feng wrote:
> A warning was hit when running xfstests/generic/068 in a Hyper-V guest:
> 
> [...] ------------[ cut here ]------------
> [...] DEBUG_LOCKS_WARN_ON(lockdep_hardirqs_enabled())
> [...] WARNING: CPU: 2 PID: 1350 at kernel/locking/lockdep.c:5280 check_flags.part.0+0x165/0x170
> [...] ...
> [...] Workqueue: events pwq_unbound_release_workfn
> [...] RIP: 0010:check_flags.part.0+0x165/0x170
> [...] ...
> [...] Call Trace:
> [...]  lock_is_held_type+0x72/0x150
> [...]  ? lock_acquire+0x16e/0x4a0
> [...]  rcu_read_lock_sched_held+0x3f/0x80
> [...]  __send_ipi_one+0x14d/0x1b0
> [...]  hv_send_ipi+0x12/0x30
> [...]  __pv_queued_spin_unlock_slowpath+0xd1/0x110
> [...]  __raw_callee_save___pv_queued_spin_unlock_slowpath+0x11/0x20
> [...]  .slowpath+0x9/0xe
> [...]  lockdep_unregister_key+0x128/0x180
> [...]  pwq_unbound_release_workfn+0xbb/0xf0
> [...]  process_one_work+0x227/0x5c0
> [...]  worker_thread+0x55/0x3c0
> [...]  ? process_one_work+0x5c0/0x5c0
> [...]  kthread+0x153/0x170
> [...]  ? __kthread_bind_mask+0x60/0x60
> [...]  ret_from_fork+0x1f/0x30
> 
> The cause of the problem is we have call chain lockdep_unregister_key()
> -> <irq disabled by raw_local_irq_save()> lockdep_unlock() ->
> arch_spin_unlock() -> __pv_queued_spin_unlock_slowpath() -> pv_kick() ->
> __send_ipi_one() -> trace_hyperv_send_ipi_one().
> 
> Although this particular warning is triggered because Hyper-V has a
> trace point in ipi sending, but in general arch_spin_unlock() may call
> another function having a trace point in it, so put the arch_spin_lock()
> and arch_spin_unlock() after lock_recursion protection to fix this
> problem and avoid similiar problems.
> 
> Signed-off-by: Boqun Feng <boqun.feng@gmail.com>

Works for me, thanks!

^ permalink raw reply	[flat|nested] 3+ messages in thread

* [tip: locking/urgent] lockdep: Put graph lock/unlock under lock_recursion protection
  2020-11-13 11:05 [RFC] lockdep: Put graph lock/unlock under lock_recursion protection Boqun Feng
  2020-11-13 11:54 ` Peter Zijlstra
@ 2020-11-19  9:55 ` tip-bot2 for Boqun Feng
  1 sibling, 0 replies; 3+ messages in thread
From: tip-bot2 for Boqun Feng @ 2020-11-19  9:55 UTC (permalink / raw)
  To: linux-tip-commits; +Cc: Boqun Feng, Peter Zijlstra (Intel), x86, linux-kernel

The following commit has been merged into the locking/urgent branch of tip:

Commit-ID:     43be4388e94b915799a24f0eaf664bf95b85231f
Gitweb:        https://git.kernel.org/tip/43be4388e94b915799a24f0eaf664bf95b85231f
Author:        Boqun Feng <boqun.feng@gmail.com>
AuthorDate:    Fri, 13 Nov 2020 19:05:03 +08:00
Committer:     Peter Zijlstra <peterz@infradead.org>
CommitterDate: Tue, 17 Nov 2020 13:15:35 +01:00

lockdep: Put graph lock/unlock under lock_recursion protection

A warning was hit when running xfstests/generic/068 in a Hyper-V guest:

[...] ------------[ cut here ]------------
[...] DEBUG_LOCKS_WARN_ON(lockdep_hardirqs_enabled())
[...] WARNING: CPU: 2 PID: 1350 at kernel/locking/lockdep.c:5280 check_flags.part.0+0x165/0x170
[...] ...
[...] Workqueue: events pwq_unbound_release_workfn
[...] RIP: 0010:check_flags.part.0+0x165/0x170
[...] ...
[...] Call Trace:
[...]  lock_is_held_type+0x72/0x150
[...]  ? lock_acquire+0x16e/0x4a0
[...]  rcu_read_lock_sched_held+0x3f/0x80
[...]  __send_ipi_one+0x14d/0x1b0
[...]  hv_send_ipi+0x12/0x30
[...]  __pv_queued_spin_unlock_slowpath+0xd1/0x110
[...]  __raw_callee_save___pv_queued_spin_unlock_slowpath+0x11/0x20
[...]  .slowpath+0x9/0xe
[...]  lockdep_unregister_key+0x128/0x180
[...]  pwq_unbound_release_workfn+0xbb/0xf0
[...]  process_one_work+0x227/0x5c0
[...]  worker_thread+0x55/0x3c0
[...]  ? process_one_work+0x5c0/0x5c0
[...]  kthread+0x153/0x170
[...]  ? __kthread_bind_mask+0x60/0x60
[...]  ret_from_fork+0x1f/0x30

The cause of the problem is we have call chain lockdep_unregister_key()
-> <irq disabled by raw_local_irq_save()> lockdep_unlock() ->
arch_spin_unlock() -> __pv_queued_spin_unlock_slowpath() -> pv_kick() ->
__send_ipi_one() -> trace_hyperv_send_ipi_one().

Although this particular warning is triggered because Hyper-V has a
trace point in ipi sending, but in general arch_spin_unlock() may call
another function having a trace point in it, so put the arch_spin_lock()
and arch_spin_unlock() after lock_recursion protection to fix this
problem and avoid similiar problems.

Signed-off-by: Boqun Feng <boqun.feng@gmail.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Link: https://lkml.kernel.org/r/20201113110512.1056501-1-boqun.feng@gmail.com
---
 kernel/locking/lockdep.c | 6 ++++--
 1 file changed, 4 insertions(+), 2 deletions(-)

diff --git a/kernel/locking/lockdep.c b/kernel/locking/lockdep.c
index d9fb9e1..c1418b4 100644
--- a/kernel/locking/lockdep.c
+++ b/kernel/locking/lockdep.c
@@ -108,19 +108,21 @@ static inline void lockdep_lock(void)
 {
 	DEBUG_LOCKS_WARN_ON(!irqs_disabled());
 
+	__this_cpu_inc(lockdep_recursion);
 	arch_spin_lock(&__lock);
 	__owner = current;
-	__this_cpu_inc(lockdep_recursion);
 }
 
 static inline void lockdep_unlock(void)
 {
+	DEBUG_LOCKS_WARN_ON(!irqs_disabled());
+
 	if (debug_locks && DEBUG_LOCKS_WARN_ON(__owner != current))
 		return;
 
-	__this_cpu_dec(lockdep_recursion);
 	__owner = NULL;
 	arch_spin_unlock(&__lock);
+	__this_cpu_dec(lockdep_recursion);
 }
 
 static inline bool lockdep_assert_locked(void)

^ permalink raw reply related	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2020-11-19  9:55 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-11-13 11:05 [RFC] lockdep: Put graph lock/unlock under lock_recursion protection Boqun Feng
2020-11-13 11:54 ` Peter Zijlstra
2020-11-19  9:55 ` [tip: locking/urgent] " tip-bot2 for Boqun Feng

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).