All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
@ 2018-10-01  7:16 Juergen Gross
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
                   ` (7 more replies)
  0 siblings, 8 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:16 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, Juergen Gross,
	Waiman.Long, peterz

The Xen specific queue spinlock wait function has two issues which
could result in a hanging system.

They have a similar root cause of clearing a pending wakeup of a
waiting vcpu and later going to sleep waiting for the just cleared
wakeup event, which of course won't ever happen.

Juergen Gross (2):
  xen: fix race in xen_qlock_wait()
  xen: make xen_qlock_wait() nestable

 arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
 1 file changed, 12 insertions(+), 21 deletions(-)

Cc: Waiman.Long@hp.com
Cc: peterz@infradead.org

-- 
2.16.4


^ permalink raw reply	[flat|nested] 35+ messages in thread

* [PATCH 1/2] xen: fix race in xen_qlock_wait()
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
@ 2018-10-01  7:16 ` Juergen Gross
  2018-10-01  7:37   ` Juergen Gross
                     ` (3 more replies)
  2018-10-01  7:16 ` Juergen Gross
                   ` (6 subsequent siblings)
  7 siblings, 4 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:16 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, Juergen Gross, stable,
	Waiman.Long, peterz

In the following situation a vcpu waiting for a lock might not be
woken up from xen_poll_irq():

CPU 1:                CPU 2:                      CPU 3:
takes a spinlock
                      tries to get lock
                      -> xen_qlock_wait()
                        -> xen_clear_irq_pending()
frees the lock
-> xen_qlock_kick(cpu2)

takes lock again
                                                  tries to get lock
                                                  -> *lock = _Q_SLOW_VAL
                        -> *lock == _Q_SLOW_VAL ?
                        -> xen_poll_irq()
frees the lock
-> xen_qlock_kick(cpu3)

And cpu 2 will sleep forever.

This can be avoided easily by modifying xen_qlock_wait() to call
xen_poll_irq() only if the related irq was not pending and to call
xen_clear_irq_pending() only if it was pending.

Cc: stable@vger.kernel.org
Cc: Waiman.Long@hp.com
Cc: peterz@infradead.org
Signed-off-by: Juergen Gross <jgross@suse.com>
---
 arch/x86/xen/spinlock.c | 15 +++++----------
 1 file changed, 5 insertions(+), 10 deletions(-)

diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
index 973f10e05211..cd210a4ba7b1 100644
--- a/arch/x86/xen/spinlock.c
+++ b/arch/x86/xen/spinlock.c
@@ -45,17 +45,12 @@ static void xen_qlock_wait(u8 *byte, u8 val)
 	if (irq == -1)
 		return;
 
-	/* clear pending */
-	xen_clear_irq_pending(irq);
-	barrier();
+	/* If irq pending already clear it and return. */
+	if (xen_test_irq_pending(irq)) {
+		xen_clear_irq_pending(irq);
+		return;
+	}
 
-	/*
-	 * We check the byte value after clearing pending IRQ to make sure
-	 * that we won't miss a wakeup event because of the clearing.
-	 *
-	 * The sync_clear_bit() call in xen_clear_irq_pending() is atomic.
-	 * So it is effectively a memory barrier for x86.
-	 */
 	if (READ_ONCE(*byte) != val)
 		return;
 
-- 
2.16.4


^ permalink raw reply related	[flat|nested] 35+ messages in thread

* [PATCH 1/2] xen: fix race in xen_qlock_wait()
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
@ 2018-10-01  7:16 ` Juergen Gross
  2018-10-01  7:16 ` [PATCH 2/2] xen: make xen_qlock_wait() nestable Juergen Gross
                   ` (5 subsequent siblings)
  7 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:16 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: Juergen Gross, Waiman.Long, peterz, stable, mingo, bp, hpa,
	boris.ostrovsky, tglx

In the following situation a vcpu waiting for a lock might not be
woken up from xen_poll_irq():

CPU 1:                CPU 2:                      CPU 3:
takes a spinlock
                      tries to get lock
                      -> xen_qlock_wait()
                        -> xen_clear_irq_pending()
frees the lock
-> xen_qlock_kick(cpu2)

takes lock again
                                                  tries to get lock
                                                  -> *lock = _Q_SLOW_VAL
                        -> *lock == _Q_SLOW_VAL ?
                        -> xen_poll_irq()
frees the lock
-> xen_qlock_kick(cpu3)

And cpu 2 will sleep forever.

This can be avoided easily by modifying xen_qlock_wait() to call
xen_poll_irq() only if the related irq was not pending and to call
xen_clear_irq_pending() only if it was pending.

Cc: stable@vger.kernel.org
Cc: Waiman.Long@hp.com
Cc: peterz@infradead.org
Signed-off-by: Juergen Gross <jgross@suse.com>
---
 arch/x86/xen/spinlock.c | 15 +++++----------
 1 file changed, 5 insertions(+), 10 deletions(-)

diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
index 973f10e05211..cd210a4ba7b1 100644
--- a/arch/x86/xen/spinlock.c
+++ b/arch/x86/xen/spinlock.c
@@ -45,17 +45,12 @@ static void xen_qlock_wait(u8 *byte, u8 val)
 	if (irq == -1)
 		return;
 
-	/* clear pending */
-	xen_clear_irq_pending(irq);
-	barrier();
+	/* If irq pending already clear it and return. */
+	if (xen_test_irq_pending(irq)) {
+		xen_clear_irq_pending(irq);
+		return;
+	}
 
-	/*
-	 * We check the byte value after clearing pending IRQ to make sure
-	 * that we won't miss a wakeup event because of the clearing.
-	 *
-	 * The sync_clear_bit() call in xen_clear_irq_pending() is atomic.
-	 * So it is effectively a memory barrier for x86.
-	 */
 	if (READ_ONCE(*byte) != val)
 		return;
 
-- 
2.16.4


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply related	[flat|nested] 35+ messages in thread

* [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
                   ` (2 preceding siblings ...)
  2018-10-01  7:16 ` [PATCH 2/2] xen: make xen_qlock_wait() nestable Juergen Gross
@ 2018-10-01  7:16 ` Juergen Gross
  2018-10-01  7:38   ` Juergen Gross
                     ` (6 more replies)
  2018-10-01  7:37 ` [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
                   ` (3 subsequent siblings)
  7 siblings, 7 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:16 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, Juergen Gross, stable,
	Waiman.Long, peterz

xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
of xen_qlock_kick() might be ignored in case a deeper nesting level
was active right before the call of xen_poll_irq():

CPU 1:                                   CPU 2:
spin_lock(lock1)
                                         spin_lock(lock1)
                                         -> xen_qlock_wait()
                                            -> xen_clear_irq_pending()
                                            Interrupt happens
spin_unlock(lock1)
-> xen_qlock_kick(CPU 2)
spin_lock_irqsave(lock2)
                                         spin_lock_irqsave(lock2)
                                         -> xen_qlock_wait()
                                            -> xen_clear_irq_pending()
                                               clears kick for lock1
                                            -> xen_poll_irq()
spin_unlock_irq_restore(lock2)
-> xen_qlock_kick(CPU 2)
                                            wakes up
                                         spin_unlock_irq_restore(lock2)
                                         IRET
                                           resumes in xen_qlock_wait()
                                           -> xen_poll_irq()
                                           never wakes up

The solution is to disable interrupts in xen_qlock_wait() and not to
poll for the irq in case xen_qlock_wait() is called in nmi context.

Cc: stable@vger.kernel.org
Cc: Waiman.Long@hp.com
Cc: peterz@infradead.org
Signed-off-by: Juergen Gross <jgross@suse.com>
---
 arch/x86/xen/spinlock.c | 24 ++++++++++--------------
 1 file changed, 10 insertions(+), 14 deletions(-)

diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
index cd210a4ba7b1..e8d880e98057 100644
--- a/arch/x86/xen/spinlock.c
+++ b/arch/x86/xen/spinlock.c
@@ -39,29 +39,25 @@ static void xen_qlock_kick(int cpu)
  */
 static void xen_qlock_wait(u8 *byte, u8 val)
 {
+	unsigned long flags;
 	int irq = __this_cpu_read(lock_kicker_irq);
 
 	/* If kicker interrupts not initialized yet, just spin */
-	if (irq == -1)
+	if (irq == -1 || in_nmi())
 		return;
 
-	/* If irq pending already clear it and return. */
+	/* Guard against reentry. */
+	local_irq_save(flags);
+
+	/* If irq pending already clear it. */
 	if (xen_test_irq_pending(irq)) {
 		xen_clear_irq_pending(irq);
-		return;
+	} else if (READ_ONCE(*byte) == val) {
+		/* Block until irq becomes pending (or a spurious wakeup) */
+		xen_poll_irq(irq);
 	}
 
-	if (READ_ONCE(*byte) != val)
-		return;
-
-	/*
-	 * If an interrupt happens here, it will leave the wakeup irq
-	 * pending, which will cause xen_poll_irq() to return
-	 * immediately.
-	 */
-
-	/* Block until irq becomes pending (or perhaps a spurious wakeup) */
-	xen_poll_irq(irq);
+	local_irq_restore(flags);
 }
 
 static irqreturn_t dummy_handler(int irq, void *dev_id)
-- 
2.16.4


^ permalink raw reply related	[flat|nested] 35+ messages in thread

* [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
  2018-10-01  7:16 ` Juergen Gross
@ 2018-10-01  7:16 ` Juergen Gross
  2018-10-01  7:16 ` Juergen Gross
                   ` (4 subsequent siblings)
  7 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:16 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: Juergen Gross, Waiman.Long, peterz, stable, mingo, bp, hpa,
	boris.ostrovsky, tglx

xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
of xen_qlock_kick() might be ignored in case a deeper nesting level
was active right before the call of xen_poll_irq():

CPU 1:                                   CPU 2:
spin_lock(lock1)
                                         spin_lock(lock1)
                                         -> xen_qlock_wait()
                                            -> xen_clear_irq_pending()
                                            Interrupt happens
spin_unlock(lock1)
-> xen_qlock_kick(CPU 2)
spin_lock_irqsave(lock2)
                                         spin_lock_irqsave(lock2)
                                         -> xen_qlock_wait()
                                            -> xen_clear_irq_pending()
                                               clears kick for lock1
                                            -> xen_poll_irq()
spin_unlock_irq_restore(lock2)
-> xen_qlock_kick(CPU 2)
                                            wakes up
                                         spin_unlock_irq_restore(lock2)
                                         IRET
                                           resumes in xen_qlock_wait()
                                           -> xen_poll_irq()
                                           never wakes up

The solution is to disable interrupts in xen_qlock_wait() and not to
poll for the irq in case xen_qlock_wait() is called in nmi context.

Cc: stable@vger.kernel.org
Cc: Waiman.Long@hp.com
Cc: peterz@infradead.org
Signed-off-by: Juergen Gross <jgross@suse.com>
---
 arch/x86/xen/spinlock.c | 24 ++++++++++--------------
 1 file changed, 10 insertions(+), 14 deletions(-)

diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
index cd210a4ba7b1..e8d880e98057 100644
--- a/arch/x86/xen/spinlock.c
+++ b/arch/x86/xen/spinlock.c
@@ -39,29 +39,25 @@ static void xen_qlock_kick(int cpu)
  */
 static void xen_qlock_wait(u8 *byte, u8 val)
 {
+	unsigned long flags;
 	int irq = __this_cpu_read(lock_kicker_irq);
 
 	/* If kicker interrupts not initialized yet, just spin */
-	if (irq == -1)
+	if (irq == -1 || in_nmi())
 		return;
 
-	/* If irq pending already clear it and return. */
+	/* Guard against reentry. */
+	local_irq_save(flags);
+
+	/* If irq pending already clear it. */
 	if (xen_test_irq_pending(irq)) {
 		xen_clear_irq_pending(irq);
-		return;
+	} else if (READ_ONCE(*byte) == val) {
+		/* Block until irq becomes pending (or a spurious wakeup) */
+		xen_poll_irq(irq);
 	}
 
-	if (READ_ONCE(*byte) != val)
-		return;
-
-	/*
-	 * If an interrupt happens here, it will leave the wakeup irq
-	 * pending, which will cause xen_poll_irq() to return
-	 * immediately.
-	 */
-
-	/* Block until irq becomes pending (or perhaps a spurious wakeup) */
-	xen_poll_irq(irq);
+	local_irq_restore(flags);
 }
 
 static irqreturn_t dummy_handler(int irq, void *dev_id)
-- 
2.16.4


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply related	[flat|nested] 35+ messages in thread

* Re: [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
                   ` (3 preceding siblings ...)
  2018-10-01  7:16 ` Juergen Gross
@ 2018-10-01  7:37 ` Juergen Gross
  2018-10-01  7:37 ` Juergen Gross
                   ` (2 subsequent siblings)
  7 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:37 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, Borislav Petkov, peterz, Waiman Long

Correcting Waiman's mail address

On 01/10/2018 09:16, Juergen Gross wrote:
> The Xen specific queue spinlock wait function has two issues which
> could result in a hanging system.
> 
> They have a similar root cause of clearing a pending wakeup of a
> waiting vcpu and later going to sleep waiting for the just cleared
> wakeup event, which of course won't ever happen.
> 
> Juergen Gross (2):
>   xen: fix race in xen_qlock_wait()
>   xen: make xen_qlock_wait() nestable
> 
>  arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
>  1 file changed, 12 insertions(+), 21 deletions(-)
> 
> Cc: longman@redhat.com
> Cc: peterz@infradead.org
> 


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
                   ` (4 preceding siblings ...)
  2018-10-01  7:37 ` [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
@ 2018-10-01  7:37 ` Juergen Gross
  2018-10-09 14:40 ` David Woodhouse
  2018-10-09 14:40 ` David Woodhouse
  7 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:37 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: peterz, mingo, Borislav Petkov, hpa, Waiman Long, boris.ostrovsky, tglx

Correcting Waiman's mail address

On 01/10/2018 09:16, Juergen Gross wrote:
> The Xen specific queue spinlock wait function has two issues which
> could result in a hanging system.
> 
> They have a similar root cause of clearing a pending wakeup of a
> waiting vcpu and later going to sleep waiting for the just cleared
> wakeup event, which of course won't ever happen.
> 
> Juergen Gross (2):
>   xen: fix race in xen_qlock_wait()
>   xen: make xen_qlock_wait() nestable
> 
>  arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
>  1 file changed, 12 insertions(+), 21 deletions(-)
> 
> Cc: longman@redhat.com
> Cc: peterz@infradead.org
> 


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 1/2] xen: fix race in xen_qlock_wait()
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
  2018-10-01  7:37   ` Juergen Gross
@ 2018-10-01  7:37   ` Juergen Gross
  2018-10-01  8:54   ` Jan Beulich
  2018-10-01  8:54   ` [Xen-devel] " Jan Beulich
  3 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:37 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, stable, Waiman Long, peterz

Correcting Waiman's mail address

On 01/10/2018 09:16, Juergen Gross wrote:
> In the following situation a vcpu waiting for a lock might not be
> woken up from xen_poll_irq():
> 
> CPU 1:                CPU 2:                      CPU 3:
> takes a spinlock
>                       tries to get lock
>                       -> xen_qlock_wait()
>                         -> xen_clear_irq_pending()
> frees the lock
> -> xen_qlock_kick(cpu2)
> 
> takes lock again
>                                                   tries to get lock
>                                                   -> *lock = _Q_SLOW_VAL
>                         -> *lock == _Q_SLOW_VAL ?
>                         -> xen_poll_irq()
> frees the lock
> -> xen_qlock_kick(cpu3)
> 
> And cpu 2 will sleep forever.
> 
> This can be avoided easily by modifying xen_qlock_wait() to call
> xen_poll_irq() only if the related irq was not pending and to call
> xen_clear_irq_pending() only if it was pending.
> 
> Cc: stable@vger.kernel.org
> Cc: longman@redhat.com
> Cc: peterz@infradead.org
> Signed-off-by: Juergen Gross <jgross@suse.com>
> ---
>  arch/x86/xen/spinlock.c | 15 +++++----------
>  1 file changed, 5 insertions(+), 10 deletions(-)
> 
> diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
> index 973f10e05211..cd210a4ba7b1 100644
> --- a/arch/x86/xen/spinlock.c
> +++ b/arch/x86/xen/spinlock.c
> @@ -45,17 +45,12 @@ static void xen_qlock_wait(u8 *byte, u8 val)
>  	if (irq == -1)
>  		return;
>  
> -	/* clear pending */
> -	xen_clear_irq_pending(irq);
> -	barrier();
> +	/* If irq pending already clear it and return. */
> +	if (xen_test_irq_pending(irq)) {
> +		xen_clear_irq_pending(irq);
> +		return;
> +	}
>  
> -	/*
> -	 * We check the byte value after clearing pending IRQ to make sure
> -	 * that we won't miss a wakeup event because of the clearing.
> -	 *
> -	 * The sync_clear_bit() call in xen_clear_irq_pending() is atomic.
> -	 * So it is effectively a memory barrier for x86.
> -	 */
>  	if (READ_ONCE(*byte) != val)
>  		return;
>  
> 


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 1/2] xen: fix race in xen_qlock_wait()
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
@ 2018-10-01  7:37   ` Juergen Gross
  2018-10-01  7:37   ` Juergen Gross
                     ` (2 subsequent siblings)
  3 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:37 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: peterz, stable, mingo, bp, hpa, Waiman Long, boris.ostrovsky, tglx

Correcting Waiman's mail address

On 01/10/2018 09:16, Juergen Gross wrote:
> In the following situation a vcpu waiting for a lock might not be
> woken up from xen_poll_irq():
> 
> CPU 1:                CPU 2:                      CPU 3:
> takes a spinlock
>                       tries to get lock
>                       -> xen_qlock_wait()
>                         -> xen_clear_irq_pending()
> frees the lock
> -> xen_qlock_kick(cpu2)
> 
> takes lock again
>                                                   tries to get lock
>                                                   -> *lock = _Q_SLOW_VAL
>                         -> *lock == _Q_SLOW_VAL ?
>                         -> xen_poll_irq()
> frees the lock
> -> xen_qlock_kick(cpu3)
> 
> And cpu 2 will sleep forever.
> 
> This can be avoided easily by modifying xen_qlock_wait() to call
> xen_poll_irq() only if the related irq was not pending and to call
> xen_clear_irq_pending() only if it was pending.
> 
> Cc: stable@vger.kernel.org
> Cc: longman@redhat.com
> Cc: peterz@infradead.org
> Signed-off-by: Juergen Gross <jgross@suse.com>
> ---
>  arch/x86/xen/spinlock.c | 15 +++++----------
>  1 file changed, 5 insertions(+), 10 deletions(-)
> 
> diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
> index 973f10e05211..cd210a4ba7b1 100644
> --- a/arch/x86/xen/spinlock.c
> +++ b/arch/x86/xen/spinlock.c
> @@ -45,17 +45,12 @@ static void xen_qlock_wait(u8 *byte, u8 val)
>  	if (irq == -1)
>  		return;
>  
> -	/* clear pending */
> -	xen_clear_irq_pending(irq);
> -	barrier();
> +	/* If irq pending already clear it and return. */
> +	if (xen_test_irq_pending(irq)) {
> +		xen_clear_irq_pending(irq);
> +		return;
> +	}
>  
> -	/*
> -	 * We check the byte value after clearing pending IRQ to make sure
> -	 * that we won't miss a wakeup event because of the clearing.
> -	 *
> -	 * The sync_clear_bit() call in xen_clear_irq_pending() is atomic.
> -	 * So it is effectively a memory barrier for x86.
> -	 */
>  	if (READ_ONCE(*byte) != val)
>  		return;
>  
> 


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 ` Juergen Gross
  2018-10-01  7:38   ` Juergen Gross
@ 2018-10-01  7:38   ` Juergen Gross
  2018-10-01  8:57   ` Jan Beulich
                     ` (4 subsequent siblings)
  6 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:38 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, stable, Waiman Long, peterz

Correcting Waiman's mail address

On 01/10/2018 09:16, Juergen Gross wrote:
> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
> of xen_qlock_kick() might be ignored in case a deeper nesting level
> was active right before the call of xen_poll_irq():
> 
> CPU 1:                                   CPU 2:
> spin_lock(lock1)
>                                          spin_lock(lock1)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                             Interrupt happens
> spin_unlock(lock1)
> -> xen_qlock_kick(CPU 2)
> spin_lock_irqsave(lock2)
>                                          spin_lock_irqsave(lock2)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                                clears kick for lock1
>                                             -> xen_poll_irq()
> spin_unlock_irq_restore(lock2)
> -> xen_qlock_kick(CPU 2)
>                                             wakes up
>                                          spin_unlock_irq_restore(lock2)
>                                          IRET
>                                            resumes in xen_qlock_wait()
>                                            -> xen_poll_irq()
>                                            never wakes up
> 
> The solution is to disable interrupts in xen_qlock_wait() and not to
> poll for the irq in case xen_qlock_wait() is called in nmi context.
> 
> Cc: stable@vger.kernel.org
> Cc: longman@redhat.com
> Cc: peterz@infradead.org
> Signed-off-by: Juergen Gross <jgross@suse.com>
> ---
>  arch/x86/xen/spinlock.c | 24 ++++++++++--------------
>  1 file changed, 10 insertions(+), 14 deletions(-)
> 
> diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
> index cd210a4ba7b1..e8d880e98057 100644
> --- a/arch/x86/xen/spinlock.c
> +++ b/arch/x86/xen/spinlock.c
> @@ -39,29 +39,25 @@ static void xen_qlock_kick(int cpu)
>   */
>  static void xen_qlock_wait(u8 *byte, u8 val)
>  {
> +	unsigned long flags;
>  	int irq = __this_cpu_read(lock_kicker_irq);
>  
>  	/* If kicker interrupts not initialized yet, just spin */
> -	if (irq == -1)
> +	if (irq == -1 || in_nmi())
>  		return;
>  
> -	/* If irq pending already clear it and return. */
> +	/* Guard against reentry. */
> +	local_irq_save(flags);
> +
> +	/* If irq pending already clear it. */
>  	if (xen_test_irq_pending(irq)) {
>  		xen_clear_irq_pending(irq);
> -		return;
> +	} else if (READ_ONCE(*byte) == val) {
> +		/* Block until irq becomes pending (or a spurious wakeup) */
> +		xen_poll_irq(irq);
>  	}
>  
> -	if (READ_ONCE(*byte) != val)
> -		return;
> -
> -	/*
> -	 * If an interrupt happens here, it will leave the wakeup irq
> -	 * pending, which will cause xen_poll_irq() to return
> -	 * immediately.
> -	 */
> -
> -	/* Block until irq becomes pending (or perhaps a spurious wakeup) */
> -	xen_poll_irq(irq);
> +	local_irq_restore(flags);
>  }
>  
>  static irqreturn_t dummy_handler(int irq, void *dev_id)
> 


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 ` Juergen Gross
@ 2018-10-01  7:38   ` Juergen Gross
  2018-10-01  7:38   ` Juergen Gross
                     ` (5 subsequent siblings)
  6 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:38 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: peterz, stable, mingo, bp, hpa, Waiman Long, boris.ostrovsky, tglx

Correcting Waiman's mail address

On 01/10/2018 09:16, Juergen Gross wrote:
> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
> of xen_qlock_kick() might be ignored in case a deeper nesting level
> was active right before the call of xen_poll_irq():
> 
> CPU 1:                                   CPU 2:
> spin_lock(lock1)
>                                          spin_lock(lock1)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                             Interrupt happens
> spin_unlock(lock1)
> -> xen_qlock_kick(CPU 2)
> spin_lock_irqsave(lock2)
>                                          spin_lock_irqsave(lock2)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                                clears kick for lock1
>                                             -> xen_poll_irq()
> spin_unlock_irq_restore(lock2)
> -> xen_qlock_kick(CPU 2)
>                                             wakes up
>                                          spin_unlock_irq_restore(lock2)
>                                          IRET
>                                            resumes in xen_qlock_wait()
>                                            -> xen_poll_irq()
>                                            never wakes up
> 
> The solution is to disable interrupts in xen_qlock_wait() and not to
> poll for the irq in case xen_qlock_wait() is called in nmi context.
> 
> Cc: stable@vger.kernel.org
> Cc: longman@redhat.com
> Cc: peterz@infradead.org
> Signed-off-by: Juergen Gross <jgross@suse.com>
> ---
>  arch/x86/xen/spinlock.c | 24 ++++++++++--------------
>  1 file changed, 10 insertions(+), 14 deletions(-)
> 
> diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
> index cd210a4ba7b1..e8d880e98057 100644
> --- a/arch/x86/xen/spinlock.c
> +++ b/arch/x86/xen/spinlock.c
> @@ -39,29 +39,25 @@ static void xen_qlock_kick(int cpu)
>   */
>  static void xen_qlock_wait(u8 *byte, u8 val)
>  {
> +	unsigned long flags;
>  	int irq = __this_cpu_read(lock_kicker_irq);
>  
>  	/* If kicker interrupts not initialized yet, just spin */
> -	if (irq == -1)
> +	if (irq == -1 || in_nmi())
>  		return;
>  
> -	/* If irq pending already clear it and return. */
> +	/* Guard against reentry. */
> +	local_irq_save(flags);
> +
> +	/* If irq pending already clear it. */
>  	if (xen_test_irq_pending(irq)) {
>  		xen_clear_irq_pending(irq);
> -		return;
> +	} else if (READ_ONCE(*byte) == val) {
> +		/* Block until irq becomes pending (or a spurious wakeup) */
> +		xen_poll_irq(irq);
>  	}
>  
> -	if (READ_ONCE(*byte) != val)
> -		return;
> -
> -	/*
> -	 * If an interrupt happens here, it will leave the wakeup irq
> -	 * pending, which will cause xen_poll_irq() to return
> -	 * immediately.
> -	 */
> -
> -	/* Block until irq becomes pending (or perhaps a spurious wakeup) */
> -	xen_poll_irq(irq);
> +	local_irq_restore(flags);
>  }
>  
>  static irqreturn_t dummy_handler(int irq, void *dev_id)
> 


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [Xen-devel] [PATCH 1/2] xen: fix race in xen_qlock_wait()
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
                     ` (2 preceding siblings ...)
  2018-10-01  8:54   ` Jan Beulich
@ 2018-10-01  8:54   ` Jan Beulich
  3 siblings, 0 replies; 35+ messages in thread
From: Jan Beulich @ 2018-10-01  8:54 UTC (permalink / raw)
  To: Juergen Gross
  Cc: Borislav Petkov, Peter Zijlstra, the arch/x86 maintainers, tglx,
	xen-devel, Boris Ostrovsky, longman, mingo, linux-kernel, stable,
	hpa

>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
> In the following situation a vcpu waiting for a lock might not be
> woken up from xen_poll_irq():
> 
> CPU 1:                CPU 2:                      CPU 3:
> takes a spinlock
>                       tries to get lock
>                       -> xen_qlock_wait()
>                         -> xen_clear_irq_pending()

Doesn't the last line above ...

> frees the lock
> -> xen_qlock_kick(cpu2)

... need to be below here?

> takes lock again
>                                                   tries to get lock
>                                                   -> *lock = _Q_SLOW_VAL
>                         -> *lock == _Q_SLOW_VAL ?
>                         -> xen_poll_irq()
> frees the lock
> -> xen_qlock_kick(cpu3)
> 
> And cpu 2 will sleep forever.
> 
> This can be avoided easily by modifying xen_qlock_wait() to call
> xen_poll_irq() only if the related irq was not pending and to call
> xen_clear_irq_pending() only if it was pending.
> 
> Cc: stable@vger.kernel.org 
> Cc: Waiman.Long@hp.com 
> Cc: peterz@infradead.org 
> Signed-off-by: Juergen Gross <jgross@suse.com>

Patch itself
Reviewed-by: Jan Beulich <jbeulich@suse.com>

Jan



^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 1/2] xen: fix race in xen_qlock_wait()
  2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
  2018-10-01  7:37   ` Juergen Gross
  2018-10-01  7:37   ` Juergen Gross
@ 2018-10-01  8:54   ` Jan Beulich
  2018-10-01  8:54   ` [Xen-devel] " Jan Beulich
  3 siblings, 0 replies; 35+ messages in thread
From: Jan Beulich @ 2018-10-01  8:54 UTC (permalink / raw)
  To: Juergen Gross
  Cc: hpa, Peter Zijlstra, the arch/x86 maintainers, linux-kernel,
	stable, mingo, Borislav Petkov, longman, xen-devel, tglx,
	Boris Ostrovsky

>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
> In the following situation a vcpu waiting for a lock might not be
> woken up from xen_poll_irq():
> 
> CPU 1:                CPU 2:                      CPU 3:
> takes a spinlock
>                       tries to get lock
>                       -> xen_qlock_wait()
>                         -> xen_clear_irq_pending()

Doesn't the last line above ...

> frees the lock
> -> xen_qlock_kick(cpu2)

... need to be below here?

> takes lock again
>                                                   tries to get lock
>                                                   -> *lock = _Q_SLOW_VAL
>                         -> *lock == _Q_SLOW_VAL ?
>                         -> xen_poll_irq()
> frees the lock
> -> xen_qlock_kick(cpu3)
> 
> And cpu 2 will sleep forever.
> 
> This can be avoided easily by modifying xen_qlock_wait() to call
> xen_poll_irq() only if the related irq was not pending and to call
> xen_clear_irq_pending() only if it was pending.
> 
> Cc: stable@vger.kernel.org 
> Cc: Waiman.Long@hp.com 
> Cc: peterz@infradead.org 
> Signed-off-by: Juergen Gross <jgross@suse.com>

Patch itself
Reviewed-by: Jan Beulich <jbeulich@suse.com>

Jan



_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [Xen-devel] [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 ` Juergen Gross
                     ` (2 preceding siblings ...)
  2018-10-01  8:57   ` Jan Beulich
@ 2018-10-01  8:57   ` Jan Beulich
       [not found]   ` <5BB1E18802000078001ED127@suse.com>
                     ` (2 subsequent siblings)
  6 siblings, 0 replies; 35+ messages in thread
From: Jan Beulich @ 2018-10-01  8:57 UTC (permalink / raw)
  To: Juergen Gross
  Cc: Borislav Petkov, Peter Zijlstra, the arch/x86 maintainers, tglx,
	xen-devel, Boris Ostrovsky, longman, mingo, linux-kernel, stable,
	hpa

>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
> of xen_qlock_kick() might be ignored in case a deeper nesting level
> was active right before the call of xen_poll_irq():
> 
> CPU 1:                                   CPU 2:
> spin_lock(lock1)
>                                          spin_lock(lock1)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                             Interrupt happens
> spin_unlock(lock1)
> -> xen_qlock_kick(CPU 2)
> spin_lock_irqsave(lock2)
>                                          spin_lock_irqsave(lock2)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                                clears kick for lock1
>                                             -> xen_poll_irq()
> spin_unlock_irq_restore(lock2)
> -> xen_qlock_kick(CPU 2)
>                                             wakes up
>                                          spin_unlock_irq_restore(lock2)
>                                          IRET
>                                            resumes in xen_qlock_wait()
>                                            -> xen_poll_irq()
>                                            never wakes up
> 
> The solution is to disable interrupts in xen_qlock_wait() and not to
> poll for the irq in case xen_qlock_wait() is called in nmi context.

Are precautions against NMI really worthwhile? Locks acquired both
in NMI context as well as outside of it are liable to deadlock anyway,
aren't they?

Jan



^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 ` Juergen Gross
  2018-10-01  7:38   ` Juergen Gross
  2018-10-01  7:38   ` Juergen Gross
@ 2018-10-01  8:57   ` Jan Beulich
  2018-10-01  8:57   ` [Xen-devel] " Jan Beulich
                     ` (3 subsequent siblings)
  6 siblings, 0 replies; 35+ messages in thread
From: Jan Beulich @ 2018-10-01  8:57 UTC (permalink / raw)
  To: Juergen Gross
  Cc: hpa, Peter Zijlstra, the arch/x86 maintainers, linux-kernel,
	stable, mingo, Borislav Petkov, longman, xen-devel, tglx,
	Boris Ostrovsky

>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
> of xen_qlock_kick() might be ignored in case a deeper nesting level
> was active right before the call of xen_poll_irq():
> 
> CPU 1:                                   CPU 2:
> spin_lock(lock1)
>                                          spin_lock(lock1)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                             Interrupt happens
> spin_unlock(lock1)
> -> xen_qlock_kick(CPU 2)
> spin_lock_irqsave(lock2)
>                                          spin_lock_irqsave(lock2)
>                                          -> xen_qlock_wait()
>                                             -> xen_clear_irq_pending()
>                                                clears kick for lock1
>                                             -> xen_poll_irq()
> spin_unlock_irq_restore(lock2)
> -> xen_qlock_kick(CPU 2)
>                                             wakes up
>                                          spin_unlock_irq_restore(lock2)
>                                          IRET
>                                            resumes in xen_qlock_wait()
>                                            -> xen_poll_irq()
>                                            never wakes up
> 
> The solution is to disable interrupts in xen_qlock_wait() and not to
> poll for the irq in case xen_qlock_wait() is called in nmi context.

Are precautions against NMI really worthwhile? Locks acquired both
in NMI context as well as outside of it are liable to deadlock anyway,
aren't they?

Jan



_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [Xen-devel] [PATCH 2/2] xen: make xen_qlock_wait() nestable
@ 2018-10-01  9:03     ` Juergen Gross
  2018-10-01  9:18       ` Jan Beulich
  2018-10-01  9:18       ` Jan Beulich
  0 siblings, 2 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  9:03 UTC (permalink / raw)
  To: Jan Beulich
  Cc: Borislav Petkov, Peter Zijlstra, the arch/x86 maintainers,
	Thomas Gleixner, xen-devel, Boris Ostrovsky, longman, mingo,
	lkml, stable, H. Peter Anvin

On 01/10/2018 10:57, Jan Beulich wrote:
>>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
>> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
>> of xen_qlock_kick() might be ignored in case a deeper nesting level
>> was active right before the call of xen_poll_irq():
>>
>> CPU 1:                                   CPU 2:
>> spin_lock(lock1)
>>                                          spin_lock(lock1)
>>                                          -> xen_qlock_wait()
>>                                             -> xen_clear_irq_pending()
>>                                             Interrupt happens
>> spin_unlock(lock1)
>> -> xen_qlock_kick(CPU 2)
>> spin_lock_irqsave(lock2)
>>                                          spin_lock_irqsave(lock2)
>>                                          -> xen_qlock_wait()
>>                                             -> xen_clear_irq_pending()
>>                                                clears kick for lock1
>>                                             -> xen_poll_irq()
>> spin_unlock_irq_restore(lock2)
>> -> xen_qlock_kick(CPU 2)
>>                                             wakes up
>>                                          spin_unlock_irq_restore(lock2)
>>                                          IRET
>>                                            resumes in xen_qlock_wait()
>>                                            -> xen_poll_irq()
>>                                            never wakes up
>>
>> The solution is to disable interrupts in xen_qlock_wait() and not to
>> poll for the irq in case xen_qlock_wait() is called in nmi context.
> 
> Are precautions against NMI really worthwhile? Locks acquired both
> in NMI context as well as outside of it are liable to deadlock anyway,
> aren't they?

The locks don't need to be the same. A NMI-only lock tried to be
acquired with xen_qlock_wait() for another lock having been interrupted
by the NMI will be enough to risk the issue.

So yes, I believe the test for NMI is good to have.


Juergen


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [Xen-devel] [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  9:03     ` Juergen Gross
@ 2018-10-01  9:18       ` Jan Beulich
  2018-10-01  9:18       ` Jan Beulich
  1 sibling, 0 replies; 35+ messages in thread
From: Jan Beulich @ 2018-10-01  9:18 UTC (permalink / raw)
  To: Juergen Gross
  Cc: Borislav Petkov, Peter Zijlstra, the arch/x86 maintainers, tglx,
	xen-devel, Boris Ostrovsky, longman, mingo, linux-kernel, stable,
	hpa

>>> On 01.10.18 at 11:03, <jgross@suse.com> wrote:
> On 01/10/2018 10:57, Jan Beulich wrote:
>>>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
>>> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
>>> of xen_qlock_kick() might be ignored in case a deeper nesting level
>>> was active right before the call of xen_poll_irq():
>>>
>>> CPU 1:                                   CPU 2:
>>> spin_lock(lock1)
>>>                                          spin_lock(lock1)
>>>                                          -> xen_qlock_wait()
>>>                                             -> xen_clear_irq_pending()
>>>                                             Interrupt happens
>>> spin_unlock(lock1)
>>> -> xen_qlock_kick(CPU 2)
>>> spin_lock_irqsave(lock2)
>>>                                          spin_lock_irqsave(lock2)
>>>                                          -> xen_qlock_wait()
>>>                                             -> xen_clear_irq_pending()
>>>                                                clears kick for lock1
>>>                                             -> xen_poll_irq()
>>> spin_unlock_irq_restore(lock2)
>>> -> xen_qlock_kick(CPU 2)
>>>                                             wakes up
>>>                                          spin_unlock_irq_restore(lock2)
>>>                                          IRET
>>>                                            resumes in xen_qlock_wait()
>>>                                            -> xen_poll_irq()
>>>                                            never wakes up
>>>
>>> The solution is to disable interrupts in xen_qlock_wait() and not to
>>> poll for the irq in case xen_qlock_wait() is called in nmi context.
>> 
>> Are precautions against NMI really worthwhile? Locks acquired both
>> in NMI context as well as outside of it are liable to deadlock anyway,
>> aren't they?
> 
> The locks don't need to be the same. A NMI-only lock tried to be
> acquired with xen_qlock_wait() for another lock having been interrupted
> by the NMI will be enough to risk the issue.

Ah, right. In which case
Reviewed-by: Jan Beulich <jbeulich@suse.com>

Jan



^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  9:03     ` Juergen Gross
  2018-10-01  9:18       ` Jan Beulich
@ 2018-10-01  9:18       ` Jan Beulich
  1 sibling, 0 replies; 35+ messages in thread
From: Jan Beulich @ 2018-10-01  9:18 UTC (permalink / raw)
  To: Juergen Gross
  Cc: hpa, Peter Zijlstra, the arch/x86 maintainers, linux-kernel,
	stable, mingo, Borislav Petkov, longman, xen-devel, tglx,
	Boris Ostrovsky

>>> On 01.10.18 at 11:03, <jgross@suse.com> wrote:
> On 01/10/2018 10:57, Jan Beulich wrote:
>>>>> On 01.10.18 at 09:16, <jgross@suse.com> wrote:
>>> xen_qlock_wait() isn't safe for nested calls due to interrupts. A call
>>> of xen_qlock_kick() might be ignored in case a deeper nesting level
>>> was active right before the call of xen_poll_irq():
>>>
>>> CPU 1:                                   CPU 2:
>>> spin_lock(lock1)
>>>                                          spin_lock(lock1)
>>>                                          -> xen_qlock_wait()
>>>                                             -> xen_clear_irq_pending()
>>>                                             Interrupt happens
>>> spin_unlock(lock1)
>>> -> xen_qlock_kick(CPU 2)
>>> spin_lock_irqsave(lock2)
>>>                                          spin_lock_irqsave(lock2)
>>>                                          -> xen_qlock_wait()
>>>                                             -> xen_clear_irq_pending()
>>>                                                clears kick for lock1
>>>                                             -> xen_poll_irq()
>>> spin_unlock_irq_restore(lock2)
>>> -> xen_qlock_kick(CPU 2)
>>>                                             wakes up
>>>                                          spin_unlock_irq_restore(lock2)
>>>                                          IRET
>>>                                            resumes in xen_qlock_wait()
>>>                                            -> xen_poll_irq()
>>>                                            never wakes up
>>>
>>> The solution is to disable interrupts in xen_qlock_wait() and not to
>>> poll for the irq in case xen_qlock_wait() is called in nmi context.
>> 
>> Are precautions against NMI really worthwhile? Locks acquired both
>> in NMI context as well as outside of it are liable to deadlock anyway,
>> aren't they?
> 
> The locks don't need to be the same. A NMI-only lock tried to be
> acquired with xen_qlock_wait() for another lock having been interrupted
> by the NMI will be enough to risk the issue.

Ah, right. In which case
Reviewed-by: Jan Beulich <jbeulich@suse.com>

Jan



_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
                   ` (5 preceding siblings ...)
  2018-10-01  7:37 ` Juergen Gross
@ 2018-10-09 14:40 ` David Woodhouse
  2018-10-09 14:52   ` Juergen Gross
  2018-10-09 14:52   ` Juergen Gross
  2018-10-09 14:40 ` David Woodhouse
  7 siblings, 2 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-09 14:40 UTC (permalink / raw)
  To: Juergen Gross, linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, Waiman.Long, peterz

[-- Attachment #1: Type: text/plain, Size: 720 bytes --]

On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> The Xen specific queue spinlock wait function has two issues which
> could result in a hanging system.
> 
> They have a similar root cause of clearing a pending wakeup of a
> waiting vcpu and later going to sleep waiting for the just cleared
> wakeup event, which of course won't ever happen.
> 
> Juergen Gross (2):
>   xen: fix race in xen_qlock_wait()
>   xen: make xen_qlock_wait() nestable
> 
>  arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
>  1 file changed, 12 insertions(+), 21 deletions(-)
> 
> Cc: Waiman.Long@hp.com
> Cc: peterz@infradead.org

LGTM. Both these should be Cc:stable@vger.kernel.org, yes?

Thanks.

[-- Attachment #2: smime.p7s --]
[-- Type: application/x-pkcs7-signature, Size: 5213 bytes --]

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
  2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
                   ` (6 preceding siblings ...)
  2018-10-09 14:40 ` David Woodhouse
@ 2018-10-09 14:40 ` David Woodhouse
  7 siblings, 0 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-09 14:40 UTC (permalink / raw)
  To: Juergen Gross, linux-kernel, xen-devel, x86
  Cc: Waiman.Long, peterz, mingo, bp, hpa, boris.ostrovsky, tglx


[-- Attachment #1.1: Type: text/plain, Size: 720 bytes --]

On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> The Xen specific queue spinlock wait function has two issues which
> could result in a hanging system.
> 
> They have a similar root cause of clearing a pending wakeup of a
> waiting vcpu and later going to sleep waiting for the just cleared
> wakeup event, which of course won't ever happen.
> 
> Juergen Gross (2):
>   xen: fix race in xen_qlock_wait()
>   xen: make xen_qlock_wait() nestable
> 
>  arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
>  1 file changed, 12 insertions(+), 21 deletions(-)
> 
> Cc: Waiman.Long@hp.com
> Cc: peterz@infradead.org

LGTM. Both these should be Cc:stable@vger.kernel.org, yes?

Thanks.

[-- Attachment #1.2: smime.p7s --]
[-- Type: application/x-pkcs7-signature, Size: 5213 bytes --]

[-- Attachment #2: Type: text/plain, Size: 157 bytes --]

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
  2018-10-09 14:40 ` David Woodhouse
  2018-10-09 14:52   ` Juergen Gross
@ 2018-10-09 14:52   ` Juergen Gross
  1 sibling, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-09 14:52 UTC (permalink / raw)
  To: David Woodhouse, linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, Waiman.Long, peterz

On 09/10/2018 16:40, David Woodhouse wrote:
> On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
>> The Xen specific queue spinlock wait function has two issues which
>> could result in a hanging system.
>>
>> They have a similar root cause of clearing a pending wakeup of a
>> waiting vcpu and later going to sleep waiting for the just cleared
>> wakeup event, which of course won't ever happen.
>>
>> Juergen Gross (2):
>>   xen: fix race in xen_qlock_wait()
>>   xen: make xen_qlock_wait() nestable
>>
>>  arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
>>  1 file changed, 12 insertions(+), 21 deletions(-)
>>
>> Cc: Waiman.Long@hp.com
>> Cc: peterz@infradead.org
> 
> LGTM. Both these should be Cc:stable@vger.kernel.org, yes?

Yes, they are.

I have them already queued in the Xen tree. As the bug is rather
old I didn't want to rush the patches in so late in the rc phase of
4.19.


Juergen

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
  2018-10-09 14:40 ` David Woodhouse
@ 2018-10-09 14:52   ` Juergen Gross
  2018-10-09 14:52   ` Juergen Gross
  1 sibling, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-09 14:52 UTC (permalink / raw)
  To: David Woodhouse, linux-kernel, xen-devel, x86
  Cc: Waiman.Long, peterz, mingo, bp, hpa, boris.ostrovsky, tglx

On 09/10/2018 16:40, David Woodhouse wrote:
> On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
>> The Xen specific queue spinlock wait function has two issues which
>> could result in a hanging system.
>>
>> They have a similar root cause of clearing a pending wakeup of a
>> waiting vcpu and later going to sleep waiting for the just cleared
>> wakeup event, which of course won't ever happen.
>>
>> Juergen Gross (2):
>>   xen: fix race in xen_qlock_wait()
>>   xen: make xen_qlock_wait() nestable
>>
>>  arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
>>  1 file changed, 12 insertions(+), 21 deletions(-)
>>
>> Cc: Waiman.Long@hp.com
>> Cc: peterz@infradead.org
> 
> LGTM. Both these should be Cc:stable@vger.kernel.org, yes?

Yes, they are.

I have them already queued in the Xen tree. As the bug is rather
old I didn't want to rush the patches in so late in the rc phase of
4.19.


Juergen

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 ` Juergen Gross
                     ` (4 preceding siblings ...)
       [not found]   ` <5BB1E18802000078001ED127@suse.com>
@ 2018-10-10 11:53   ` David Woodhouse
  2018-10-10 12:30     ` Thomas Gleixner
  2018-10-10 12:30     ` Thomas Gleixner
  2018-10-10 11:53   ` David Woodhouse
  6 siblings, 2 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-10 11:53 UTC (permalink / raw)
  To: Juergen Gross, linux-kernel, xen-devel, x86
  Cc: boris.ostrovsky, hpa, tglx, mingo, bp, stable, Waiman.Long, peterz

[-- Attachment #1: Type: text/plain, Size: 770 bytes --]

On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> -       /* If irq pending already clear it and return. */
> +       /* Guard against reentry. */
> +       local_irq_save(flags);
> +
> +       /* If irq pending already clear it. */
>         if (xen_test_irq_pending(irq)) {
>                 xen_clear_irq_pending(irq);
> -               return;
> +       } else if (READ_ONCE(*byte) == val) {
> +               /* Block until irq becomes pending (or a spurious wakeup) */
> +               xen_poll_irq(irq);
>         }


Does this still allow other IRQs to wake it from xen_poll_irq()?

In the case where process-context code is spinning for a lock without
disabling interrupts, we *should* allow interrupts to occur still...
does this?

[-- Attachment #2: smime.p7s --]
[-- Type: application/x-pkcs7-signature, Size: 5213 bytes --]

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-01  7:16 ` Juergen Gross
                     ` (5 preceding siblings ...)
  2018-10-10 11:53   ` David Woodhouse
@ 2018-10-10 11:53   ` David Woodhouse
  6 siblings, 0 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-10 11:53 UTC (permalink / raw)
  To: Juergen Gross, linux-kernel, xen-devel, x86
  Cc: Waiman.Long, peterz, stable, mingo, bp, hpa, boris.ostrovsky, tglx


[-- Attachment #1.1: Type: text/plain, Size: 770 bytes --]

On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> -       /* If irq pending already clear it and return. */
> +       /* Guard against reentry. */
> +       local_irq_save(flags);
> +
> +       /* If irq pending already clear it. */
>         if (xen_test_irq_pending(irq)) {
>                 xen_clear_irq_pending(irq);
> -               return;
> +       } else if (READ_ONCE(*byte) == val) {
> +               /* Block until irq becomes pending (or a spurious wakeup) */
> +               xen_poll_irq(irq);
>         }


Does this still allow other IRQs to wake it from xen_poll_irq()?

In the case where process-context code is spinning for a lock without
disabling interrupts, we *should* allow interrupts to occur still...
does this?

[-- Attachment #1.2: smime.p7s --]
[-- Type: application/x-pkcs7-signature, Size: 5213 bytes --]

[-- Attachment #2: Type: text/plain, Size: 157 bytes --]

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 11:53   ` David Woodhouse
  2018-10-10 12:30     ` Thomas Gleixner
@ 2018-10-10 12:30     ` Thomas Gleixner
  2018-10-10 12:44       ` David Woodhouse
  2018-10-10 12:44       ` David Woodhouse
  1 sibling, 2 replies; 35+ messages in thread
From: Thomas Gleixner @ 2018-10-10 12:30 UTC (permalink / raw)
  To: David Woodhouse
  Cc: Juergen Gross, linux-kernel, xen-devel, x86, boris.ostrovsky,
	hpa, mingo, bp, stable, Waiman.Long, peterz

On Wed, 10 Oct 2018, David Woodhouse wrote:

> On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> > -       /* If irq pending already clear it and return. */
> > +       /* Guard against reentry. */
> > +       local_irq_save(flags);
> > +
> > +       /* If irq pending already clear it. */
> >         if (xen_test_irq_pending(irq)) {
> >                 xen_clear_irq_pending(irq);
> > -               return;
> > +       } else if (READ_ONCE(*byte) == val) {
> > +               /* Block until irq becomes pending (or a spurious wakeup) */
> > +               xen_poll_irq(irq);
> >         }
> 
> 
> Does this still allow other IRQs to wake it from xen_poll_irq()?
> 
> In the case where process-context code is spinning for a lock without
> disabling interrupts, we *should* allow interrupts to occur still...
> does this?

Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
checking the condition and then the hardware or in this case the hypervisor
has to bring you back when an interrupt is raised.

If that would not work then the check would be racy, because the interrupt
could hit and be handled after the check and before going into
HLT/WFI/hypercall and then the thing is out until the next interrupt comes
along, which might be never.

Thanks,

	tglx


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 11:53   ` David Woodhouse
@ 2018-10-10 12:30     ` Thomas Gleixner
  2018-10-10 12:30     ` Thomas Gleixner
  1 sibling, 0 replies; 35+ messages in thread
From: Thomas Gleixner @ 2018-10-10 12:30 UTC (permalink / raw)
  To: David Woodhouse
  Cc: Juergen Gross, Waiman.Long, peterz, x86, linux-kernel, stable,
	mingo, bp, hpa, xen-devel, boris.ostrovsky

On Wed, 10 Oct 2018, David Woodhouse wrote:

> On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> > -       /* If irq pending already clear it and return. */
> > +       /* Guard against reentry. */
> > +       local_irq_save(flags);
> > +
> > +       /* If irq pending already clear it. */
> >         if (xen_test_irq_pending(irq)) {
> >                 xen_clear_irq_pending(irq);
> > -               return;
> > +       } else if (READ_ONCE(*byte) == val) {
> > +               /* Block until irq becomes pending (or a spurious wakeup) */
> > +               xen_poll_irq(irq);
> >         }
> 
> 
> Does this still allow other IRQs to wake it from xen_poll_irq()?
> 
> In the case where process-context code is spinning for a lock without
> disabling interrupts, we *should* allow interrupts to occur still...
> does this?

Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
checking the condition and then the hardware or in this case the hypervisor
has to bring you back when an interrupt is raised.

If that would not work then the check would be racy, because the interrupt
could hit and be handled after the check and before going into
HLT/WFI/hypercall and then the thing is out until the next interrupt comes
along, which might be never.

Thanks,

	tglx


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 12:30     ` Thomas Gleixner
@ 2018-10-10 12:44       ` David Woodhouse
  2018-10-10 12:47         ` Thomas Gleixner
  2018-10-10 12:47         ` Thomas Gleixner
  2018-10-10 12:44       ` David Woodhouse
  1 sibling, 2 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-10 12:44 UTC (permalink / raw)
  To: Thomas Gleixner
  Cc: Juergen Gross, linux-kernel, xen-devel, x86, boris.ostrovsky,
	hpa, mingo, bp, stable, Waiman.Long, peterz

[-- Attachment #1: Type: text/plain, Size: 1739 bytes --]

On Wed, 2018-10-10 at 14:30 +0200, Thomas Gleixner wrote:
> On Wed, 10 Oct 2018, David Woodhouse wrote:
> 
> > On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> > > -       /* If irq pending already clear it and return. */
> > > +       /* Guard against reentry. */
> > > +       local_irq_save(flags);
> > > +
> > > +       /* If irq pending already clear it. */
> > >         if (xen_test_irq_pending(irq)) {
> > >                 xen_clear_irq_pending(irq);
> > > -               return;
> > > +       } else if (READ_ONCE(*byte) == val) {
> > > +               /* Block until irq becomes pending (or a spurious wakeup) */
> > > +               xen_poll_irq(irq);
> > >         }
> > 
> > 
> > Does this still allow other IRQs to wake it from xen_poll_irq()?
> > 
> > In the case where process-context code is spinning for a lock without
> > disabling interrupts, we *should* allow interrupts to occur still...
> > does this?
> 
> Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
> checking the condition and then the hardware or in this case the hypervisor
> has to bring you back when an interrupt is raised.
> 
> If that would not work then the check would be racy, because the interrupt
> could hit and be handled after the check and before going into
> HLT/WFI/hypercall and then the thing is out until the next interrupt comes
> along, which might be never.

Right, but in this case we're calling into the hypervisor to poll for
one *specific* IRQ. Everything you say is true for that specific IRQ.

My question is what happens to *other* IRQs. We want them, but are they
masked? I'm staring at the Xen do_poll() code and haven't quite worked
that out...


[-- Attachment #2: smime.p7s --]
[-- Type: application/x-pkcs7-signature, Size: 5213 bytes --]

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 12:30     ` Thomas Gleixner
  2018-10-10 12:44       ` David Woodhouse
@ 2018-10-10 12:44       ` David Woodhouse
  1 sibling, 0 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-10 12:44 UTC (permalink / raw)
  To: Thomas Gleixner
  Cc: Juergen Gross, Waiman.Long, peterz, x86, linux-kernel, stable,
	mingo, bp, hpa, xen-devel, boris.ostrovsky


[-- Attachment #1.1: Type: text/plain, Size: 1739 bytes --]

On Wed, 2018-10-10 at 14:30 +0200, Thomas Gleixner wrote:
> On Wed, 10 Oct 2018, David Woodhouse wrote:
> 
> > On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> > > -       /* If irq pending already clear it and return. */
> > > +       /* Guard against reentry. */
> > > +       local_irq_save(flags);
> > > +
> > > +       /* If irq pending already clear it. */
> > >         if (xen_test_irq_pending(irq)) {
> > >                 xen_clear_irq_pending(irq);
> > > -               return;
> > > +       } else if (READ_ONCE(*byte) == val) {
> > > +               /* Block until irq becomes pending (or a spurious wakeup) */
> > > +               xen_poll_irq(irq);
> > >         }
> > 
> > 
> > Does this still allow other IRQs to wake it from xen_poll_irq()?
> > 
> > In the case where process-context code is spinning for a lock without
> > disabling interrupts, we *should* allow interrupts to occur still...
> > does this?
> 
> Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
> checking the condition and then the hardware or in this case the hypervisor
> has to bring you back when an interrupt is raised.
> 
> If that would not work then the check would be racy, because the interrupt
> could hit and be handled after the check and before going into
> HLT/WFI/hypercall and then the thing is out until the next interrupt comes
> along, which might be never.

Right, but in this case we're calling into the hypervisor to poll for
one *specific* IRQ. Everything you say is true for that specific IRQ.

My question is what happens to *other* IRQs. We want them, but are they
masked? I'm staring at the Xen do_poll() code and haven't quite worked
that out...


[-- Attachment #1.2: smime.p7s --]
[-- Type: application/x-pkcs7-signature, Size: 5213 bytes --]

[-- Attachment #2: Type: text/plain, Size: 157 bytes --]

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 12:44       ` David Woodhouse
  2018-10-10 12:47         ` Thomas Gleixner
@ 2018-10-10 12:47         ` Thomas Gleixner
  2018-10-10 13:38           ` Juergen Gross
  2018-10-10 13:38           ` Juergen Gross
  1 sibling, 2 replies; 35+ messages in thread
From: Thomas Gleixner @ 2018-10-10 12:47 UTC (permalink / raw)
  To: David Woodhouse
  Cc: Juergen Gross, linux-kernel, xen-devel, x86, boris.ostrovsky,
	hpa, mingo, bp, stable, Waiman.Long, peterz

On Wed, 10 Oct 2018, David Woodhouse wrote:
> On Wed, 2018-10-10 at 14:30 +0200, Thomas Gleixner wrote:
> > On Wed, 10 Oct 2018, David Woodhouse wrote:
> > 
> > > On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> > > > -       /* If irq pending already clear it and return. */
> > > > +       /* Guard against reentry. */
> > > > +       local_irq_save(flags);
> > > > +
> > > > +       /* If irq pending already clear it. */
> > > >         if (xen_test_irq_pending(irq)) {
> > > >                 xen_clear_irq_pending(irq);
> > > > -               return;
> > > > +       } else if (READ_ONCE(*byte) == val) {
> > > > +               /* Block until irq becomes pending (or a spurious wakeup) */
> > > > +               xen_poll_irq(irq);
> > > >         }
> > > 
> > > 
> > > Does this still allow other IRQs to wake it from xen_poll_irq()?
> > > 
> > > In the case where process-context code is spinning for a lock without
> > > disabling interrupts, we *should* allow interrupts to occur still...
> > > does this?
> > 
> > Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
> > checking the condition and then the hardware or in this case the hypervisor
> > has to bring you back when an interrupt is raised.
> > 
> > If that would not work then the check would be racy, because the interrupt
> > could hit and be handled after the check and before going into
> > HLT/WFI/hypercall and then the thing is out until the next interrupt comes
> > along, which might be never.
> 
> Right, but in this case we're calling into the hypervisor to poll for
> one *specific* IRQ. Everything you say is true for that specific IRQ.
> 
> My question is what happens to *other* IRQs. We want them, but are they
> masked? I'm staring at the Xen do_poll() code and haven't quite worked
> that out...

Ah, sorry. That of course has to come back like HLT/WFI for any interrupt,
but I have no idea what the Xen HV is doing there.

Thanks,

	tglx

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 12:44       ` David Woodhouse
@ 2018-10-10 12:47         ` Thomas Gleixner
  2018-10-10 12:47         ` Thomas Gleixner
  1 sibling, 0 replies; 35+ messages in thread
From: Thomas Gleixner @ 2018-10-10 12:47 UTC (permalink / raw)
  To: David Woodhouse
  Cc: Juergen Gross, Waiman.Long, peterz, x86, linux-kernel, stable,
	mingo, bp, hpa, xen-devel, boris.ostrovsky

On Wed, 10 Oct 2018, David Woodhouse wrote:
> On Wed, 2018-10-10 at 14:30 +0200, Thomas Gleixner wrote:
> > On Wed, 10 Oct 2018, David Woodhouse wrote:
> > 
> > > On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
> > > > -       /* If irq pending already clear it and return. */
> > > > +       /* Guard against reentry. */
> > > > +       local_irq_save(flags);
> > > > +
> > > > +       /* If irq pending already clear it. */
> > > >         if (xen_test_irq_pending(irq)) {
> > > >                 xen_clear_irq_pending(irq);
> > > > -               return;
> > > > +       } else if (READ_ONCE(*byte) == val) {
> > > > +               /* Block until irq becomes pending (or a spurious wakeup) */
> > > > +               xen_poll_irq(irq);
> > > >         }
> > > 
> > > 
> > > Does this still allow other IRQs to wake it from xen_poll_irq()?
> > > 
> > > In the case where process-context code is spinning for a lock without
> > > disabling interrupts, we *should* allow interrupts to occur still...
> > > does this?
> > 
> > Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
> > checking the condition and then the hardware or in this case the hypervisor
> > has to bring you back when an interrupt is raised.
> > 
> > If that would not work then the check would be racy, because the interrupt
> > could hit and be handled after the check and before going into
> > HLT/WFI/hypercall and then the thing is out until the next interrupt comes
> > along, which might be never.
> 
> Right, but in this case we're calling into the hypervisor to poll for
> one *specific* IRQ. Everything you say is true for that specific IRQ.
> 
> My question is what happens to *other* IRQs. We want them, but are they
> masked? I'm staring at the Xen do_poll() code and haven't quite worked
> that out...

Ah, sorry. That of course has to come back like HLT/WFI for any interrupt,
but I have no idea what the Xen HV is doing there.

Thanks,

	tglx

_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 12:47         ` Thomas Gleixner
@ 2018-10-10 13:38           ` Juergen Gross
  2018-10-10 13:53             ` David Woodhouse
  2018-10-10 13:53             ` David Woodhouse
  2018-10-10 13:38           ` Juergen Gross
  1 sibling, 2 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-10 13:38 UTC (permalink / raw)
  To: Thomas Gleixner, David Woodhouse
  Cc: linux-kernel, xen-devel, x86, boris.ostrovsky, hpa, mingo, bp,
	stable, Waiman.Long, peterz

On 10/10/2018 14:47, Thomas Gleixner wrote:
> On Wed, 10 Oct 2018, David Woodhouse wrote:
>> On Wed, 2018-10-10 at 14:30 +0200, Thomas Gleixner wrote:
>>> On Wed, 10 Oct 2018, David Woodhouse wrote:
>>>
>>>> On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
>>>>> -       /* If irq pending already clear it and return. */
>>>>> +       /* Guard against reentry. */
>>>>> +       local_irq_save(flags);
>>>>> +
>>>>> +       /* If irq pending already clear it. */
>>>>>         if (xen_test_irq_pending(irq)) {
>>>>>                 xen_clear_irq_pending(irq);
>>>>> -               return;
>>>>> +       } else if (READ_ONCE(*byte) == val) {
>>>>> +               /* Block until irq becomes pending (or a spurious wakeup) */
>>>>> +               xen_poll_irq(irq);
>>>>>         }
>>>>
>>>>
>>>> Does this still allow other IRQs to wake it from xen_poll_irq()?
>>>>
>>>> In the case where process-context code is spinning for a lock without
>>>> disabling interrupts, we *should* allow interrupts to occur still...
>>>> does this?
>>>
>>> Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
>>> checking the condition and then the hardware or in this case the hypervisor
>>> has to bring you back when an interrupt is raised.
>>>
>>> If that would not work then the check would be racy, because the interrupt
>>> could hit and be handled after the check and before going into
>>> HLT/WFI/hypercall and then the thing is out until the next interrupt comes
>>> along, which might be never.
>>
>> Right, but in this case we're calling into the hypervisor to poll for
>> one *specific* IRQ. Everything you say is true for that specific IRQ.
>>
>> My question is what happens to *other* IRQs. We want them, but are they
>> masked? I'm staring at the Xen do_poll() code and haven't quite worked
>> that out...
> 
> Ah, sorry. That of course has to come back like HLT/WFI for any interrupt,
> but I have no idea what the Xen HV is doing there.

The Xen HV is doing it right. It is blocking the vcpu in do_poll() and
any interrupt will unblock it.


Juergen


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 12:47         ` Thomas Gleixner
  2018-10-10 13:38           ` Juergen Gross
@ 2018-10-10 13:38           ` Juergen Gross
  1 sibling, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-10 13:38 UTC (permalink / raw)
  To: Thomas Gleixner, David Woodhouse
  Cc: Waiman.Long, peterz, x86, linux-kernel, stable, mingo, bp, hpa,
	xen-devel, boris.ostrovsky

On 10/10/2018 14:47, Thomas Gleixner wrote:
> On Wed, 10 Oct 2018, David Woodhouse wrote:
>> On Wed, 2018-10-10 at 14:30 +0200, Thomas Gleixner wrote:
>>> On Wed, 10 Oct 2018, David Woodhouse wrote:
>>>
>>>> On Mon, 2018-10-01 at 09:16 +0200, Juergen Gross wrote:
>>>>> -       /* If irq pending already clear it and return. */
>>>>> +       /* Guard against reentry. */
>>>>> +       local_irq_save(flags);
>>>>> +
>>>>> +       /* If irq pending already clear it. */
>>>>>         if (xen_test_irq_pending(irq)) {
>>>>>                 xen_clear_irq_pending(irq);
>>>>> -               return;
>>>>> +       } else if (READ_ONCE(*byte) == val) {
>>>>> +               /* Block until irq becomes pending (or a spurious wakeup) */
>>>>> +               xen_poll_irq(irq);
>>>>>         }
>>>>
>>>>
>>>> Does this still allow other IRQs to wake it from xen_poll_irq()?
>>>>
>>>> In the case where process-context code is spinning for a lock without
>>>> disabling interrupts, we *should* allow interrupts to occur still...
>>>> does this?
>>>
>>> Yes. Look at it like idle HLT or WFI. You have to disable interrupt before
>>> checking the condition and then the hardware or in this case the hypervisor
>>> has to bring you back when an interrupt is raised.
>>>
>>> If that would not work then the check would be racy, because the interrupt
>>> could hit and be handled after the check and before going into
>>> HLT/WFI/hypercall and then the thing is out until the next interrupt comes
>>> along, which might be never.
>>
>> Right, but in this case we're calling into the hypervisor to poll for
>> one *specific* IRQ. Everything you say is true for that specific IRQ.
>>
>> My question is what happens to *other* IRQs. We want them, but are they
>> masked? I'm staring at the Xen do_poll() code and haven't quite worked
>> that out...
> 
> Ah, sorry. That of course has to come back like HLT/WFI for any interrupt,
> but I have no idea what the Xen HV is doing there.

The Xen HV is doing it right. It is blocking the vcpu in do_poll() and
any interrupt will unblock it.


Juergen


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 13:38           ` Juergen Gross
  2018-10-10 13:53             ` David Woodhouse
@ 2018-10-10 13:53             ` David Woodhouse
  1 sibling, 0 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-10 13:53 UTC (permalink / raw)
  To: Juergen Gross
  Cc: Thomas Gleixner, David Woodhouse, linux-kernel, xen-devel, x86,
	boris.ostrovsky, hpa, mingo, bp, stable, waiman.long, peterz


> The Xen HV is doing it right. It is blocking the vcpu in do_poll() and
> any interrupt will unblock it.

Great. Thanks for the confirmation.

-- 
dwmw2


^ permalink raw reply	[flat|nested] 35+ messages in thread

* Re: [PATCH 2/2] xen: make xen_qlock_wait() nestable
  2018-10-10 13:38           ` Juergen Gross
@ 2018-10-10 13:53             ` David Woodhouse
  2018-10-10 13:53             ` David Woodhouse
  1 sibling, 0 replies; 35+ messages in thread
From: David Woodhouse @ 2018-10-10 13:53 UTC (permalink / raw)
  To: Juergen Gross
  Cc: waiman.long, peterz, x86, linux-kernel, stable, mingo, bp, hpa,
	xen-devel, boris.ostrovsky, David Woodhouse, Thomas Gleixner


> The Xen HV is doing it right. It is blocking the vcpu in do_poll() and
> any interrupt will unblock it.

Great. Thanks for the confirmation.

-- 
dwmw2


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

* [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling
@ 2018-10-01  7:16 Juergen Gross
  0 siblings, 0 replies; 35+ messages in thread
From: Juergen Gross @ 2018-10-01  7:16 UTC (permalink / raw)
  To: linux-kernel, xen-devel, x86
  Cc: Juergen Gross, Waiman.Long, peterz, mingo, bp, hpa,
	boris.ostrovsky, tglx

The Xen specific queue spinlock wait function has two issues which
could result in a hanging system.

They have a similar root cause of clearing a pending wakeup of a
waiting vcpu and later going to sleep waiting for the just cleared
wakeup event, which of course won't ever happen.

Juergen Gross (2):
  xen: fix race in xen_qlock_wait()
  xen: make xen_qlock_wait() nestable

 arch/x86/xen/spinlock.c | 33 ++++++++++++---------------------
 1 file changed, 12 insertions(+), 21 deletions(-)

Cc: Waiman.Long@hp.com
Cc: peterz@infradead.org

-- 
2.16.4


_______________________________________________
Xen-devel mailing list
Xen-devel@lists.xenproject.org
https://lists.xenproject.org/mailman/listinfo/xen-devel

^ permalink raw reply	[flat|nested] 35+ messages in thread

end of thread, other threads:[~2018-10-10 13:53 UTC | newest]

Thread overview: 35+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2018-10-01  7:16 [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
2018-10-01  7:16 ` [PATCH 1/2] xen: fix race in xen_qlock_wait() Juergen Gross
2018-10-01  7:37   ` Juergen Gross
2018-10-01  7:37   ` Juergen Gross
2018-10-01  8:54   ` Jan Beulich
2018-10-01  8:54   ` [Xen-devel] " Jan Beulich
2018-10-01  7:16 ` Juergen Gross
2018-10-01  7:16 ` [PATCH 2/2] xen: make xen_qlock_wait() nestable Juergen Gross
2018-10-01  7:16 ` Juergen Gross
2018-10-01  7:38   ` Juergen Gross
2018-10-01  7:38   ` Juergen Gross
2018-10-01  8:57   ` Jan Beulich
2018-10-01  8:57   ` [Xen-devel] " Jan Beulich
     [not found]   ` <5BB1E18802000078001ED127@suse.com>
2018-10-01  9:03     ` Juergen Gross
2018-10-01  9:18       ` Jan Beulich
2018-10-01  9:18       ` Jan Beulich
2018-10-10 11:53   ` David Woodhouse
2018-10-10 12:30     ` Thomas Gleixner
2018-10-10 12:30     ` Thomas Gleixner
2018-10-10 12:44       ` David Woodhouse
2018-10-10 12:47         ` Thomas Gleixner
2018-10-10 12:47         ` Thomas Gleixner
2018-10-10 13:38           ` Juergen Gross
2018-10-10 13:53             ` David Woodhouse
2018-10-10 13:53             ` David Woodhouse
2018-10-10 13:38           ` Juergen Gross
2018-10-10 12:44       ` David Woodhouse
2018-10-10 11:53   ` David Woodhouse
2018-10-01  7:37 ` [PATCH 0/2] xen: fix two issues in Xen pv qspinlock handling Juergen Gross
2018-10-01  7:37 ` Juergen Gross
2018-10-09 14:40 ` David Woodhouse
2018-10-09 14:52   ` Juergen Gross
2018-10-09 14:52   ` Juergen Gross
2018-10-09 14:40 ` David Woodhouse
  -- strict thread matches above, loose matches on Subject: below --
2018-10-01  7:16 Juergen Gross

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.