linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [patch 2/8] genirq/generic_pending: Do not lose pending affinity update
@ 2018-06-04 15:33 Thomas Gleixner
  2018-06-05  7:07 ` Song Liu
                   ` (2 more replies)
  0 siblings, 3 replies; 4+ messages in thread
From: Thomas Gleixner @ 2018-06-04 15:33 UTC (permalink / raw)
  To: LKML
  Cc: Ingo Molnar, Peter Zijlstra, Borislav Petkov, Dmitry Safonov,
	Tariq Toukan, Song Liu, Joerg Roedel, Mike Travis, stable

[-- Attachment #1: genirq-generic_pending--Do-not-lose-pending-affinity-update.patch --]
[-- Type: text/plain, Size: 2401 bytes --]

The generic pending interrupt mechanism moves interrupts from the interrupt
handler on the original target CPU to the new destination CPU. This is
required for x86 and ia64 due to the way the interrupt delivery and
acknowledge works if the interrupts are not remapped.

However that update can fail for various reasons. Some of them are valid
reasons to discard the pending update, but the case, when the previous move
has not been fully cleaned up is not a legit reason to fail.

Check the return value of irq_do_set_affinity() for -EBUSY, which indicates
a pending cleanup, and rearm the pending move in the irq dexcriptor so it's
tried again when the next interrupt arrives.

Fixes: 996c591227d9 ("x86/irq: Plug vector cleanup race")
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Cc: stable@vger.kernel.org
---
 kernel/irq/migration.c |   24 ++++++++++++++++++------
 1 file changed, 18 insertions(+), 6 deletions(-)

--- a/kernel/irq/migration.c
+++ b/kernel/irq/migration.c
@@ -38,17 +38,18 @@ bool irq_fixup_move_pending(struct irq_d
 void irq_move_masked_irq(struct irq_data *idata)
 {
 	struct irq_desc *desc = irq_data_to_desc(idata);
-	struct irq_chip *chip = desc->irq_data.chip;
+	struct irq_data *data = &desc->irq_data;
+	struct irq_chip *chip = data->chip;
 
-	if (likely(!irqd_is_setaffinity_pending(&desc->irq_data)))
+	if (likely(!irqd_is_setaffinity_pending(data)))
 		return;
 
-	irqd_clr_move_pending(&desc->irq_data);
+	irqd_clr_move_pending(data);
 
 	/*
 	 * Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
 	 */
-	if (irqd_is_per_cpu(&desc->irq_data)) {
+	if (irqd_is_per_cpu(data)) {
 		WARN_ON(1);
 		return;
 	}
@@ -73,9 +74,20 @@ void irq_move_masked_irq(struct irq_data
 	 * For correct operation this depends on the caller
 	 * masking the irqs.
 	 */
-	if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids)
-		irq_do_set_affinity(&desc->irq_data, desc->pending_mask, false);
+	if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
+		int ret;
 
+		ret = irq_do_set_affinity(data, desc->pending_mask, false);
+		/*
+		 * If the there is a cleanup pending in the underlying
+		 * vector management, reschedule the move for the next
+		 * interrupt. Leave desc->pending_mask intact.
+		 */
+		if (ret == -EBUSY) {
+			irqd_set_move_pending(data);
+			return;
+		}
+	}
 	cpumask_clear(desc->pending_mask);
 }
 

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [patch 2/8] genirq/generic_pending: Do not lose pending affinity update
  2018-06-04 15:33 [patch 2/8] genirq/generic_pending: Do not lose pending affinity update Thomas Gleixner
@ 2018-06-05  7:07 ` Song Liu
  2018-06-05 11:20 ` Dou Liyang
  2018-06-06 13:31 ` [tip:x86/urgent] " tip-bot for Thomas Gleixner
  2 siblings, 0 replies; 4+ messages in thread
From: Song Liu @ 2018-06-05  7:07 UTC (permalink / raw)
  To: Thomas Gleixner
  Cc: LKML, Ingo Molnar, Peter Zijlstra, Borislav Petkov,
	Dmitry Safonov, Tariq Toukan, Joerg Roedel, Mike Travis, stable

On Mon, Jun 4, 2018 at 8:33 AM, Thomas Gleixner <tglx@linutronix.de> wrote:
> The generic pending interrupt mechanism moves interrupts from the interrupt
> handler on the original target CPU to the new destination CPU. This is
> required for x86 and ia64 due to the way the interrupt delivery and
> acknowledge works if the interrupts are not remapped.
>
> However that update can fail for various reasons. Some of them are valid
> reasons to discard the pending update, but the case, when the previous move
> has not been fully cleaned up is not a legit reason to fail.
>
> Check the return value of irq_do_set_affinity() for -EBUSY, which indicates
> a pending cleanup, and rearm the pending move in the irq dexcriptor so it's
> tried again when the next interrupt arrives.
>
> Fixes: 996c591227d9 ("x86/irq: Plug vector cleanup race")
> Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
> Cc: stable@vger.kernel.org

Tested-by: Song Liu <songliubraving@fb.com>

> ---
>  kernel/irq/migration.c |   24 ++++++++++++++++++------
>  1 file changed, 18 insertions(+), 6 deletions(-)
>
> --- a/kernel/irq/migration.c
> +++ b/kernel/irq/migration.c
> @@ -38,17 +38,18 @@ bool irq_fixup_move_pending(struct irq_d
>  void irq_move_masked_irq(struct irq_data *idata)
>  {
>         struct irq_desc *desc = irq_data_to_desc(idata);
> -       struct irq_chip *chip = desc->irq_data.chip;
> +       struct irq_data *data = &desc->irq_data;
> +       struct irq_chip *chip = data->chip;
>
> -       if (likely(!irqd_is_setaffinity_pending(&desc->irq_data)))
> +       if (likely(!irqd_is_setaffinity_pending(data)))
>                 return;
>
> -       irqd_clr_move_pending(&desc->irq_data);
> +       irqd_clr_move_pending(data);
>
>         /*
>          * Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
>          */
> -       if (irqd_is_per_cpu(&desc->irq_data)) {
> +       if (irqd_is_per_cpu(data)) {
>                 WARN_ON(1);
>                 return;
>         }
> @@ -73,9 +74,20 @@ void irq_move_masked_irq(struct irq_data
>          * For correct operation this depends on the caller
>          * masking the irqs.
>          */
> -       if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids)
> -               irq_do_set_affinity(&desc->irq_data, desc->pending_mask, false);
> +       if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
> +               int ret;
>
> +               ret = irq_do_set_affinity(data, desc->pending_mask, false);
> +               /*
> +                * If the there is a cleanup pending in the underlying
> +                * vector management, reschedule the move for the next
> +                * interrupt. Leave desc->pending_mask intact.
> +                */
> +               if (ret == -EBUSY) {
> +                       irqd_set_move_pending(data);
> +                       return;
> +               }
> +       }
>         cpumask_clear(desc->pending_mask);
>  }
>
>
>

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [patch 2/8] genirq/generic_pending: Do not lose pending affinity update
  2018-06-04 15:33 [patch 2/8] genirq/generic_pending: Do not lose pending affinity update Thomas Gleixner
  2018-06-05  7:07 ` Song Liu
@ 2018-06-05 11:20 ` Dou Liyang
  2018-06-06 13:31 ` [tip:x86/urgent] " tip-bot for Thomas Gleixner
  2 siblings, 0 replies; 4+ messages in thread
From: Dou Liyang @ 2018-06-05 11:20 UTC (permalink / raw)
  To: Thomas Gleixner, LKML
  Cc: Ingo Molnar, Peter Zijlstra, Borislav Petkov, Dmitry Safonov,
	Tariq Toukan, Song Liu, Joerg Roedel, Mike Travis, stable

Hi Thomas,

At 06/04/2018 11:33 PM, Thomas Gleixner wrote:
> The generic pending interrupt mechanism moves interrupts from the interrupt
> handler on the original target CPU to the new destination CPU. This is
> required for x86 and ia64 due to the way the interrupt delivery and
> acknowledge works if the interrupts are not remapped.
> 
> However that update can fail for various reasons. Some of them are valid
> reasons to discard the pending update, but the case, when the previous move
> has not been fully cleaned up is not a legit reason to fail.
> 
> Check the return value of irq_do_set_affinity() for -EBUSY, which indicates
> a pending cleanup, and rearm the pending move in the irq dexcriptor so it's

s/dexcriptor/descriptor

> tried again when the next interrupt arrives.
> 
> Fixes: 996c591227d9 ("x86/irq: Plug vector cleanup race")
> Signed-off-by: Thomas Gleixner<tglx@linutronix.de>
> Cc:stable@vger.kernel.org
> ---
>   kernel/irq/migration.c |   24 ++++++++++++++++++------
>   1 file changed, 18 insertions(+), 6 deletions(-)
> 
> --- a/kernel/irq/migration.c
> +++ b/kernel/irq/migration.c
> @@ -38,17 +38,18 @@ bool irq_fixup_move_pending(struct irq_d
>   void irq_move_masked_irq(struct irq_data *idata)
>   {
>   	struct irq_desc *desc = irq_data_to_desc(idata);
> -	struct irq_chip *chip = desc->irq_data.chip;
> +	struct irq_data *data = &desc->irq_data;
> +	struct irq_chip *chip = data->chip;
>   
> -	if (likely(!irqd_is_setaffinity_pending(&desc->irq_data)))
> +	if (likely(!irqd_is_setaffinity_pending(data)))
>   		return;
>   
> -	irqd_clr_move_pending(&desc->irq_data);
> +	irqd_clr_move_pending(data);
>   
>   	/*
>   	 * Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
>   	 */
> -	if (irqd_is_per_cpu(&desc->irq_data)) {
> +	if (irqd_is_per_cpu(data)) {
>   		WARN_ON(1);
>   		return;
>   	}
> @@ -73,9 +74,20 @@ void irq_move_masked_irq(struct irq_data
>   	 * For correct operation this depends on the caller
>   	 * masking the irqs.
>   	 */
> -	if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids)
> -		irq_do_set_affinity(&desc->irq_data, desc->pending_mask, false);
> +	if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
> +		int ret;
>   
> +		ret = irq_do_set_affinity(data, desc->pending_mask, false);
> +		/*
> +		 * If the there is a cleanup pending in the underlying

s/If the there is/If there is/

Thanks,
	dou

^ permalink raw reply	[flat|nested] 4+ messages in thread

* [tip:x86/urgent] genirq/generic_pending: Do not lose pending affinity update
  2018-06-04 15:33 [patch 2/8] genirq/generic_pending: Do not lose pending affinity update Thomas Gleixner
  2018-06-05  7:07 ` Song Liu
  2018-06-05 11:20 ` Dou Liyang
@ 2018-06-06 13:31 ` tip-bot for Thomas Gleixner
  2 siblings, 0 replies; 4+ messages in thread
From: tip-bot for Thomas Gleixner @ 2018-06-06 13:31 UTC (permalink / raw)
  To: linux-tip-commits
  Cc: tariqt, bp, mingo, liu.song.a23, jroedel, tglx, linux-kernel,
	songliubraving, 0x7f454c46, mike.travis, hpa, peterz

Commit-ID:  a33a5d2d16cb84bea8d5f5510f3a41aa48b5c467
Gitweb:     https://git.kernel.org/tip/a33a5d2d16cb84bea8d5f5510f3a41aa48b5c467
Author:     Thomas Gleixner <tglx@linutronix.de>
AuthorDate: Mon, 4 Jun 2018 17:33:54 +0200
Committer:  Thomas Gleixner <tglx@linutronix.de>
CommitDate: Wed, 6 Jun 2018 15:18:19 +0200

genirq/generic_pending: Do not lose pending affinity update

The generic pending interrupt mechanism moves interrupts from the interrupt
handler on the original target CPU to the new destination CPU. This is
required for x86 and ia64 due to the way the interrupt delivery and
acknowledge works if the interrupts are not remapped.

However that update can fail for various reasons. Some of them are valid
reasons to discard the pending update, but the case, when the previous move
has not been fully cleaned up is not a legit reason to fail.

Check the return value of irq_do_set_affinity() for -EBUSY, which indicates
a pending cleanup, and rearm the pending move in the irq dexcriptor so it's
tried again when the next interrupt arrives.

Fixes: 996c591227d9 ("x86/irq: Plug vector cleanup race")
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Tested-by: Song Liu <songliubraving@fb.com>
Cc: Joerg Roedel <jroedel@suse.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Song Liu <liu.song.a23@gmail.com>
Cc: Dmitry Safonov <0x7f454c46@gmail.com>
Cc: stable@vger.kernel.org
Cc: Mike Travis <mike.travis@hpe.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Tariq Toukan <tariqt@mellanox.com>
Link: https://lkml.kernel.org/r/20180604162224.386544292@linutronix.de

---
 kernel/irq/migration.c | 26 +++++++++++++++++++-------
 1 file changed, 19 insertions(+), 7 deletions(-)

diff --git a/kernel/irq/migration.c b/kernel/irq/migration.c
index 86ae0eb80b53..8b8cecd18cce 100644
--- a/kernel/irq/migration.c
+++ b/kernel/irq/migration.c
@@ -38,17 +38,18 @@ bool irq_fixup_move_pending(struct irq_desc *desc, bool force_clear)
 void irq_move_masked_irq(struct irq_data *idata)
 {
 	struct irq_desc *desc = irq_data_to_desc(idata);
-	struct irq_chip *chip = desc->irq_data.chip;
+	struct irq_data *data = &desc->irq_data;
+	struct irq_chip *chip = data->chip;
 
-	if (likely(!irqd_is_setaffinity_pending(&desc->irq_data)))
+	if (likely(!irqd_is_setaffinity_pending(data)))
 		return;
 
-	irqd_clr_move_pending(&desc->irq_data);
+	irqd_clr_move_pending(data);
 
 	/*
 	 * Paranoia: cpu-local interrupts shouldn't be calling in here anyway.
 	 */
-	if (irqd_is_per_cpu(&desc->irq_data)) {
+	if (irqd_is_per_cpu(data)) {
 		WARN_ON(1);
 		return;
 	}
@@ -73,9 +74,20 @@ void irq_move_masked_irq(struct irq_data *idata)
 	 * For correct operation this depends on the caller
 	 * masking the irqs.
 	 */
-	if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids)
-		irq_do_set_affinity(&desc->irq_data, desc->pending_mask, false);
-
+	if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) {
+		int ret;
+
+		ret = irq_do_set_affinity(data, desc->pending_mask, false);
+		/*
+		 * If the there is a cleanup pending in the underlying
+		 * vector management, reschedule the move for the next
+		 * interrupt. Leave desc->pending_mask intact.
+		 */
+		if (ret == -EBUSY) {
+			irqd_set_move_pending(data);
+			return;
+		}
+	}
 	cpumask_clear(desc->pending_mask);
 }
 

^ permalink raw reply related	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2018-06-06 13:33 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2018-06-04 15:33 [patch 2/8] genirq/generic_pending: Do not lose pending affinity update Thomas Gleixner
2018-06-05  7:07 ` Song Liu
2018-06-05 11:20 ` Dou Liyang
2018-06-06 13:31 ` [tip:x86/urgent] " tip-bot for Thomas Gleixner

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).