All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH] x86, mm: Send tlb flush IPIs to online cpus only
@ 2012-07-19 12:57 Srivatsa S. Bhat
  2012-07-19 13:32 ` Borislav Petkov
  0 siblings, 1 reply; 4+ messages in thread
From: Srivatsa S. Bhat @ 2012-07-19 12:57 UTC (permalink / raw)
  To: msb, mingo, linux-kernel, shaohua.li, yinghai
  Cc: tglx, hpa, x86, tj, akpm, sfr, cl, olofj, paulmck, Srivatsa S. Bhat

Mandeep reports:
We are seeing a softlockup reporting during shutdown. The stack
trace shows us that we are inside default_send_IPI_mask_logical:

 BUG: soft lockup - CPU#0 stuck for 11s! [lmt-udev:23605]
 Pid: 23605, comm: lmt-udev Tainted: G        WC   3.2.7 #1
 EIP: 0060:[<8101eec6>] EFLAGS: 00000202 CPU: 0
 EIP is at flush_tlb_others_ipi+0x8a/0xba
 Call Trace:
  [<8101f0bb>] flush_tlb_mm+0x5e/0x62
  [<8101e36c>] pud_populate+0x2c/0x31
  [<8101e409>] pgd_alloc+0x98/0xc7
  [<8102c881>] mm_init.isra.38+0xcc/0xf3
  [<8102cbc2>] dup_mm+0x68/0x34e
  [<8139bbae>] ? _cond_resched+0xd/0x21
  [<810a5b7c>] ? kmem_cache_alloc+0x26/0xe2
  [<8102d421>] ? copy_process+0x556/0xda6
  [<8102d641>] copy_process+0x776/0xda6
  [<8102dd5e>] do_fork+0xcb/0x1d4
  [<810a8c96>] ? do_sync_write+0xd3/0xd3
  [<810a94ab>] ? vfs_read+0x95/0xa2
  [<81008850>] sys_clone+0x20/0x25
  [<8139d8c5>] ptregs_clone+0x15/0x30
  [<8139d7f7>] ? sysenter_do_call+0x12/0x26

Before the softlock, we see the following kernel warning:

 WARNING: at ../../arch/x86/kernel/apic/ipi.c:113 default_send_IPI_mask_logical+0x58/0x73()
 Pid: 23605, comm: lmt-udev Tainted: G         C   3.2.7 #1
 Call Trace:
  [<8102e666>] warn_slowpath_common+0x68/0x7d
  [<81016c36>] ? default_send_IPI_mask_logical+0x58/0x73
  [<8102e68f>] warn_slowpath_null+0x14/0x18
  [<81016c36>] default_send_IPI_mask_logical+0x58/0x73
  [<8101eec2>] flush_tlb_others_ipi+0x86/0xba
  [<8101f0bb>] flush_tlb_mm+0x5e/0x62
  [<8101e36c>] pud_populate+0x2c/0x31
  [<8101e409>] pgd_alloc+0x98/0xc7
  [<8102c881>] mm_init.isra.38+0xcc/0xf3
  [<8102cbc2>] dup_mm+0x68/0x34e
  [<8139bbae>] ? _cond_resched+0xd/0x21
  [<810a5b7c>] ? kmem_cache_alloc+0x26/0xe2
  [<8102d421>] ? copy_process+0x556/0xda6
  [<8102d641>] copy_process+0x776/0xda6
  [<8102dd5e>] do_fork+0xcb/0x1d4
  [<810a8c96>] ? do_sync_write+0xd3/0xd3
  [<810a94ab>] ? vfs_read+0x95/0xa2
  [<81008850>] sys_clone+0x20/0x25
  [<8139d8c5>] ptregs_clone+0x15/0x30
  [<8139d7f7>] ? sysenter_do_call+0x12/0x26

So we are sending an IPI to a cpu which is now offline. Once a cpu is offline,
it will no longer respond to IPIs. This explains the softlockup.

A cpu in the mm_cpumask could go offline before we send the invalidate
IPI causing us to wait forever. Avoid this by sending the IPI to only the
online cpus.

[Since flush_tlb_others_ipi() is always called with preempt disabled, it is
not possible for a CPU to go offline once we enter this function, because
CPU offline goes through the stop_machine() stuff (which cannot proceed until
all preempt disabled sections are exited). So we don't have to worry about
any race between CPU offline and the target cpumask calculation in
flush_tlb_others_ipi().]

Addresses http://crosbug.com/31737

Reported-and-debugged-by: Mandeep Singh Baines <msb@chromium.org>
Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com>
Acked-by: Mandeep Singh Baines <msb@chromium.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: x86@kernel.org
Cc: Tejun Heo <tj@kernel.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Stephen Rothwell <sfr@canb.auug.org.au>
Cc: Christoph Lameter <cl@gentwo.org>
Cc: Olof Johansson <olofj@chromium.org>
---

 arch/x86/mm/tlb.c |    6 +++++-
 1 files changed, 5 insertions(+), 1 deletions(-)

diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c
index 5e57e11..9d387a9 100644
--- a/arch/x86/mm/tlb.c
+++ b/arch/x86/mm/tlb.c
@@ -186,7 +186,11 @@ static void flush_tlb_others_ipi(const struct cpumask *cpumask,
 
 	f->flush_mm = mm;
 	f->flush_va = va;
-	if (cpumask_andnot(to_cpumask(f->flush_cpumask), cpumask, cpumask_of(smp_processor_id()))) {
+
+	cpumask_and(to_cpumask(f->flush_cpumask), cpumask, cpu_online_mask);
+	cpumask_clear_cpu(smp_processor_id(), to_cpumask(f->flush_cpumask));
+
+	if (!cpumask_empty(to_cpumask(f->flush_cpumask))) {
 		/*
 		 * We have to send the IPI only to
 		 * CPUs affected.


^ permalink raw reply related	[flat|nested] 4+ messages in thread

* Re: [PATCH] x86, mm: Send tlb flush IPIs to online cpus only
  2012-07-19 12:57 [PATCH] x86, mm: Send tlb flush IPIs to online cpus only Srivatsa S. Bhat
@ 2012-07-19 13:32 ` Borislav Petkov
  2012-07-19 13:35   ` Srivatsa S. Bhat
  0 siblings, 1 reply; 4+ messages in thread
From: Borislav Petkov @ 2012-07-19 13:32 UTC (permalink / raw)
  To: Srivatsa S. Bhat
  Cc: msb, mingo, linux-kernel, shaohua.li, yinghai, tglx, hpa, x86,
	tj, akpm, sfr, cl, olofj, paulmck, Alex Shi, Conny Seidel

On Thu, Jul 19, 2012 at 06:27:33PM +0530, Srivatsa S. Bhat wrote:

[ … ]
 
> So we are sending an IPI to a cpu which is now offline. Once a cpu is offline,
> it will no longer respond to IPIs. This explains the softlockup.
> 
> A cpu in the mm_cpumask could go offline before we send the invalidate
> IPI causing us to wait forever. Avoid this by sending the IPI to only the
> online cpus.
> 
> [Since flush_tlb_others_ipi() is always called with preempt disabled, it is
> not possible for a CPU to go offline once we enter this function, because
> CPU offline goes through the stop_machine() stuff (which cannot proceed until
> all preempt disabled sections are exited). So we don't have to worry about
> any race between CPU offline and the target cpumask calculation in
> flush_tlb_others_ipi().]
> 
> Addresses http://crosbug.com/31737
> 
> Reported-and-debugged-by: Mandeep Singh Baines <msb@chromium.org>
> Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com>
> Acked-by: Mandeep Singh Baines <msb@chromium.org>
> Cc: Thomas Gleixner <tglx@linutronix.de>
> Cc: Ingo Molnar <mingo@redhat.com>
> Cc: "H. Peter Anvin" <hpa@zytor.com>
> Cc: x86@kernel.org
> Cc: Tejun Heo <tj@kernel.org>
> Cc: Andrew Morton <akpm@linux-foundation.org>
> Cc: Stephen Rothwell <sfr@canb.auug.org.au>
> Cc: Christoph Lameter <cl@gentwo.org>
> Cc: Olof Johansson <olofj@chromium.org>
> ---
> 
>  arch/x86/mm/tlb.c |    6 +++++-
>  1 files changed, 5 insertions(+), 1 deletions(-)
> 
> diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c
> index 5e57e11..9d387a9 100644
> --- a/arch/x86/mm/tlb.c
> +++ b/arch/x86/mm/tlb.c
> @@ -186,7 +186,11 @@ static void flush_tlb_others_ipi(const struct cpumask *cpumask,
>  
>  	f->flush_mm = mm;
>  	f->flush_va = va;
> -	if (cpumask_andnot(to_cpumask(f->flush_cpumask), cpumask, cpumask_of(smp_processor_id()))) {
> +
> +	cpumask_and(to_cpumask(f->flush_cpumask), cpumask, cpu_online_mask);
> +	cpumask_clear_cpu(smp_processor_id(), to_cpumask(f->flush_cpumask));
> +
> +	if (!cpumask_empty(to_cpumask(f->flush_cpumask))) {

FWIW, there's code in tip/x86/mm which reworks all that and
flush_tlb_others_ipi along with the 32 TLB flush vectors are being
removed in favor of a smp_call_function_many thing. And it should be
hotplug-safe since it must be called with preemption disabled anyway.

-- 
Regards/Gruss,
Boris.

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] x86, mm: Send tlb flush IPIs to online cpus only
  2012-07-19 13:32 ` Borislav Petkov
@ 2012-07-19 13:35   ` Srivatsa S. Bhat
  2012-07-19 13:39     ` Borislav Petkov
  0 siblings, 1 reply; 4+ messages in thread
From: Srivatsa S. Bhat @ 2012-07-19 13:35 UTC (permalink / raw)
  To: Borislav Petkov, msb, mingo, linux-kernel, shaohua.li, yinghai,
	tglx, hpa, x86, tj, akpm, sfr, cl, olofj, paulmck, Alex Shi,
	Conny Seidel

On 07/19/2012 07:02 PM, Borislav Petkov wrote:
> On Thu, Jul 19, 2012 at 06:27:33PM +0530, Srivatsa S. Bhat wrote:
> 
> [ … ]
> 
>> So we are sending an IPI to a cpu which is now offline. Once a cpu is offline,
>> it will no longer respond to IPIs. This explains the softlockup.
>>
>> A cpu in the mm_cpumask could go offline before we send the invalidate
>> IPI causing us to wait forever. Avoid this by sending the IPI to only the
>> online cpus.
>>
>> [Since flush_tlb_others_ipi() is always called with preempt disabled, it is
>> not possible for a CPU to go offline once we enter this function, because
>> CPU offline goes through the stop_machine() stuff (which cannot proceed until
>> all preempt disabled sections are exited). So we don't have to worry about
>> any race between CPU offline and the target cpumask calculation in
>> flush_tlb_others_ipi().]
>>
>> Addresses http://crosbug.com/31737
>>
>> Reported-and-debugged-by: Mandeep Singh Baines <msb@chromium.org>
>> Signed-off-by: Srivatsa S. Bhat <srivatsa.bhat@linux.vnet.ibm.com>
>> Acked-by: Mandeep Singh Baines <msb@chromium.org>
>> Cc: Thomas Gleixner <tglx@linutronix.de>
>> Cc: Ingo Molnar <mingo@redhat.com>
>> Cc: "H. Peter Anvin" <hpa@zytor.com>
>> Cc: x86@kernel.org
>> Cc: Tejun Heo <tj@kernel.org>
>> Cc: Andrew Morton <akpm@linux-foundation.org>
>> Cc: Stephen Rothwell <sfr@canb.auug.org.au>
>> Cc: Christoph Lameter <cl@gentwo.org>
>> Cc: Olof Johansson <olofj@chromium.org>
>> ---
>>
>>  arch/x86/mm/tlb.c |    6 +++++-
>>  1 files changed, 5 insertions(+), 1 deletions(-)
>>
>> diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c
>> index 5e57e11..9d387a9 100644
>> --- a/arch/x86/mm/tlb.c
>> +++ b/arch/x86/mm/tlb.c
>> @@ -186,7 +186,11 @@ static void flush_tlb_others_ipi(const struct cpumask *cpumask,
>>  
>>  	f->flush_mm = mm;
>>  	f->flush_va = va;
>> -	if (cpumask_andnot(to_cpumask(f->flush_cpumask), cpumask, cpumask_of(smp_processor_id()))) {
>> +
>> +	cpumask_and(to_cpumask(f->flush_cpumask), cpumask, cpu_online_mask);
>> +	cpumask_clear_cpu(smp_processor_id(), to_cpumask(f->flush_cpumask));
>> +
>> +	if (!cpumask_empty(to_cpumask(f->flush_cpumask))) {
> 
> FWIW, there's code in tip/x86/mm which reworks all that and
> flush_tlb_others_ipi along with the 32 TLB flush vectors are being
> removed in favor of a smp_call_function_many thing. And it should be
> hotplug-safe since it must be called with preemption disabled anyway.
> 

Oh.. ok.. Then this patch is perhaps unnecessary..
Thanks for the info!

Regards,
Srivatsa S. Bhat


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH] x86, mm: Send tlb flush IPIs to online cpus only
  2012-07-19 13:35   ` Srivatsa S. Bhat
@ 2012-07-19 13:39     ` Borislav Petkov
  0 siblings, 0 replies; 4+ messages in thread
From: Borislav Petkov @ 2012-07-19 13:39 UTC (permalink / raw)
  To: Srivatsa S. Bhat
  Cc: msb, mingo, linux-kernel, shaohua.li, yinghai, tglx, hpa, x86,
	tj, akpm, sfr, cl, olofj, paulmck, Alex Shi, Conny Seidel

On Thu, Jul 19, 2012 at 07:05:51PM +0530, Srivatsa S. Bhat wrote:
> Oh.. ok.. Then this patch is perhaps unnecessary.. Thanks for the
> info!

Probably, if the TLB flushrange set from Alex goes upstream in the next
merge window...

-- 
Regards/Gruss,
Boris.

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2012-07-19 13:39 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2012-07-19 12:57 [PATCH] x86, mm: Send tlb flush IPIs to online cpus only Srivatsa S. Bhat
2012-07-19 13:32 ` Borislav Petkov
2012-07-19 13:35   ` Srivatsa S. Bhat
2012-07-19 13:39     ` Borislav Petkov

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.