* lockdep splat in 3.0-rc7 (circular locking, scheduler related?)
@ 2011-07-12 20:51 Ben Greear
0 siblings, 0 replies; only message in thread
From: Ben Greear @ 2011-07-12 20:51 UTC (permalink / raw)
To: Linux Kernel Mailing List
Been seeing this one for a while in 3.0, and still in -rc7.
Kernel has lots of debug enabled, and some nfs patches.
=======================================================
[ INFO: possible circular locking dependency detected ]
3.0.0-rc7+ #23
-------------------------------------------------------
ip/30705 is trying to acquire lock:
(rcu_node_level_0){..-...}, at: [<ffffffff810a7455>] rcu_report_unblock_qs_rnp+0x52/0x72
but task is already holding lock:
(&rq->lock){-.-.-.}, at: [<ffffffff81045d88>] sched_ttwu_pending+0x34/0x58
which lock already depends on the new lock.
the existing dependency chain (in reverse order) is:
-> #3 (&rq->lock){-.-.-.}:
[<ffffffff8107b4e5>] lock_acquire+0xf4/0x14b
[<ffffffff8147e6d1>] _raw_spin_lock+0x36/0x45
[<ffffffff8103d9fe>] __task_rq_lock+0x5b/0x89
[<ffffffff81046a99>] wake_up_new_task+0x41/0x116
[<ffffffff810496b9>] do_fork+0x207/0x2f1
[<ffffffff81010d25>] kernel_thread+0x70/0x72
[<ffffffff81465c0d>] rest_init+0x21/0xd7
[<ffffffff81aa9c76>] start_kernel+0x3bd/0x3c8
[<ffffffff81aa92cd>] x86_64_start_reservations+0xb8/0xbc
[<ffffffff81aa93d2>] x86_64_start_kernel+0x101/0x110
-> #2 (&p->pi_lock){-.-.-.}:
[<ffffffff8107b4e5>] lock_acquire+0xf4/0x14b
[<ffffffff8147e7e5>] _raw_spin_lock_irqsave+0x4e/0x60
[<ffffffff810468b3>] try_to_wake_up+0x29/0x1a0
[<ffffffff81046a37>] default_wake_function+0xd/0xf
[<ffffffff810674fe>] autoremove_wake_function+0x13/0x38
[<ffffffff810395d0>] __wake_up_common+0x49/0x7f
[<ffffffff8103c78a>] __wake_up+0x34/0x48
[<ffffffff810a74c5>] rcu_report_exp_rnp+0x50/0x89
[<ffffffff810a804e>] __rcu_read_unlock+0x1e9/0x24e
[<ffffffff81110c01>] rcu_read_unlock+0x21/0x23
[<ffffffff81113320>] __mem_cgroup_try_charge+0x195/0x4f5
[<ffffffff81114c8c>] mem_cgroup_charge_common+0xa7/0xdb
[<ffffffff81114e22>] mem_cgroup_newpage_charge+0x34/0x43
[<ffffffff810ea0cb>] handle_pte_fault+0x1b6/0x84d
[<ffffffff810ea90e>] handle_mm_fault+0x1ac/0x1c4
[<ffffffff81481e34>] do_page_fault+0x32d/0x374
[<ffffffff8147f605>] page_fault+0x25/0x30
-> #1 (sync_rcu_preempt_exp_wq.lock){......}:
[<ffffffff8107b4e5>] lock_acquire+0xf4/0x14b
[<ffffffff8147e7e5>] _raw_spin_lock_irqsave+0x4e/0x60
[<ffffffff8103c773>] __wake_up+0x1d/0x48
[<ffffffff810a74c5>] rcu_report_exp_rnp+0x50/0x89
[<ffffffff810a8c48>] sync_rcu_preempt_exp_init.clone.0+0x3e/0x53
[<ffffffff810a8d38>] synchronize_rcu_expedited+0xdb/0x1c3
[<ffffffff813c0fd3>] synchronize_net+0x25/0x2e
[<ffffffff813c35a2>] rollback_registered_many+0xee/0x1e1
[<ffffffff813c36a9>] unregister_netdevice_many+0x14/0x55
[<ffffffffa036a138>] 0xffffffffa036a138
[<ffffffff813bdb5d>] ops_exit_list+0x25/0x4e
[<ffffffff813bdda9>] unregister_pernet_operations+0x5c/0x8e
[<ffffffff813bde42>] unregister_pernet_subsys+0x22/0x32
[<ffffffffa0372e2c>] 0xffffffffa0372e2c
[<ffffffff81083d56>] sys_delete_module+0x1aa/0x20e
[<ffffffff81484fd2>] system_call_fastpath+0x16/0x1b
-> #0 (rcu_node_level_0){..-...}:
[<ffffffff8107acf2>] __lock_acquire+0xae6/0xdd5
[<ffffffff8107b4e5>] lock_acquire+0xf4/0x14b
[<ffffffff8147e6d1>] _raw_spin_lock+0x36/0x45
[<ffffffff810a7455>] rcu_report_unblock_qs_rnp+0x52/0x72
[<ffffffff810a800c>] __rcu_read_unlock+0x1a7/0x24e
[<ffffffff8103d33d>] rcu_read_unlock+0x21/0x23
[<ffffffff8103d392>] cpuacct_charge+0x53/0x5b
[<ffffffff81044cea>] update_curr+0x11f/0x15a
[<ffffffff81045a1a>] enqueue_task_fair+0x46/0x22a
[<ffffffff8103d2b0>] enqueue_task+0x61/0x68
[<ffffffff8103d2df>] activate_task+0x28/0x30
[<ffffffff81040b26>] ttwu_activate+0x12/0x34
[<ffffffff81045d42>] ttwu_do_activate.clone.4+0x2d/0x3f
[<ffffffff81045d97>] sched_ttwu_pending+0x43/0x58
[<ffffffff81045db5>] scheduler_ipi+0x9/0xb
[<ffffffff81021e40>] smp_reschedule_interrupt+0x25/0x27
[<ffffffff81485ef3>] reschedule_interrupt+0x13/0x20
[<ffffffff810c98dc>] rcu_read_unlock+0x21/0x23
[<ffffffff810c999f>] find_get_pages_tag+0xc1/0xed
[<ffffffff810d3480>] pagevec_lookup_tag+0x20/0x29
[<ffffffff811979e7>] write_cache_pages_da+0xeb/0x37b
[<ffffffff81197f65>] ext4_da_writepages+0x2ee/0x565
[<ffffffff810d2ccf>] do_writepages+0x1f/0x28
[<ffffffff810ca39c>] __filemap_fdatawrite_range+0x4e/0x50
[<ffffffff810ca8a6>] filemap_flush+0x17/0x19
[<ffffffff81193447>] ext4_alloc_da_blocks+0x81/0xab
[<ffffffff8118d669>] ext4_release_file+0x35/0xb3
[<ffffffff8111b07c>] fput+0x117/0x1b2
[<ffffffff81117b56>] filp_close+0x6d/0x78
[<ffffffff8104c2f7>] put_files_struct+0xca/0x190
[<ffffffff8104c403>] exit_files+0x46/0x4e
[<ffffffff8104deb7>] do_exit+0x2b5/0x760
[<ffffffff8104e3e0>] do_group_exit+0x7e/0xa9
[<ffffffff8104e41d>] sys_exit_group+0x12/0x16
[<ffffffff81484fd2>] system_call_fastpath+0x16/0x1b
other info that might help us debug this:
Chain exists of:
rcu_node_level_0 --> &p->pi_lock --> &rq->lock
Possible unsafe locking scenario:
CPU0 CPU1
---- ----
lock(&rq->lock);
lock(&p->pi_lock);
lock(&rq->lock);
lock(rcu_node_level_0);
*** DEADLOCK ***
2 locks held by ip/30705:
#0: (jbd2_handle){+.+...}, at: [<ffffffff811c9e68>] start_this_handle+0x605/0x64d
#1: (&rq->lock){-.-.-.}, at: [<ffffffff81045d88>] sched_ttwu_pending+0x34/0x58
stack backtrace:
Pid: 30705, comm: ip Not tainted 3.0.0-rc7+ #23
Call Trace:
<IRQ> [<ffffffff8147ed33>] ? _raw_spin_unlock_irqrestore+0x6b/0x79
[<ffffffff8107a14d>] print_circular_bug+0x1fe/0x20f
[<ffffffff8107acf2>] __lock_acquire+0xae6/0xdd5
[<ffffffff810a7455>] ? rcu_report_unblock_qs_rnp+0x52/0x72
[<ffffffff8107b4e5>] lock_acquire+0xf4/0x14b
[<ffffffff810a7455>] ? rcu_report_unblock_qs_rnp+0x52/0x72
[<ffffffff8147e6d1>] _raw_spin_lock+0x36/0x45
[<ffffffff810a7455>] ? rcu_report_unblock_qs_rnp+0x52/0x72
[<ffffffff8147ecbb>] ? _raw_spin_unlock+0x45/0x52
[<ffffffff810a7455>] rcu_report_unblock_qs_rnp+0x52/0x72
[<ffffffff810a7f41>] ? __rcu_read_unlock+0xdc/0x24e
[<ffffffff810a800c>] __rcu_read_unlock+0x1a7/0x24e
[<ffffffff8103d33d>] rcu_read_unlock+0x21/0x23
[<ffffffff8103d392>] cpuacct_charge+0x53/0x5b
[<ffffffff81044cea>] update_curr+0x11f/0x15a
[<ffffffff81045a1a>] enqueue_task_fair+0x46/0x22a
[<ffffffff8103d2b0>] enqueue_task+0x61/0x68
[<ffffffff8103d2df>] activate_task+0x28/0x30
[<ffffffff81040b26>] ttwu_activate+0x12/0x34
[<ffffffff81045d42>] ttwu_do_activate.clone.4+0x2d/0x3f
[<ffffffff81045d97>] sched_ttwu_pending+0x43/0x58
[<ffffffff81045db5>] scheduler_ipi+0x9/0xb
[<ffffffff81021e40>] smp_reschedule_interrupt+0x25/0x27
[<ffffffff81485ef3>] reschedule_interrupt+0x13/0x20
<EOI> [<ffffffff8122d6ce>] ? trace_hardirqs_on_thunk+0x3a/0x3f
[<ffffffff810a7eb8>] ? __rcu_read_unlock+0x53/0x24e
[<ffffffff810c98bb>] ? radix_tree_deref_slot+0x57/0x57
[<ffffffff8107b20e>] ? lock_release+0x6/0x1e9
[<ffffffff810c98dc>] rcu_read_unlock+0x21/0x23
[<ffffffff810c999f>] find_get_pages_tag+0xc1/0xed
[<ffffffff811c9e68>] ? start_this_handle+0x605/0x64d
[<ffffffff810d3480>] pagevec_lookup_tag+0x20/0x29
[<ffffffff811979e7>] write_cache_pages_da+0xeb/0x37b
[<ffffffff811ca10d>] ? jbd2_journal_start+0xe/0x10
[<ffffffff811aafae>] ? ext4_journal_start_sb+0x115/0x124
[<ffffffff81197f65>] ext4_da_writepages+0x2ee/0x565
[<ffffffff810d2ccf>] do_writepages+0x1f/0x28
[<ffffffff810ca39c>] __filemap_fdatawrite_range+0x4e/0x50
[<ffffffff810ca8a6>] filemap_flush+0x17/0x19
[<ffffffff81193447>] ext4_alloc_da_blocks+0x81/0xab
[<ffffffff8118d669>] ext4_release_file+0x35/0xb3
[<ffffffff8111b07c>] fput+0x117/0x1b2
[<ffffffff81117b56>] filp_close+0x6d/0x78
[<ffffffff8104c2f7>] put_files_struct+0xca/0x190
[<ffffffff8104c403>] exit_files+0x46/0x4e
[<ffffffff8104deb7>] do_exit+0x2b5/0x760
[<ffffffff81119196>] ? fsnotify_modify+0x5d/0x65
[<ffffffff8147f33d>] ? retint_swapgs+0x13/0x1b
[<ffffffff8104e3e0>] do_group_exit+0x7e/0xa9
[<ffffffff8104e41d>] sys_exit_group+0x12/0x16
[<ffffffff81484fd2>] system_call_fastpath+0x16/0x1b
Thanks,
Ben
--
Ben Greear <greearb@candelatech.com>
Candela Technologies Inc http://www.candelatech.com
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2011-07-12 20:51 UTC | newest]
Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2011-07-12 20:51 lockdep splat in 3.0-rc7 (circular locking, scheduler related?) Ben Greear
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.