* [syzbot] WARNING: locking bug in __perf_event_task_sched_in (2)
@ 2023-01-10 20:50 syzbot
2023-01-11 1:47 ` Jakub Kicinski
0 siblings, 1 reply; 2+ messages in thread
From: syzbot @ 2023-01-10 20:50 UTC (permalink / raw)
To: acme, alexander.shishkin, bpf, jolsa, linux-kernel,
linux-perf-users, mark.rutland, mingo, namhyung, netdev, peterz,
syzkaller-bugs
Hello,
syzbot found the following issue on:
HEAD commit: 6d0c4b11e743 libbpf: Poison strlcpy()
git tree: bpf-next
console output: https://syzkaller.appspot.com/x/log.txt?x=17764f3a480000
kernel config: https://syzkaller.appspot.com/x/.config?x=46221e8203c7aca6
dashboard link: https://syzkaller.appspot.com/bug?extid=d94d214ea473e218fc89
compiler: gcc (Debian 10.2.1-6) 10.2.1 20210110, GNU ld (GNU Binutils for Debian) 2.35.2
Unfortunately, I don't have any reproducer for this issue yet.
Downloadable assets:
disk image: https://storage.googleapis.com/syzbot-assets/83567aa48724/disk-6d0c4b11.raw.xz
vmlinux: https://storage.googleapis.com/syzbot-assets/6047fdb8660e/vmlinux-6d0c4b11.xz
kernel image: https://storage.googleapis.com/syzbot-assets/a94d1047d7b7/bzImage-6d0c4b11.xz
IMPORTANT: if you fix the issue, please add the following tag to the commit:
Reported-by: syzbot+d94d214ea473e218fc89@syzkaller.appspotmail.com
------------[ cut here ]------------
DEBUG_LOCKS_WARN_ON(1)
WARNING: CPU: 1 PID: 6975 at kernel/locking/lockdep.c:231 hlock_class kernel/locking/lockdep.c:231 [inline]
WARNING: CPU: 1 PID: 6975 at kernel/locking/lockdep.c:231 hlock_class kernel/locking/lockdep.c:220 [inline]
WARNING: CPU: 1 PID: 6975 at kernel/locking/lockdep.c:231 check_wait_context kernel/locking/lockdep.c:4754 [inline]
WARNING: CPU: 1 PID: 6975 at kernel/locking/lockdep.c:231 __lock_acquire+0xecf/0x56d0 kernel/locking/lockdep.c:5005
Modules linked in:
CPU: 1 PID: 6975 Comm: kworker/u4:13 Not tainted 6.2.0-rc2-syzkaller-00302-g6d0c4b11e743 #0
Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 10/26/2022
Workqueue: netns cleanup_net
RIP: 0010:hlock_class kernel/locking/lockdep.c:231 [inline]
RIP: 0010:hlock_class kernel/locking/lockdep.c:220 [inline]
RIP: 0010:check_wait_context kernel/locking/lockdep.c:4754 [inline]
RIP: 0010:__lock_acquire+0xecf/0x56d0 kernel/locking/lockdep.c:5005
Code: 28 14 73 8e e8 02 bb 6b 00 8b 05 94 78 0f 0d 85 c0 0f 85 79 f8 ff ff 48 c7 c6 40 51 4c 8a 48 c7 c7 a0 4a 4c 8a e8 51 d6 5b 08 <0f> 0b 31 c0 e9 73 f7 ff ff 48 63 5c 24 18 be 08 00 00 00 48 89 d8
RSP: 0018:ffffc900035ef518 EFLAGS: 00010082
RAX: 0000000000000000 RBX: 000000000000070e RCX: 0000000000000000
RDX: ffff888079031d40 RSI: ffffffff8166724c RDI: fffff520006bde95
RBP: 0000000000000003 R08: 0000000000000005 R09: 0000000000000000
R10: 0000000080000002 R11: 0000000000000001 R12: 0000000000000003
R13: ffff888079031d40 R14: ffff888079032778 R15: dffffc0000000000
FS: 0000000000000000(0000) GS:ffff8880b9900000(0000) knlGS:0000000000000000
CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
CR2: 00007ffc92edd748 CR3: 0000000044fa3000 CR4: 00000000003506e0
DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
Call Trace:
<TASK>
lock_acquire kernel/locking/lockdep.c:5668 [inline]
lock_acquire+0x1e3/0x630 kernel/locking/lockdep.c:5633
rcu_lock_acquire include/linux/rcupdate.h:325 [inline]
rcu_read_lock include/linux/rcupdate.h:764 [inline]
perf_event_context_sched_in kernel/events/core.c:3913 [inline]
__perf_event_task_sched_in+0xe2/0x6c0 kernel/events/core.c:3980
perf_event_task_sched_in include/linux/perf_event.h:1328 [inline]
finish_task_switch.isra.0+0x5e5/0xc80 kernel/sched/core.c:5118
context_switch kernel/sched/core.c:5247 [inline]
__schedule+0xb92/0x5450 kernel/sched/core.c:6555
schedule+0xde/0x1b0 kernel/sched/core.c:6631
schedule_preempt_disabled+0x13/0x20 kernel/sched/core.c:6690
__mutex_lock_common kernel/locking/mutex.c:679 [inline]
__mutex_lock+0xa48/0x1360 kernel/locking/mutex.c:747
devl_lock net/devlink/core.c:54 [inline]
devlink_pernet_pre_exit+0x10a/0x220 net/devlink/core.c:301
ops_pre_exit_list net/core/net_namespace.c:159 [inline]
cleanup_net+0x455/0xb10 net/core/net_namespace.c:594
process_one_work+0x9bf/0x1710 kernel/workqueue.c:2289
worker_thread+0x669/0x1090 kernel/workqueue.c:2436
kthread+0x2e8/0x3a0 kernel/kthread.c:376
ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:308
</TASK>
---
This report is generated by a bot. It may contain errors.
See https://goo.gl/tpsmEJ for more information about syzbot.
syzbot engineers can be reached at syzkaller@googlegroups.com.
syzbot will keep track of this issue. See:
https://goo.gl/tpsmEJ#status for how to communicate with syzbot.
^ permalink raw reply [flat|nested] 2+ messages in thread
* Re: [syzbot] WARNING: locking bug in __perf_event_task_sched_in (2)
2023-01-10 20:50 [syzbot] WARNING: locking bug in __perf_event_task_sched_in (2) syzbot
@ 2023-01-11 1:47 ` Jakub Kicinski
0 siblings, 0 replies; 2+ messages in thread
From: Jakub Kicinski @ 2023-01-11 1:47 UTC (permalink / raw)
To: syzbot
Cc: acme, alexander.shishkin, bpf, jolsa, linux-kernel,
linux-perf-users, mark.rutland, mingo, namhyung, netdev, peterz,
syzkaller-bugs
On Tue, 10 Jan 2023 12:50:48 -0800 syzbot wrote:
> <TASK>
> lock_acquire kernel/locking/lockdep.c:5668 [inline]
> lock_acquire+0x1e3/0x630 kernel/locking/lockdep.c:5633
> rcu_lock_acquire include/linux/rcupdate.h:325 [inline]
> rcu_read_lock include/linux/rcupdate.h:764 [inline]
> perf_event_context_sched_in kernel/events/core.c:3913 [inline]
> __perf_event_task_sched_in+0xe2/0x6c0 kernel/events/core.c:3980
> perf_event_task_sched_in include/linux/perf_event.h:1328 [inline]
> finish_task_switch.isra.0+0x5e5/0xc80 kernel/sched/core.c:5118
> context_switch kernel/sched/core.c:5247 [inline]
> __schedule+0xb92/0x5450 kernel/sched/core.c:6555
> schedule+0xde/0x1b0 kernel/sched/core.c:6631
> schedule_preempt_disabled+0x13/0x20 kernel/sched/core.c:6690
> __mutex_lock_common kernel/locking/mutex.c:679 [inline]
> __mutex_lock+0xa48/0x1360 kernel/locking/mutex.c:747
> devl_lock net/devlink/core.c:54 [inline]
> devlink_pernet_pre_exit+0x10a/0x220 net/devlink/core.c:301
> ops_pre_exit_list net/core/net_namespace.c:159 [inline]
> cleanup_net+0x455/0xb10 net/core/net_namespace.c:594
> process_one_work+0x9bf/0x1710 kernel/workqueue.c:2289
> worker_thread+0x669/0x1090 kernel/workqueue.c:2436
> kthread+0x2e8/0x3a0 kernel/kthread.c:376
> ret_from_fork+0x1f/0x30 arch/x86/entry/entry_64.S:308
> </TASK>
Yes, I pooped it. We need to keep the mutex around as well as
the devlink instance memory, otherwise locked screams.
Fix building..
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2023-01-11 1:47 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2023-01-10 20:50 [syzbot] WARNING: locking bug in __perf_event_task_sched_in (2) syzbot
2023-01-11 1:47 ` Jakub Kicinski
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.