From: Hans de Goede <hdegoede@redhat.com> To: Ben Skeggs <bskeggs@redhat.com>, Karol Herbst <kherbst@redhat.com>, Lyude Paul <lyude@redhat.com> Cc: "nouveau@lists.freedesktop.org" <nouveau@lists.freedesktop.org>, "dri-devel@lists.freedesktop.org" <dri-devel@lists.freedesktop.org> Subject: [Nouveau] nouveau lockdep deadlock report with 5.18-rc6 Date: Tue, 17 May 2022 13:10:07 +0200 [thread overview] Message-ID: <ac39455b-b85c-4cf7-8cd0-089325c9514a@redhat.com> (raw) Hi All, I just noticed the below lockdep possible deadlock report with a 5.18-rc6 kernel on a Dell Latitude E6430 laptop with the following nvidia GPU: 01:00.0 VGA compatible controller [0300]: NVIDIA Corporation GF108GLM [NVS 5200M] [10de:0dfc] (rev a1) 01:00.1 Audio device [0403]: NVIDIA Corporation GF108 High Definition Audio Controller [10de:0bea] (rev a1) This is with the laptop in Optimus mode, so with the Intel integrated gfx from the i5-3320M CPU driving the LCD panel and with nothing connected to the HDMI connector, which is always routed to the NVIDIA GPU on this laptop. The lockdep possible deadlock warning seems to happen when the NVIDIA GPU is runtime suspended shortly after gdm has loaded: [ 24.859171] ====================================================== [ 24.859173] WARNING: possible circular locking dependency detected [ 24.859175] 5.18.0-rc6+ #34 Tainted: G E [ 24.859178] ------------------------------------------------------ [ 24.859179] kworker/1:1/46 is trying to acquire lock: [ 24.859181] ffff92b0c0ee0518 (&cli->mutex){+.+.}-{3:3}, at: nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859231] but task is already holding lock: [ 24.859233] ffff92b0c4bf35a0 (reservation_ww_class_mutex){+.+.}-{3:3}, at: ttm_bo_wait+0x7d/0x140 [ttm] [ 24.859243] which lock already depends on the new lock. [ 24.859244] the existing dependency chain (in reverse order) is: [ 24.859246] -> #1 (reservation_ww_class_mutex){+.+.}-{3:3}: [ 24.859249] __ww_mutex_lock.constprop.0+0xb3/0xfb0 [ 24.859256] ww_mutex_lock+0x38/0xa0 [ 24.859259] nouveau_bo_pin+0x30/0x380 [nouveau] [ 24.859297] nouveau_channel_del+0x1d7/0x3e0 [nouveau] [ 24.859328] nouveau_channel_new+0x48/0x730 [nouveau] [ 24.859358] nouveau_abi16_ioctl_channel_alloc+0x113/0x360 [nouveau] [ 24.859389] drm_ioctl_kernel+0xa1/0x150 [ 24.859392] drm_ioctl+0x21c/0x410 [ 24.859395] nouveau_drm_ioctl+0x56/0x1820 [nouveau] [ 24.859431] __x64_sys_ioctl+0x8d/0xc0 [ 24.859436] do_syscall_64+0x5b/0x80 [ 24.859440] entry_SYSCALL_64_after_hwframe+0x44/0xae [ 24.859443] -> #0 (&cli->mutex){+.+.}-{3:3}: [ 24.859446] __lock_acquire+0x12e2/0x1f90 [ 24.859450] lock_acquire+0xad/0x290 [ 24.859453] __mutex_lock+0x90/0x830 [ 24.859456] nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859493] ttm_bo_move_to_lru_tail+0x32c/0x980 [ttm] [ 24.859498] ttm_mem_evict_first+0x25c/0x4b0 [ttm] [ 24.859503] ttm_resource_manager_evict_all+0x93/0x1b0 [ttm] [ 24.859509] nouveau_debugfs_fini+0x161/0x260 [nouveau] [ 24.859545] nouveau_drm_ioctl+0xa4a/0x1820 [nouveau] [ 24.859582] pci_pm_runtime_suspend+0x5c/0x180 [ 24.859585] __rpm_callback+0x48/0x1b0 [ 24.859589] rpm_callback+0x5a/0x70 [ 24.859591] rpm_suspend+0x10a/0x6f0 [ 24.859594] pm_runtime_work+0xa0/0xb0 [ 24.859596] process_one_work+0x254/0x560 [ 24.859601] worker_thread+0x4f/0x390 [ 24.859604] kthread+0xe6/0x110 [ 24.859607] ret_from_fork+0x22/0x30 [ 24.859611] other info that might help us debug this: [ 24.859612] Possible unsafe locking scenario: [ 24.859613] CPU0 CPU1 [ 24.859615] ---- ---- [ 24.859616] lock(reservation_ww_class_mutex); [ 24.859618] lock(&cli->mutex); [ 24.859620] lock(reservation_ww_class_mutex); [ 24.859622] lock(&cli->mutex); [ 24.859624] *** DEADLOCK *** [ 24.859625] 3 locks held by kworker/1:1/46: [ 24.859627] #0: ffff92b0c0bb4338 ((wq_completion)pm){+.+.}-{0:0}, at: process_one_work+0x1d0/0x560 [ 24.859634] #1: ffffa8ffc02dfe80 ((work_completion)(&dev->power.work)){+.+.}-{0:0}, at: process_one_work+0x1d0/0x560 [ 24.859641] #2: ffff92b0c4bf35a0 (reservation_ww_class_mutex){+.+.}-{3:3}, at: ttm_bo_wait+0x7d/0x140 [ttm] [ 24.859649] stack backtrace: [ 24.859651] CPU: 1 PID: 46 Comm: kworker/1:1 Tainted: G E 5.18.0-rc6+ #34 [ 24.859654] Hardware name: Dell Inc. Latitude E6430/0H3MT5, BIOS A21 05/08/2017 [ 24.859656] Workqueue: pm pm_runtime_work [ 24.859660] Call Trace: [ 24.859662] <TASK> [ 24.859665] dump_stack_lvl+0x5b/0x74 [ 24.859669] check_noncircular+0xdf/0x100 [ 24.859672] ? register_lock_class+0x38/0x470 [ 24.859678] __lock_acquire+0x12e2/0x1f90 [ 24.859683] lock_acquire+0xad/0x290 [ 24.859686] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859724] ? lock_is_held_type+0xa6/0x120 [ 24.859730] __mutex_lock+0x90/0x830 [ 24.859733] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859770] ? nvif_vmm_map+0x114/0x130 [nouveau] [ 24.859791] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859829] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859866] nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859905] ttm_bo_move_to_lru_tail+0x32c/0x980 [ttm] [ 24.859912] ttm_mem_evict_first+0x25c/0x4b0 [ttm] [ 24.859919] ? lock_release+0x20/0x2a0 [ 24.859923] ttm_resource_manager_evict_all+0x93/0x1b0 [ttm] [ 24.859930] nouveau_debugfs_fini+0x161/0x260 [nouveau] [ 24.859968] nouveau_drm_ioctl+0xa4a/0x1820 [nouveau] [ 24.860005] pci_pm_runtime_suspend+0x5c/0x180 [ 24.860008] ? pci_dev_put+0x20/0x20 [ 24.860011] __rpm_callback+0x48/0x1b0 [ 24.860014] ? pci_dev_put+0x20/0x20 [ 24.860018] rpm_callback+0x5a/0x70 [ 24.860020] ? pci_dev_put+0x20/0x20 [ 24.860023] rpm_suspend+0x10a/0x6f0 [ 24.860025] ? process_one_work+0x1d0/0x560 [ 24.860031] pm_runtime_work+0xa0/0xb0 [ 24.860034] process_one_work+0x254/0x560 [ 24.860039] worker_thread+0x4f/0x390 [ 24.860043] ? process_one_work+0x560/0x560 [ 24.860046] kthread+0xe6/0x110 [ 24.860049] ? kthread_complete_and_exit+0x20/0x20 [ 24.860053] ret_from_fork+0x22/0x30 [ 24.860059] </TASK> Regards, Hans
WARNING: multiple messages have this Message-ID (diff)
From: Hans de Goede <hdegoede@redhat.com> To: Ben Skeggs <bskeggs@redhat.com>, Karol Herbst <kherbst@redhat.com>, Lyude Paul <lyude@redhat.com> Cc: "nouveau@lists.freedesktop.org" <nouveau@lists.freedesktop.org>, "dri-devel@lists.freedesktop.org" <dri-devel@lists.freedesktop.org> Subject: nouveau lockdep deadlock report with 5.18-rc6 Date: Tue, 17 May 2022 13:10:07 +0200 [thread overview] Message-ID: <ac39455b-b85c-4cf7-8cd0-089325c9514a@redhat.com> (raw) Hi All, I just noticed the below lockdep possible deadlock report with a 5.18-rc6 kernel on a Dell Latitude E6430 laptop with the following nvidia GPU: 01:00.0 VGA compatible controller [0300]: NVIDIA Corporation GF108GLM [NVS 5200M] [10de:0dfc] (rev a1) 01:00.1 Audio device [0403]: NVIDIA Corporation GF108 High Definition Audio Controller [10de:0bea] (rev a1) This is with the laptop in Optimus mode, so with the Intel integrated gfx from the i5-3320M CPU driving the LCD panel and with nothing connected to the HDMI connector, which is always routed to the NVIDIA GPU on this laptop. The lockdep possible deadlock warning seems to happen when the NVIDIA GPU is runtime suspended shortly after gdm has loaded: [ 24.859171] ====================================================== [ 24.859173] WARNING: possible circular locking dependency detected [ 24.859175] 5.18.0-rc6+ #34 Tainted: G E [ 24.859178] ------------------------------------------------------ [ 24.859179] kworker/1:1/46 is trying to acquire lock: [ 24.859181] ffff92b0c0ee0518 (&cli->mutex){+.+.}-{3:3}, at: nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859231] but task is already holding lock: [ 24.859233] ffff92b0c4bf35a0 (reservation_ww_class_mutex){+.+.}-{3:3}, at: ttm_bo_wait+0x7d/0x140 [ttm] [ 24.859243] which lock already depends on the new lock. [ 24.859244] the existing dependency chain (in reverse order) is: [ 24.859246] -> #1 (reservation_ww_class_mutex){+.+.}-{3:3}: [ 24.859249] __ww_mutex_lock.constprop.0+0xb3/0xfb0 [ 24.859256] ww_mutex_lock+0x38/0xa0 [ 24.859259] nouveau_bo_pin+0x30/0x380 [nouveau] [ 24.859297] nouveau_channel_del+0x1d7/0x3e0 [nouveau] [ 24.859328] nouveau_channel_new+0x48/0x730 [nouveau] [ 24.859358] nouveau_abi16_ioctl_channel_alloc+0x113/0x360 [nouveau] [ 24.859389] drm_ioctl_kernel+0xa1/0x150 [ 24.859392] drm_ioctl+0x21c/0x410 [ 24.859395] nouveau_drm_ioctl+0x56/0x1820 [nouveau] [ 24.859431] __x64_sys_ioctl+0x8d/0xc0 [ 24.859436] do_syscall_64+0x5b/0x80 [ 24.859440] entry_SYSCALL_64_after_hwframe+0x44/0xae [ 24.859443] -> #0 (&cli->mutex){+.+.}-{3:3}: [ 24.859446] __lock_acquire+0x12e2/0x1f90 [ 24.859450] lock_acquire+0xad/0x290 [ 24.859453] __mutex_lock+0x90/0x830 [ 24.859456] nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859493] ttm_bo_move_to_lru_tail+0x32c/0x980 [ttm] [ 24.859498] ttm_mem_evict_first+0x25c/0x4b0 [ttm] [ 24.859503] ttm_resource_manager_evict_all+0x93/0x1b0 [ttm] [ 24.859509] nouveau_debugfs_fini+0x161/0x260 [nouveau] [ 24.859545] nouveau_drm_ioctl+0xa4a/0x1820 [nouveau] [ 24.859582] pci_pm_runtime_suspend+0x5c/0x180 [ 24.859585] __rpm_callback+0x48/0x1b0 [ 24.859589] rpm_callback+0x5a/0x70 [ 24.859591] rpm_suspend+0x10a/0x6f0 [ 24.859594] pm_runtime_work+0xa0/0xb0 [ 24.859596] process_one_work+0x254/0x560 [ 24.859601] worker_thread+0x4f/0x390 [ 24.859604] kthread+0xe6/0x110 [ 24.859607] ret_from_fork+0x22/0x30 [ 24.859611] other info that might help us debug this: [ 24.859612] Possible unsafe locking scenario: [ 24.859613] CPU0 CPU1 [ 24.859615] ---- ---- [ 24.859616] lock(reservation_ww_class_mutex); [ 24.859618] lock(&cli->mutex); [ 24.859620] lock(reservation_ww_class_mutex); [ 24.859622] lock(&cli->mutex); [ 24.859624] *** DEADLOCK *** [ 24.859625] 3 locks held by kworker/1:1/46: [ 24.859627] #0: ffff92b0c0bb4338 ((wq_completion)pm){+.+.}-{0:0}, at: process_one_work+0x1d0/0x560 [ 24.859634] #1: ffffa8ffc02dfe80 ((work_completion)(&dev->power.work)){+.+.}-{0:0}, at: process_one_work+0x1d0/0x560 [ 24.859641] #2: ffff92b0c4bf35a0 (reservation_ww_class_mutex){+.+.}-{3:3}, at: ttm_bo_wait+0x7d/0x140 [ttm] [ 24.859649] stack backtrace: [ 24.859651] CPU: 1 PID: 46 Comm: kworker/1:1 Tainted: G E 5.18.0-rc6+ #34 [ 24.859654] Hardware name: Dell Inc. Latitude E6430/0H3MT5, BIOS A21 05/08/2017 [ 24.859656] Workqueue: pm pm_runtime_work [ 24.859660] Call Trace: [ 24.859662] <TASK> [ 24.859665] dump_stack_lvl+0x5b/0x74 [ 24.859669] check_noncircular+0xdf/0x100 [ 24.859672] ? register_lock_class+0x38/0x470 [ 24.859678] __lock_acquire+0x12e2/0x1f90 [ 24.859683] lock_acquire+0xad/0x290 [ 24.859686] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859724] ? lock_is_held_type+0xa6/0x120 [ 24.859730] __mutex_lock+0x90/0x830 [ 24.859733] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859770] ? nvif_vmm_map+0x114/0x130 [nouveau] [ 24.859791] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859829] ? nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859866] nouveau_vga_lastclose+0x910/0x1030 [nouveau] [ 24.859905] ttm_bo_move_to_lru_tail+0x32c/0x980 [ttm] [ 24.859912] ttm_mem_evict_first+0x25c/0x4b0 [ttm] [ 24.859919] ? lock_release+0x20/0x2a0 [ 24.859923] ttm_resource_manager_evict_all+0x93/0x1b0 [ttm] [ 24.859930] nouveau_debugfs_fini+0x161/0x260 [nouveau] [ 24.859968] nouveau_drm_ioctl+0xa4a/0x1820 [nouveau] [ 24.860005] pci_pm_runtime_suspend+0x5c/0x180 [ 24.860008] ? pci_dev_put+0x20/0x20 [ 24.860011] __rpm_callback+0x48/0x1b0 [ 24.860014] ? pci_dev_put+0x20/0x20 [ 24.860018] rpm_callback+0x5a/0x70 [ 24.860020] ? pci_dev_put+0x20/0x20 [ 24.860023] rpm_suspend+0x10a/0x6f0 [ 24.860025] ? process_one_work+0x1d0/0x560 [ 24.860031] pm_runtime_work+0xa0/0xb0 [ 24.860034] process_one_work+0x254/0x560 [ 24.860039] worker_thread+0x4f/0x390 [ 24.860043] ? process_one_work+0x560/0x560 [ 24.860046] kthread+0xe6/0x110 [ 24.860049] ? kthread_complete_and_exit+0x20/0x20 [ 24.860053] ret_from_fork+0x22/0x30 [ 24.860059] </TASK> Regards, Hans
next reply other threads:[~2022-05-17 11:10 UTC|newest] Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top 2022-05-17 11:10 Hans de Goede [this message] 2022-05-17 11:10 ` nouveau lockdep deadlock report with 5.18-rc6 Hans de Goede 2022-05-17 22:24 ` [Nouveau] " Lyude Paul 2022-05-17 22:24 ` Lyude Paul 2022-05-18 17:42 ` [Nouveau] " Lyude Paul 2022-05-18 17:42 ` Lyude Paul 2022-05-20 11:46 ` [Nouveau] " Computer Enthusiastic 2022-05-20 11:46 ` Computer Enthusiastic 2022-05-23 19:59 ` Lyude Paul 2022-05-23 19:59 ` Lyude Paul
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=ac39455b-b85c-4cf7-8cd0-089325c9514a@redhat.com \ --to=hdegoede@redhat.com \ --cc=bskeggs@redhat.com \ --cc=dri-devel@lists.freedesktop.org \ --cc=kherbst@redhat.com \ --cc=lyude@redhat.com \ --cc=nouveau@lists.freedesktop.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.