linux-arm-msm.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* i.MX53 error during GPU use
@ 2021-04-23 11:58 Otavio Salvador
  2021-04-23 14:35 ` Rob Clark
  0 siblings, 1 reply; 3+ messages in thread
From: Otavio Salvador @ 2021-04-23 11:58 UTC (permalink / raw)
  To: Rob Clark, Sean Paul, David Airlie, Daniel Vetter, linux-arm-msm,
	DRI mailing list, freedreno

Hello all,

We found this error when using Freedreno driver on an i.MX53 device
with Wayland. Any idea how to fix this?

[   32.414110] [drm:msm_ioctl_gem_submit] *ERROR* invalid cmdstream size: 0
[   39.177075]
[   39.178617] ======================================================
[   39.184804] WARNING: possible circular locking dependency detected
[   39.190997] 5.10.31+g7ae1de1d2bd3 #1 Not tainted
[   39.195619] ------------------------------------------------------
[   39.201805] kworker/u2:1/68 is trying to acquire lock:
[   39.206953] c2b952fc (reservation_ww_class_mutex){+.+.}-{3:3}, at:
dma_buf_detach+0xd0/0x1e0
[   39.215448]
[   39.215448] but task is already holding lock:
[   39.221286] c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at:
msm_gem_free_work+0x118/0x1b0
[   39.229062]
[   39.229062] which lock already depends on the new lock.
[   39.229062]
[   39.237243]
[   39.237243] the existing dependency chain (in reverse order) is:
[   39.244729]
[   39.244729] -> #1 (&msm_obj->lock){+.+.}-{3:3}:
[   39.250769]        mutex_lock_nested+0x1c/0x24
[   39.255220]        msm_gem_get_and_pin_iova_range+0x38/0x170
[   39.260885]        msm_gem_get_and_pin_iova+0x24/0x2c
[   39.265946]        msm_ioctl_gem_submit+0x12c0/0x1964
[   39.271008]        drm_ioctl+0x1f0/0x3e0
[   39.274940]        sys_ioctl+0x3bc/0xbe0
[   39.278873]        ret_fast_syscall+0x0/0x28
[   39.283150]        0xbed4deac
[   39.286122]
[   39.286122] -> #0 (reservation_ww_class_mutex){+.+.}-{3:3}:
[   39.293208]        lock_acquire+0x134/0x560
[   39.297398]        __ww_mutex_lock.constprop.0+0xbc/0x1124
[   39.302889]        ww_mutex_lock+0x68/0xb4
[   39.306993]        dma_buf_detach+0xd0/0x1e0
[   39.311271]        drm_prime_gem_destroy+0x2c/0x38
[   39.316068]        msm_gem_free_work+0xa0/0x1b0
[   39.320611]        process_one_work+0x2a8/0x7ec
[   39.325149]        worker_thread+0x4c/0x554
[   39.329339]        kthread+0x140/0x154
[   39.333095]        ret_from_fork+0x14/0x20
[   39.337196]        0x0
[   39.339560]
[   39.339560] other info that might help us debug this:
[   39.339560]
[   39.347568]  Possible unsafe locking scenario:
[   39.347568]
[   39.353491]        CPU0                    CPU1
[   39.358025]        ----                    ----
[   39.362557]   lock(&msm_obj->lock);
[   39.366061]                                lock(reservation_ww_class_mutex);
[   39.373123]                                lock(&msm_obj->lock);
[   39.379143]   lock(reservation_ww_class_mutex);
[   39.383689]
[   39.383689]  *** DEADLOCK ***
[   39.383689]
[   39.389614] 4 locks held by kworker/u2:1/68:
[   39.393888]  #0: c24138a8 ((wq_completion)msm){+.+.}-{0:0}, at:
process_one_work+0x1f4/0x7ec
[   39.402362]  #1: c26a1f20
((work_completion)(&priv->free_work)){+.+.}-{0:0}, at:
process_one_work+0x1f4/0x7ec
[   39.412312]  #2: c24188a4 (&dev->struct_mutex){+.+.}-{3:3}, at:
msm_gem_free_work+0x68/0x1b0
[   39.420784]  #3: c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at:
msm_gem_free_work+0x118/0x1b0
[   39.428996]
[   39.428996] stack backtrace:
[   39.433363] CPU: 0 PID: 68 Comm: kworker/u2:1 Not tainted
5.10.31+g7ae1de1d2bd3 #1
[   39.440937] Hardware name: Freescale i.MX53 (Device Tree Support)
[   39.447044] Workqueue: msm msm_gem_free_work
[   39.451346] [<c0111258>] (unwind_backtrace) from [<c010ba00>]
(show_stack+0x10/0x14)
[   39.459107] [<c010ba00>] (show_stack) from [<c0e2b560>]
(dump_stack+0xdc/0x104)
[   39.466430] [<c0e2b560>] (dump_stack) from [<c01838d8>]
(check_noncircular+0xf0/0x160)
[   39.474358] [<c01838d8>] (check_noncircular) from [<c0187060>]
(__lock_acquire+0x1628/0x2e04)
[   39.482893] [<c0187060>] (__lock_acquire) from [<c01892e8>]
(lock_acquire+0x134/0x560)
[   39.490821] [<c01892e8>] (lock_acquire) from [<c0e3da60>]
(__ww_mutex_lock.constprop.0+0xbc/0x1124)
[   39.499877] [<c0e3da60>] (__ww_mutex_lock.constprop.0) from
[<c0e3ebe4>] (ww_mutex_lock+0x68/0xb4)
[   39.508844] [<c0e3ebe4>] (ww_mutex_lock) from [<c080e6a8>]
(dma_buf_detach+0xd0/0x1e0)
[   39.516771] [<c080e6a8>] (dma_buf_detach) from [<c0701d04>]
(drm_prime_gem_destroy+0x2c/0x38)
[   39.525305] [<c0701d04>] (drm_prime_gem_destroy) from [<c077d334>]
(msm_gem_free_work+0xa0/0x1b0)
[   39.534188] [<c077d334>] (msm_gem_free_work) from [<c0147588>]
(process_one_work+0x2a8/0x7ec)
[   39.542725] [<c0147588>] (process_one_work) from [<c0147b18>]
(worker_thread+0x4c/0x554)
[   39.550825] [<c0147b18>] (worker_thread) from [<c014ebb0>]
(kthread+0x140/0x154)
[   39.558229] [<c014ebb0>] (kthread) from [<c0100134>]
(ret_from_fork+0x14/0x20)
[   39.565457] Exception stack(0xc26a1fb0 to 0xc26a1ff8)
[   39.570517] 1fa0:                                     00000000
00000000 00000000 00000000
[   39.578703] 1fc0: 00000000 00000000 00000000 00000000 00000000
00000000 00000000 00000000
[   39.586890] 1fe0: 00000000 00000000 00000000 00000000 00000013 00000000

This looks like a regression as it used to run fine on 5.4 Linux kernel.

Any idea of patch reference to test is welcome.

-- 
Otavio Salvador                             O.S. Systems
http://www.ossystems.com.br        http://code.ossystems.com.br
Mobile: +55 (53) 9 9981-7854          Mobile: +1 (347) 903-9750

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2021-05-04 17:11 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-04-23 11:58 i.MX53 error during GPU use Otavio Salvador
2021-04-23 14:35 ` Rob Clark
2021-05-04 17:11   ` Otavio Salvador

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).