* i.MX53 error during GPU use
@ 2021-04-23 11:58 Otavio Salvador
2021-04-23 14:35 ` Rob Clark
0 siblings, 1 reply; 3+ messages in thread
From: Otavio Salvador @ 2021-04-23 11:58 UTC (permalink / raw)
To: Rob Clark, Sean Paul, David Airlie, Daniel Vetter, linux-arm-msm,
DRI mailing list, freedreno
Hello all,
We found this error when using Freedreno driver on an i.MX53 device
with Wayland. Any idea how to fix this?
[ 32.414110] [drm:msm_ioctl_gem_submit] *ERROR* invalid cmdstream size: 0
[ 39.177075]
[ 39.178617] ======================================================
[ 39.184804] WARNING: possible circular locking dependency detected
[ 39.190997] 5.10.31+g7ae1de1d2bd3 #1 Not tainted
[ 39.195619] ------------------------------------------------------
[ 39.201805] kworker/u2:1/68 is trying to acquire lock:
[ 39.206953] c2b952fc (reservation_ww_class_mutex){+.+.}-{3:3}, at:
dma_buf_detach+0xd0/0x1e0
[ 39.215448]
[ 39.215448] but task is already holding lock:
[ 39.221286] c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at:
msm_gem_free_work+0x118/0x1b0
[ 39.229062]
[ 39.229062] which lock already depends on the new lock.
[ 39.229062]
[ 39.237243]
[ 39.237243] the existing dependency chain (in reverse order) is:
[ 39.244729]
[ 39.244729] -> #1 (&msm_obj->lock){+.+.}-{3:3}:
[ 39.250769] mutex_lock_nested+0x1c/0x24
[ 39.255220] msm_gem_get_and_pin_iova_range+0x38/0x170
[ 39.260885] msm_gem_get_and_pin_iova+0x24/0x2c
[ 39.265946] msm_ioctl_gem_submit+0x12c0/0x1964
[ 39.271008] drm_ioctl+0x1f0/0x3e0
[ 39.274940] sys_ioctl+0x3bc/0xbe0
[ 39.278873] ret_fast_syscall+0x0/0x28
[ 39.283150] 0xbed4deac
[ 39.286122]
[ 39.286122] -> #0 (reservation_ww_class_mutex){+.+.}-{3:3}:
[ 39.293208] lock_acquire+0x134/0x560
[ 39.297398] __ww_mutex_lock.constprop.0+0xbc/0x1124
[ 39.302889] ww_mutex_lock+0x68/0xb4
[ 39.306993] dma_buf_detach+0xd0/0x1e0
[ 39.311271] drm_prime_gem_destroy+0x2c/0x38
[ 39.316068] msm_gem_free_work+0xa0/0x1b0
[ 39.320611] process_one_work+0x2a8/0x7ec
[ 39.325149] worker_thread+0x4c/0x554
[ 39.329339] kthread+0x140/0x154
[ 39.333095] ret_from_fork+0x14/0x20
[ 39.337196] 0x0
[ 39.339560]
[ 39.339560] other info that might help us debug this:
[ 39.339560]
[ 39.347568] Possible unsafe locking scenario:
[ 39.347568]
[ 39.353491] CPU0 CPU1
[ 39.358025] ---- ----
[ 39.362557] lock(&msm_obj->lock);
[ 39.366061] lock(reservation_ww_class_mutex);
[ 39.373123] lock(&msm_obj->lock);
[ 39.379143] lock(reservation_ww_class_mutex);
[ 39.383689]
[ 39.383689] *** DEADLOCK ***
[ 39.383689]
[ 39.389614] 4 locks held by kworker/u2:1/68:
[ 39.393888] #0: c24138a8 ((wq_completion)msm){+.+.}-{0:0}, at:
process_one_work+0x1f4/0x7ec
[ 39.402362] #1: c26a1f20
((work_completion)(&priv->free_work)){+.+.}-{0:0}, at:
process_one_work+0x1f4/0x7ec
[ 39.412312] #2: c24188a4 (&dev->struct_mutex){+.+.}-{3:3}, at:
msm_gem_free_work+0x68/0x1b0
[ 39.420784] #3: c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at:
msm_gem_free_work+0x118/0x1b0
[ 39.428996]
[ 39.428996] stack backtrace:
[ 39.433363] CPU: 0 PID: 68 Comm: kworker/u2:1 Not tainted
5.10.31+g7ae1de1d2bd3 #1
[ 39.440937] Hardware name: Freescale i.MX53 (Device Tree Support)
[ 39.447044] Workqueue: msm msm_gem_free_work
[ 39.451346] [<c0111258>] (unwind_backtrace) from [<c010ba00>]
(show_stack+0x10/0x14)
[ 39.459107] [<c010ba00>] (show_stack) from [<c0e2b560>]
(dump_stack+0xdc/0x104)
[ 39.466430] [<c0e2b560>] (dump_stack) from [<c01838d8>]
(check_noncircular+0xf0/0x160)
[ 39.474358] [<c01838d8>] (check_noncircular) from [<c0187060>]
(__lock_acquire+0x1628/0x2e04)
[ 39.482893] [<c0187060>] (__lock_acquire) from [<c01892e8>]
(lock_acquire+0x134/0x560)
[ 39.490821] [<c01892e8>] (lock_acquire) from [<c0e3da60>]
(__ww_mutex_lock.constprop.0+0xbc/0x1124)
[ 39.499877] [<c0e3da60>] (__ww_mutex_lock.constprop.0) from
[<c0e3ebe4>] (ww_mutex_lock+0x68/0xb4)
[ 39.508844] [<c0e3ebe4>] (ww_mutex_lock) from [<c080e6a8>]
(dma_buf_detach+0xd0/0x1e0)
[ 39.516771] [<c080e6a8>] (dma_buf_detach) from [<c0701d04>]
(drm_prime_gem_destroy+0x2c/0x38)
[ 39.525305] [<c0701d04>] (drm_prime_gem_destroy) from [<c077d334>]
(msm_gem_free_work+0xa0/0x1b0)
[ 39.534188] [<c077d334>] (msm_gem_free_work) from [<c0147588>]
(process_one_work+0x2a8/0x7ec)
[ 39.542725] [<c0147588>] (process_one_work) from [<c0147b18>]
(worker_thread+0x4c/0x554)
[ 39.550825] [<c0147b18>] (worker_thread) from [<c014ebb0>]
(kthread+0x140/0x154)
[ 39.558229] [<c014ebb0>] (kthread) from [<c0100134>]
(ret_from_fork+0x14/0x20)
[ 39.565457] Exception stack(0xc26a1fb0 to 0xc26a1ff8)
[ 39.570517] 1fa0: 00000000
00000000 00000000 00000000
[ 39.578703] 1fc0: 00000000 00000000 00000000 00000000 00000000
00000000 00000000 00000000
[ 39.586890] 1fe0: 00000000 00000000 00000000 00000000 00000013 00000000
This looks like a regression as it used to run fine on 5.4 Linux kernel.
Any idea of patch reference to test is welcome.
--
Otavio Salvador O.S. Systems
http://www.ossystems.com.br http://code.ossystems.com.br
Mobile: +55 (53) 9 9981-7854 Mobile: +1 (347) 903-9750
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: i.MX53 error during GPU use
2021-04-23 11:58 i.MX53 error during GPU use Otavio Salvador
@ 2021-04-23 14:35 ` Rob Clark
2021-05-04 17:11 ` Otavio Salvador
0 siblings, 1 reply; 3+ messages in thread
From: Rob Clark @ 2021-04-23 14:35 UTC (permalink / raw)
To: Otavio Salvador
Cc: Sean Paul, David Airlie, Daniel Vetter, linux-arm-msm,
DRI mailing list, freedreno
On Fri, Apr 23, 2021 at 4:58 AM Otavio Salvador
<otavio.salvador@ossystems.com.br> wrote:
>
> Hello all,
>
> We found this error when using Freedreno driver on an i.MX53 device
> with Wayland. Any idea how to fix this?
>
> [ 32.414110] [drm:msm_ioctl_gem_submit] *ERROR* invalid cmdstream size: 0
The invalid cmdstream size is some sort of userspace error
> [ 39.177075]
> [ 39.178617] ======================================================
> [ 39.184804] WARNING: possible circular locking dependency detected
> [ 39.190997] 5.10.31+g7ae1de1d2bd3 #1 Not tainted
> [ 39.195619] ------------------------------------------------------
But possibly it is triggering the lockdep anger? It looks like the
gem locking re-work landed in v5.11.. any chance you can try a newer
kernel?
BR,
-R
> [ 39.201805] kworker/u2:1/68 is trying to acquire lock:
> [ 39.206953] c2b952fc (reservation_ww_class_mutex){+.+.}-{3:3}, at:
> dma_buf_detach+0xd0/0x1e0
> [ 39.215448]
> [ 39.215448] but task is already holding lock:
> [ 39.221286] c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at:
> msm_gem_free_work+0x118/0x1b0
> [ 39.229062]
> [ 39.229062] which lock already depends on the new lock.
> [ 39.229062]
> [ 39.237243]
> [ 39.237243] the existing dependency chain (in reverse order) is:
> [ 39.244729]
> [ 39.244729] -> #1 (&msm_obj->lock){+.+.}-{3:3}:
> [ 39.250769] mutex_lock_nested+0x1c/0x24
> [ 39.255220] msm_gem_get_and_pin_iova_range+0x38/0x170
> [ 39.260885] msm_gem_get_and_pin_iova+0x24/0x2c
> [ 39.265946] msm_ioctl_gem_submit+0x12c0/0x1964
> [ 39.271008] drm_ioctl+0x1f0/0x3e0
> [ 39.274940] sys_ioctl+0x3bc/0xbe0
> [ 39.278873] ret_fast_syscall+0x0/0x28
> [ 39.283150] 0xbed4deac
> [ 39.286122]
> [ 39.286122] -> #0 (reservation_ww_class_mutex){+.+.}-{3:3}:
> [ 39.293208] lock_acquire+0x134/0x560
> [ 39.297398] __ww_mutex_lock.constprop.0+0xbc/0x1124
> [ 39.302889] ww_mutex_lock+0x68/0xb4
> [ 39.306993] dma_buf_detach+0xd0/0x1e0
> [ 39.311271] drm_prime_gem_destroy+0x2c/0x38
> [ 39.316068] msm_gem_free_work+0xa0/0x1b0
> [ 39.320611] process_one_work+0x2a8/0x7ec
> [ 39.325149] worker_thread+0x4c/0x554
> [ 39.329339] kthread+0x140/0x154
> [ 39.333095] ret_from_fork+0x14/0x20
> [ 39.337196] 0x0
> [ 39.339560]
> [ 39.339560] other info that might help us debug this:
> [ 39.339560]
> [ 39.347568] Possible unsafe locking scenario:
> [ 39.347568]
> [ 39.353491] CPU0 CPU1
> [ 39.358025] ---- ----
> [ 39.362557] lock(&msm_obj->lock);
> [ 39.366061] lock(reservation_ww_class_mutex);
> [ 39.373123] lock(&msm_obj->lock);
> [ 39.379143] lock(reservation_ww_class_mutex);
> [ 39.383689]
> [ 39.383689] *** DEADLOCK ***
> [ 39.383689]
> [ 39.389614] 4 locks held by kworker/u2:1/68:
> [ 39.393888] #0: c24138a8 ((wq_completion)msm){+.+.}-{0:0}, at:
> process_one_work+0x1f4/0x7ec
> [ 39.402362] #1: c26a1f20
> ((work_completion)(&priv->free_work)){+.+.}-{0:0}, at:
> process_one_work+0x1f4/0x7ec
> [ 39.412312] #2: c24188a4 (&dev->struct_mutex){+.+.}-{3:3}, at:
> msm_gem_free_work+0x68/0x1b0
> [ 39.420784] #3: c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at:
> msm_gem_free_work+0x118/0x1b0
> [ 39.428996]
> [ 39.428996] stack backtrace:
> [ 39.433363] CPU: 0 PID: 68 Comm: kworker/u2:1 Not tainted
> 5.10.31+g7ae1de1d2bd3 #1
> [ 39.440937] Hardware name: Freescale i.MX53 (Device Tree Support)
> [ 39.447044] Workqueue: msm msm_gem_free_work
> [ 39.451346] [<c0111258>] (unwind_backtrace) from [<c010ba00>]
> (show_stack+0x10/0x14)
> [ 39.459107] [<c010ba00>] (show_stack) from [<c0e2b560>]
> (dump_stack+0xdc/0x104)
> [ 39.466430] [<c0e2b560>] (dump_stack) from [<c01838d8>]
> (check_noncircular+0xf0/0x160)
> [ 39.474358] [<c01838d8>] (check_noncircular) from [<c0187060>]
> (__lock_acquire+0x1628/0x2e04)
> [ 39.482893] [<c0187060>] (__lock_acquire) from [<c01892e8>]
> (lock_acquire+0x134/0x560)
> [ 39.490821] [<c01892e8>] (lock_acquire) from [<c0e3da60>]
> (__ww_mutex_lock.constprop.0+0xbc/0x1124)
> [ 39.499877] [<c0e3da60>] (__ww_mutex_lock.constprop.0) from
> [<c0e3ebe4>] (ww_mutex_lock+0x68/0xb4)
> [ 39.508844] [<c0e3ebe4>] (ww_mutex_lock) from [<c080e6a8>]
> (dma_buf_detach+0xd0/0x1e0)
> [ 39.516771] [<c080e6a8>] (dma_buf_detach) from [<c0701d04>]
> (drm_prime_gem_destroy+0x2c/0x38)
> [ 39.525305] [<c0701d04>] (drm_prime_gem_destroy) from [<c077d334>]
> (msm_gem_free_work+0xa0/0x1b0)
> [ 39.534188] [<c077d334>] (msm_gem_free_work) from [<c0147588>]
> (process_one_work+0x2a8/0x7ec)
> [ 39.542725] [<c0147588>] (process_one_work) from [<c0147b18>]
> (worker_thread+0x4c/0x554)
> [ 39.550825] [<c0147b18>] (worker_thread) from [<c014ebb0>]
> (kthread+0x140/0x154)
> [ 39.558229] [<c014ebb0>] (kthread) from [<c0100134>]
> (ret_from_fork+0x14/0x20)
> [ 39.565457] Exception stack(0xc26a1fb0 to 0xc26a1ff8)
> [ 39.570517] 1fa0: 00000000
> 00000000 00000000 00000000
> [ 39.578703] 1fc0: 00000000 00000000 00000000 00000000 00000000
> 00000000 00000000 00000000
> [ 39.586890] 1fe0: 00000000 00000000 00000000 00000000 00000013 00000000
>
> This looks like a regression as it used to run fine on 5.4 Linux kernel.
>
> Any idea of patch reference to test is welcome.
>
> --
> Otavio Salvador O.S. Systems
> http://www.ossystems.com.br http://code.ossystems.com.br
> Mobile: +55 (53) 9 9981-7854 Mobile: +1 (347) 903-9750
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: i.MX53 error during GPU use
2021-04-23 14:35 ` Rob Clark
@ 2021-05-04 17:11 ` Otavio Salvador
0 siblings, 0 replies; 3+ messages in thread
From: Otavio Salvador @ 2021-05-04 17:11 UTC (permalink / raw)
To: Rob Clark
Cc: Sean Paul, David Airlie, Daniel Vetter, linux-arm-msm,
DRI mailing list, freedreno, Fabio Estevam, vinicius.aquino,
luan.rafael
Hello Rob,
Em sex., 23 de abr. de 2021 às 11:35, Rob Clark <robdclark@gmail.com> escreveu:
> On Fri, Apr 23, 2021 at 4:58 AM Otavio Salvador
> <otavio.salvador@ossystems.com.br> wrote:
> > We found this error when using Freedreno driver on an i.MX53 device
> > with Wayland. Any idea how to fix this?
> >
> > [ 32.414110] [drm:msm_ioctl_gem_submit] *ERROR* invalid cmdstream size: 0
>
> The invalid cmdstream size is some sort of userspace error
>
> > [ 39.177075]
> > [ 39.178617] ======================================================
> > [ 39.184804] WARNING: possible circular locking dependency detected
> > [ 39.190997] 5.10.31+g7ae1de1d2bd3 #1 Not tainted
> > [ 39.195619] ------------------------------------------------------
>
> But possibly it is triggering the lockdep anger? It looks like the
> gem locking re-work landed in v5.11.. any chance you can try a newer
> kernel?
Sure; we tried the 5.12.1 Linux kernel and it "worked". We have used
following versions:
- Linux kernel 5.12.1
- mesa 21.0.3
- libdrm 2.4.105
It improved a lot and it opens. We now have some rendering issues:
https://photos.app.goo.gl/fBKoe5C8tsq4xU556
and an error in serial console:
[ 262.319890] schedule_timeout: wrong timeout value bf946f6e
[ 262.325845] CPU: 0 PID: 216 Comm: eadedCompositor Not tainted
5.12.1+g1a5fea11bc2f #1
[ 262.333727] Hardware name: Freescale i.MX53 (Device Tree Support)
[ 262.339854] [<c01114e8>] (unwind_backtrace) from [<c010bb60>]
(show_stack+0x10/0x14)
[ 262.347659] [<c010bb60>] (show_stack) from [<c0e446b0>]
(dump_stack+0xdc/0x104)
[ 262.355007] [<c0e446b0>] (dump_stack) from [<c0e5928c>]
(schedule_timeout+0xf0/0x128)
[ 262.362875] [<c0e5928c>] (schedule_timeout) from [<c0789260>]
(msm_wait_fence+0x1c0/0x320)
[ 262.371190] [<c0789260>] (msm_wait_fence) from [<c0788510>]
(msm_ioctl_wait_fence+0xa8/0x154)
[ 262.379749] [<c0788510>] (msm_ioctl_wait_fence) from [<c0700064>]
(drm_ioctl+0x1f0/0x3dc)
[ 262.387966] [<c0700064>] (drm_ioctl) from [<c02de8ec>]
(sys_ioctl+0x3cc/0xbac)
[ 262.395226] [<c02de8ec>] (sys_ioctl) from [<c0100080>]
(ret_fast_syscall+0x0/0x2c)
[ 262.402829] Exception stack(0xc315ffa8 to 0xc315fff0)
[ 262.407911] ffa0: 00000000 abc10840 00000010
40206447 abc10840 00000020
[ 262.416118] ffc0: 00000000 abc10840 40206447 00000036 afd32cb0
abc108b8 00000000 abc1087c
[ 262.424320] ffe0: b075aef0 abc10804 b0740214 b40a11fc
Any idea what might be causing it?
--
Otavio Salvador O.S. Systems
http://www.ossystems.com.br http://code.ossystems.com.br
Mobile: +55 (53) 9 9981-7854 Mobile: +1 (347) 903-9750
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2021-05-04 17:11 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-04-23 11:58 i.MX53 error during GPU use Otavio Salvador
2021-04-23 14:35 ` Rob Clark
2021-05-04 17:11 ` Otavio Salvador
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).