From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.7 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE, SPF_PASS,URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 105DCC433ED for ; Fri, 23 Apr 2021 11:58:55 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id B6FFD6141C for ; Fri, 23 Apr 2021 11:58:54 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234439AbhDWL73 (ORCPT ); Fri, 23 Apr 2021 07:59:29 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:37144 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234417AbhDWL73 (ORCPT ); Fri, 23 Apr 2021 07:59:29 -0400 Received: from mail-qv1-xf31.google.com (mail-qv1-xf31.google.com [IPv6:2607:f8b0:4864:20::f31]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 755B1C061574 for ; Fri, 23 Apr 2021 04:58:51 -0700 (PDT) Received: by mail-qv1-xf31.google.com with SMTP id ef17so17934013qvb.0 for ; Fri, 23 Apr 2021 04:58:51 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ossystems-com-br.20150623.gappssmtp.com; s=20150623; h=from:mime-version:date:message-id:subject:to; bh=jr8jZmW3hWeOFIHhOThCMqo1kOiVcgfycAvoHOOzfOs=; b=jd2ZSHi+A82UPBSdJfxVNxKNDia794DDn9zlNQU4hMLaZTLJtq8kDI7dtMRgqBYM4v sFMWrsBHoMwWEekSxMNVKC0ZGrd1RLEMp3tNmq6eoRQFrhY35ZenFAr0F12sqngope9P Ox0oOl7Mw1AClWBQfR2fhdjwEGTejjbmULWAPVxomQ/QW5YRNIRqRmhEOKsOyt8F9yQq 7arRaaJL8heG7ppTQgH24gTUvJlHZenYrA8zFou0xpnZkHi2nHFUlqTSRdqemKzQ2Cfy 2c0g2HrEuzGUlQAYRkUrEfJ5E5vtDvBvf+QjGl3tbl/v47JesPHeeGiklCM3u9ZTIzKK DO5w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:mime-version:date:message-id:subject:to; bh=jr8jZmW3hWeOFIHhOThCMqo1kOiVcgfycAvoHOOzfOs=; b=RxqeFH3a/k1j7kJ/XhMlp68VVEdqqlEKRK56t29KuRbW0VvcrxgS80T4cEdQIJoa2S 8+BbWdJib2rK53prh7j/98JeBM8J64gdm8HCkelQmLufePuCzDBrs1QNJ6/Pc4nwCFir hPwSfJfWJQ6SvJZFn/rHQZdBbZeNHW9gkAkZM2c+IXkDdkRquLXsZqNRZbAdtVy02veI zdmUEMI+lNncD3b2BOZGtPHka5MuWmoWzi+wAKzpmidv8QWEVKm1k9DbUNtfwthPphhN 3LdXifSXYwVQcuz20KyPkfgORHdLvAlaiOcRvFqgJ2xNowBlX4Bs9nza2mUJZKoBvCaU GiyA== X-Gm-Message-State: AOAM533fDlC9zvhmV1dO28ZGTBRDEU7J8RF56zH9FMcFKKLNhzBV1MiV gqpcDwgd5oRNBwjVJTRAZsVoMqejC6Qxj5kk X-Google-Smtp-Source: ABdhPJyTix0AstQK1dA0oaq1PZ/JllvL5bpm8TyZMk4+k4luCH4YPm0l9LeJLZQY6pQr5wI7FPAdWA== X-Received: by 2002:ad4:5742:: with SMTP id q2mr3973415qvx.11.1619179130438; Fri, 23 Apr 2021 04:58:50 -0700 (PDT) Received: from mail-qv1-f52.google.com (mail-qv1-f52.google.com. [209.85.219.52]) by smtp.gmail.com with ESMTPSA id e13sm4447762qtm.35.2021.04.23.04.58.49 for (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Fri, 23 Apr 2021 04:58:50 -0700 (PDT) From: Otavio Salvador X-Google-Original-From: Otavio Salvador Received: by mail-qv1-f52.google.com with SMTP id bs7so23094891qvb.12 for ; Fri, 23 Apr 2021 04:58:49 -0700 (PDT) X-Received: by 2002:a0c:c352:: with SMTP id j18mr4237446qvi.12.1619179129716; Fri, 23 Apr 2021 04:58:49 -0700 (PDT) MIME-Version: 1.0 Date: Fri, 23 Apr 2021 08:58:38 -0300 X-Gmail-Original-Message-ID: Message-ID: Subject: i.MX53 error during GPU use To: Rob Clark , Sean Paul , David Airlie , Daniel Vetter , linux-arm-msm@vger.kernel.org, DRI mailing list , freedreno@lists.freedesktop.org Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-arm-msm@vger.kernel.org Hello all, We found this error when using Freedreno driver on an i.MX53 device with Wayland. Any idea how to fix this? [ 32.414110] [drm:msm_ioctl_gem_submit] *ERROR* invalid cmdstream size: 0 [ 39.177075] [ 39.178617] ====================================================== [ 39.184804] WARNING: possible circular locking dependency detected [ 39.190997] 5.10.31+g7ae1de1d2bd3 #1 Not tainted [ 39.195619] ------------------------------------------------------ [ 39.201805] kworker/u2:1/68 is trying to acquire lock: [ 39.206953] c2b952fc (reservation_ww_class_mutex){+.+.}-{3:3}, at: dma_buf_detach+0xd0/0x1e0 [ 39.215448] [ 39.215448] but task is already holding lock: [ 39.221286] c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at: msm_gem_free_work+0x118/0x1b0 [ 39.229062] [ 39.229062] which lock already depends on the new lock. [ 39.229062] [ 39.237243] [ 39.237243] the existing dependency chain (in reverse order) is: [ 39.244729] [ 39.244729] -> #1 (&msm_obj->lock){+.+.}-{3:3}: [ 39.250769] mutex_lock_nested+0x1c/0x24 [ 39.255220] msm_gem_get_and_pin_iova_range+0x38/0x170 [ 39.260885] msm_gem_get_and_pin_iova+0x24/0x2c [ 39.265946] msm_ioctl_gem_submit+0x12c0/0x1964 [ 39.271008] drm_ioctl+0x1f0/0x3e0 [ 39.274940] sys_ioctl+0x3bc/0xbe0 [ 39.278873] ret_fast_syscall+0x0/0x28 [ 39.283150] 0xbed4deac [ 39.286122] [ 39.286122] -> #0 (reservation_ww_class_mutex){+.+.}-{3:3}: [ 39.293208] lock_acquire+0x134/0x560 [ 39.297398] __ww_mutex_lock.constprop.0+0xbc/0x1124 [ 39.302889] ww_mutex_lock+0x68/0xb4 [ 39.306993] dma_buf_detach+0xd0/0x1e0 [ 39.311271] drm_prime_gem_destroy+0x2c/0x38 [ 39.316068] msm_gem_free_work+0xa0/0x1b0 [ 39.320611] process_one_work+0x2a8/0x7ec [ 39.325149] worker_thread+0x4c/0x554 [ 39.329339] kthread+0x140/0x154 [ 39.333095] ret_from_fork+0x14/0x20 [ 39.337196] 0x0 [ 39.339560] [ 39.339560] other info that might help us debug this: [ 39.339560] [ 39.347568] Possible unsafe locking scenario: [ 39.347568] [ 39.353491] CPU0 CPU1 [ 39.358025] ---- ---- [ 39.362557] lock(&msm_obj->lock); [ 39.366061] lock(reservation_ww_class_mutex); [ 39.373123] lock(&msm_obj->lock); [ 39.379143] lock(reservation_ww_class_mutex); [ 39.383689] [ 39.383689] *** DEADLOCK *** [ 39.383689] [ 39.389614] 4 locks held by kworker/u2:1/68: [ 39.393888] #0: c24138a8 ((wq_completion)msm){+.+.}-{0:0}, at: process_one_work+0x1f4/0x7ec [ 39.402362] #1: c26a1f20 ((work_completion)(&priv->free_work)){+.+.}-{0:0}, at: process_one_work+0x1f4/0x7ec [ 39.412312] #2: c24188a4 (&dev->struct_mutex){+.+.}-{3:3}, at: msm_gem_free_work+0x68/0x1b0 [ 39.420784] #3: c2b957b0 (&msm_obj->lock){+.+.}-{3:3}, at: msm_gem_free_work+0x118/0x1b0 [ 39.428996] [ 39.428996] stack backtrace: [ 39.433363] CPU: 0 PID: 68 Comm: kworker/u2:1 Not tainted 5.10.31+g7ae1de1d2bd3 #1 [ 39.440937] Hardware name: Freescale i.MX53 (Device Tree Support) [ 39.447044] Workqueue: msm msm_gem_free_work [ 39.451346] [] (unwind_backtrace) from [] (show_stack+0x10/0x14) [ 39.459107] [] (show_stack) from [] (dump_stack+0xdc/0x104) [ 39.466430] [] (dump_stack) from [] (check_noncircular+0xf0/0x160) [ 39.474358] [] (check_noncircular) from [] (__lock_acquire+0x1628/0x2e04) [ 39.482893] [] (__lock_acquire) from [] (lock_acquire+0x134/0x560) [ 39.490821] [] (lock_acquire) from [] (__ww_mutex_lock.constprop.0+0xbc/0x1124) [ 39.499877] [] (__ww_mutex_lock.constprop.0) from [] (ww_mutex_lock+0x68/0xb4) [ 39.508844] [] (ww_mutex_lock) from [] (dma_buf_detach+0xd0/0x1e0) [ 39.516771] [] (dma_buf_detach) from [] (drm_prime_gem_destroy+0x2c/0x38) [ 39.525305] [] (drm_prime_gem_destroy) from [] (msm_gem_free_work+0xa0/0x1b0) [ 39.534188] [] (msm_gem_free_work) from [] (process_one_work+0x2a8/0x7ec) [ 39.542725] [] (process_one_work) from [] (worker_thread+0x4c/0x554) [ 39.550825] [] (worker_thread) from [] (kthread+0x140/0x154) [ 39.558229] [] (kthread) from [] (ret_from_fork+0x14/0x20) [ 39.565457] Exception stack(0xc26a1fb0 to 0xc26a1ff8) [ 39.570517] 1fa0: 00000000 00000000 00000000 00000000 [ 39.578703] 1fc0: 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000000 [ 39.586890] 1fe0: 00000000 00000000 00000000 00000000 00000013 00000000 This looks like a regression as it used to run fine on 5.4 Linux kernel. Any idea of patch reference to test is welcome. -- Otavio Salvador O.S. Systems http://www.ossystems.com.br http://code.ossystems.com.br Mobile: +55 (53) 9 9981-7854 Mobile: +1 (347) 903-9750