From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.2 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 84674C433F5 for ; Tue, 21 Sep 2021 08:59:45 +0000 (UTC) Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 48E4A61131 for ; Tue, 21 Sep 2021 08:59:45 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.4.1 mail.kernel.org 48E4A61131 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=pengutronix.de Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=lists.infradead.org DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=bombadil.20210309; h=Sender: Content-Transfer-Encoding:Content-Type:List-Subscribe:List-Help:List-Post: List-Archive:List-Unsubscribe:List-Id:MIME-Version:Message-ID:Date:Subject:Cc :To:From:Reply-To:Content-ID:Content-Description:Resent-Date:Resent-From: Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:In-Reply-To:References: List-Owner; bh=uCyV4VRdLu9p/N02kfHhUNDnEuvojc4QTKl5QnH2JyI=; b=JyLEBj3gPDjxo9 5GmluKVsKGR1rLLG3OejxINgu1sZ9u+2qEWs8gjsLPbTJcI2XdfmO8m/9yqxhWxpevt7zBhaHVNTe UqlZ7XRVHLnyq4utcep+kT+tUoRFm2q6MYbdwX9dM7kMCTw5zA6rpXMcC85+hZIkSXBNfYyChFlSG 2xwxfBcVjHdF2fWn0c/lwhG4yLb1LCxlwjWiosAYZLZN7bkcxEY65X/dfcGDHXH8UMt+32Ujz3Odz aeTb1curZnhAgrvVrgXRncWJ7wbzl/O76G5IIOS26QVMHDbihb3M2dF6x8yY7xeaJ41WL7QBV/AxJ m3gDYw5Oq4oVs8Cx3Vow==; Received: from localhost ([::1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.94.2 #2 (Red Hat Linux)) id 1mSbaV-003zsD-E2; Tue, 21 Sep 2021 08:57:15 +0000 Received: from metis.ext.pengutronix.de ([2001:67c:670:201:290:27ff:fe1d:cc33]) by bombadil.infradead.org with esmtps (Exim 4.94.2 #2 (Red Hat Linux)) id 1mSbaP-003zrY-6q for linux-arm-kernel@lists.infradead.org; Tue, 21 Sep 2021 08:57:12 +0000 Received: from ptx.hi.pengutronix.de ([2001:67c:670:100:1d::c0] helo=igor.pengutronix.de) by metis.ext.pengutronix.de with esmtps (TLS1.3:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.92) (envelope-from ) id 1mSbaK-0003us-Cx; Tue, 21 Sep 2021 10:57:04 +0200 User-agent: mu4e 1.4.13; emacs 28.0.50 From: Steffen Trumtrar To: Valentin Schneider , Marc Zyngier Cc: Andrew Lunn , Gregory Clement , Sebastion Hesselbarth , linux-arm-kernel@lists.infradead.org Subject: [BUG] irqchip: armada-370-xp: workqueue lockup Date: Tue, 21 Sep 2021 10:40:59 +0200 Message-ID: <87tuiexq5f.fsf@pengutronix.de> MIME-Version: 1.0 X-SA-Exim-Connect-IP: 2001:67c:670:100:1d::c0 X-SA-Exim-Mail-From: s.trumtrar@pengutronix.de X-SA-Exim-Scanned: No (on metis.ext.pengutronix.de); SAEximRunCond expanded to false X-PTX-Original-Recipient: linux-arm-kernel@lists.infradead.org X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20210921_015709_644844_4B94B129 X-CRM114-Status: GOOD ( 14.22 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.34 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org Hi, I noticed that after the patch e52e73b7e9f7d08b8c2ef6fb1657105093e22a03 From: Valentin Schneider Date: Mon, 9 Nov 2020 09:41:18 +0000 Subject: [PATCH] irqchip/armada-370-xp: Make IPIs use handle_percpu_devid_irq() As done for the Arm GIC irqchips, move IPIs to handle_percpu_devid_irq() as handle_percpu_devid_fasteoi_ipi() isn't actually required. Signed-off-by: Valentin Schneider Signed-off-by: Marc Zyngier Link: https://lore.kernel.org/r/20201109094121.29975-3-valentin.schneider@arm.com --- drivers/irqchip/irq-armada-370-xp.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/irqchip/irq-armada-370-xp.c b/drivers/irqchip/irq-armada-370-xp.c index d7eb2e93db8f..32938dfc0e46 100644 --- a/drivers/irqchip/irq-armada-370-xp.c +++ b/drivers/irqchip/irq-armada-370-xp.c @@ -382,7 +382,7 @@ static int armada_370_xp_ipi_alloc(struct irq_domain *d, irq_set_percpu_devid(virq + i); irq_domain_set_info(d, virq + i, i, &ipi_irqchip, d->host_data, - handle_percpu_devid_fasteoi_ipi, + handle_percpu_devid_irq, NULL, NULL); } I get workqueue lockups on my Armada-XP based board. When I run the following test on v5.15-rc2 stress-ng --cpu 8 --io 4 --vm 2 --vm-bytes 128M --fork 4 --timeout 120s I get a backtrace like this: stress-ng: info: [7740] dispatching hogs: 8 cpu, 4 io, 2 vm, 4 fork [ 1670.169087] rcu: INFO: rcu_preempt detected stalls on CPUs/tasks: [ 1670.169102] (detected by 0, t=5252 jiffies, g=50257, q=3369) [ 1670.169112] rcu: All QSes seen, last rcu_preempt kthread activity 5252 (342543-337291), jiffies_till_next_fqs=1, root ->qsmask 0x0 [ 1670.169121] rcu: rcu_preempt kthread timer wakeup didn't happen for 5251 jiffies! g50257 f0x2 RCU_GP_WAIT_FQS(5) ->state=0x200 [ 1670.169128] rcu: Possible timer handling issue on cpu=1 timer-softirq=20398 [ 1670.169132] rcu: rcu_preempt kthread starved for 5252 jiffies! g50257 f0x2 RCU_GP_WAIT_FQS(5) ->state=0x200 ->cpu=1 [ 1670.169140] rcu: Unless rcu_preempt kthread gets sufficient CPU time, OOM is now expected behavior. [ 1670.169143] rcu: RCU grace-period kthread stack dump: [ 1670.169146] task:rcu_preempt state:R stack: 0 pid: 13 ppid: 2 flags:0x00000000 [ 1670.169157] Backtrace: [ 1670.169163] [] (__schedule) from [] (schedule+0x64/0x110) [ 1670.169185] r10:00000001 r9:c190e000 r8:c137b690 r7:c137b69c r6:c190fed4 r5:c190e000 [ 1670.169189] r4:c197c880 [ 1670.169192] [] (schedule) from [] (schedule_timeout+0xa8/0x1c0) [ 1670.169206] r5:c1303d00 r4:0005258c [ 1670.169209] [] (schedule_timeout) from [] (rcu_gp_fqs_loop+0x120/0x3ac) [ 1670.169227] r7:c137b69c r6:c1303d00 r5:c137b4c0 r4:00000000 [ 1670.169230] [] (rcu_gp_fqs_loop) from [] (rcu_gp_kthread+0xfc/0x1b0) [ 1670.169247] r10:c190ff5c r9:c1303d00 r8:c137b4c0 r7:c190e000 r6:c137b69e r5:c137b690 [ 1670.169251] r4:c137b69c [ 1670.169253] [] (rcu_gp_kthread) from [] (kthread+0x16c/0x1a0) [ 1670.169268] r7:00000000 [ 1670.169271] [] (kthread) from [] (ret_from_fork+0x14/0x38) [ 1670.169282] Exception stack(0xc190ffb0 to 0xc190fff8) [ 1670.169288] ffa0: ???????? ???????? ???????? ???????? [ 1670.169293] ffc0: ???????? ???????? ???????? ???????? ???????? ???????? ???????? ???????? [ 1670.169297] ffe0: ???????? ???????? ???????? ???????? ???????? ???????? [ 1670.169305] r10:00000000 r9:00000000 r8:00000000 r7:00000000 r6:00000000 r5:c01539a8 [ 1670.169310] r4:c19320c0 r3:00000000 [ 1670.169313] rcu: Stack dump where RCU GP kthread last ran: [ 1670.169316] Sending NMI from CPU 0 to CPUs 1: [ 1670.169327] NMI backtrace for cpu 1 [ 1670.169335] CPU: 1 PID: 7764 Comm: stress-ng-cpu Tainted: G W 5.15.0-rc2+ #5 [ 1670.169343] Hardware name: Marvell Armada 370/XP (Device Tree) [ 1670.169346] PC is at 0x4bde7a [ 1670.169354] LR is at 0x4bdf21 [ 1670.169359] pc : [<004bde7a>] lr : [<004bdf21>] psr: 20030030 [ 1670.169363] sp : beb8270c ip : 00004650 fp : beb8289c [ 1670.169367] r10: 00e5e800 r9 : 00514760 r8 : 0000036b [ 1670.169371] r7 : beb828a8 r6 : 000001f7 r5 : 000001fd r4 : 000bacd7 [ 1670.169375] r3 : 004bde30 r2 : 0000000b r1 : 000001fd r0 : 0001bbd7 [ 1670.169380] Flags: nzCv IRQs on FIQs on Mode USER_32 ISA Thumb Segment user [ 1670.169386] Control: 10c5387d Table: 0334806a DAC: 00000055 [ 1670.169389] CPU: 1 PID: 7764 Comm: stress-ng-cpu Tainted: G W 5.15.0-rc2+ #5 [ 1670.169395] Hardware name: Marvell Armada 370/XP (Device Tree) [ 1670.169398] Backtrace: [ 1670.169402] [] (dump_backtrace) from [] (show_stack+0x20/0x24) [ 1670.169418] r7:c18db400 r6:c7875fb0 r5:60030193 r4:c1099c7c [ 1670.169421] [] (show_stack) from [] (dump_stack_lvl+0x48/0x54) [ 1670.169433] [] (dump_stack_lvl) from [] (dump_stack+0x18/0x1c) [ 1670.169445] r5:00000001 r4:20030193 [ 1670.169447] [] (dump_stack) from [] (show_regs+0x1c/0x20) [ 1670.169461] [] (show_regs) from [] (nmi_cpu_backtrace+0xc0/0x10c) [ 1670.169474] [] (nmi_cpu_backtrace) from [] (do_handle_IPI+0x54/0x3b8) [ 1670.169489] r7:c18db400 r6:00000017 r5:00000001 r4:00000007 [ 1670.169491] [] (do_handle_IPI) from [] (ipi_handler+0x28/0x30) [ 1670.169505] r10:c7875f58 r9:c7875fb0 r8:c7875f30 r7:c18db400 r6:00000017 r5:c13ecadc [ 1670.169509] r4:c18d9300 r3:00000010 [ 1670.169511] [] (ipi_handler) from [] (handle_percpu_devid_irq+0xb4/0x288) [ 1670.169525] [] (handle_percpu_devid_irq) from [] (handle_domain_irq+0x8c/0xc0) [ 1670.169539] r9:c7875fb0 r8:00000007 r7:00000000 r6:c1863d80 r5:00000000 r4:c12781e0 [ 1670.169542] [] (handle_domain_irq) from [] (armada_370_xp_handle_irq+0xdc/0x124) [ 1670.169556] r10:00e5e800 r9:00514760 r8:10c5387d r7:c147d604 r6:c7875fb0 r5:000003fe [ 1670.169560] r4:00000007 r3:00000007 [ 1670.169562] [] (armada_370_xp_handle_irq) from [] (__irq_usr+0x58/0x80) [ 1670.169571] Exception stack(0xc7875fb0 to 0xc7875ff8) [ 1670.169576] 5fa0: ???????? ???????? ???????? ???????? [ 1670.169580] 5fc0: ???????? ???????? ???????? ???????? ???????? ???????? ???????? ???????? [ 1670.169584] 5fe0: ???????? ???????? ???????? ???????? ???????? ???????? [ 1670.169590] r7:10c5387d r6:ffffffff r5:20030030 r4:004bde7a [ 1690.589098] BUG: workqueue lockup - pool cpus=1 node=0 flags=0x0 nice=0 stuck for 38s! [ 1690.589133] Showing busy workqueues and worker pools: [ 1690.589138] workqueue events_unbound: flags=0x2 [ 1690.589142] pwq 4: cpus=0-1 flags=0x4 nice=0 active=3/512 refcnt=5 [ 1690.589157] in-flight: 7:call_usermodehelper_exec_work [ 1690.589177] pending: flush_memcg_stats_work, flush_memcg_stats_dwork [ 1690.589198] workqueue events_power_efficient: flags=0x80 [ 1690.589203] pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=5/256 refcnt=6 [ 1690.589218] in-flight: 53:fb_flashcursor fb_flashcursor [ 1690.589236] pending: neigh_periodic_work, neigh_periodic_work, do_cache_clean [ 1690.589265] workqueue mm_percpu_wq: flags=0x8 [ 1690.589269] pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 [ 1690.589284] pending: vmstat_update [ 1690.589301] workqueue edac-poller: flags=0xa000a [ 1690.589305] pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=4 [ 1690.589318] pending: edac_mc_workq_function [ 1690.589331] inactive: edac_device_workq_function [ 1690.589346] pool 2: cpus=1 node=0 flags=0x0 nice=0 hung=38s workers=3 idle: 7621 6478 [ 1690.589370] pool 4: cpus=0-1 flags=0x4 nice=0 hung=41s workers=3 idle: 6967 5672 [ 1721.313097] BUG: workqueue lockup - pool cpus=1 node=0 flags=0x0 nice=0 stuck for 69s! [ 1721.313136] BUG: workqueue lockup - pool cpus=0-1 flags=0x4 nice=0 stuck for 72s! [ 1721.313149] Showing busy workqueues and worker pools: [ 1721.313154] workqueue events_unbound: flags=0x2 [ 1721.313158] pwq 4: cpus=0-1 flags=0x4 nice=0 active=3/512 refcnt=5 [ 1721.313173] in-flight: 7:call_usermodehelper_exec_work [ 1721.313193] pending: flush_memcg_stats_work, flush_memcg_stats_dwork [ 1721.313213] workqueue events_power_efficient: flags=0x80 [ 1721.313218] pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=5/256 refcnt=6 [ 1721.313234] in-flight: 53:fb_flashcursor fb_flashcursor [ 1721.313251] pending: neigh_periodic_work, neigh_periodic_work, do_cache_clean [ 1721.313282] workqueue mm_percpu_wq: flags=0x8 [ 1721.313285] pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 [ 1721.313301] pending: vmstat_update [ 1721.313319] workqueue edac-poller: flags=0xa000a [ 1721.313323] pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=4 [ 1721.313336] pending: edac_mc_workq_function [ 1721.313349] inactive: edac_device_workq_function [ 1721.313366] pool 2: cpus=1 node=0 flags=0x0 nice=0 hung=69s workers=3 idle: 7621 6478 [ 1721.313390] pool 4: cpus=0-1 flags=0x4 nice=0 hung=72s workers=3 idle: 6967 5672 [ 1733.189086] rcu: INFO: rcu_preempt detected stalls on CPUs/tasks: [ 1733.189101] (detected by 0, t=21007 jiffies, g=50257, q=13112) [ 1733.189111] rcu: All QSes seen, last rcu_preempt kthread activity 21007 (358298-337291), jiffies_till_next_fqs=1, root ->qsmask 0x0 [ 1733.189119] rcu: rcu_preempt kthread timer wakeup didn't happen for 21006 jiffies! g50257 f0x2 RCU_GP_WAIT_FQS(5) ->state=0x200 [ 1733.189126] rcu: Possible timer handling issue on cpu=1 timer-softirq=20834 [ 1733.189131] rcu: rcu_preempt kthread starved for 21007 jiffies! g50257 f0x2 RCU_GP_WAIT_FQS(5) ->state=0x200 ->cpu=1 [ 1733.189138] rcu: Unless rcu_preempt kthread gets sufficient CPU time, OOM is now expected behavior. [ 1733.189141] rcu: RCU grace-period kthread stack dump: [ 1733.189144] task:rcu_preempt state:R stack: 0 pid: 13 ppid: 2 flags:0x00000000 [ 1733.189156] Backtrace: [ 1733.189162] [] (__schedule) from [] (schedule+0x64/0x110) [ 1733.189184] r10:00000001 r9:c190e000 r8:c137b690 r7:c137b69c r6:c190fed4 r5:c190e000 [ 1733.189188] r4:c197c880 [ 1733.189191] [] (schedule) from [] (schedule_timeout+0xa8/0x1c0) [ 1733.189205] r5:c1303d00 r4:0005258c [ 1733.189208] [] (schedule_timeout) from [] (rcu_gp_fqs_loop+0x120/0x3ac) [ 1733.189226] r7:c137b69c r6:c1303d00 r5:c137b4c0 r4:00000000 [ 1733.189229] [] (rcu_gp_fqs_loop) from [] (rcu_gp_kthread+0xfc/0x1b0) [ 1733.189246] r10:c190ff5c r9:c1303d00 r8:c137b4c0 r7:c190e000 r6:c137b69e r5:c137b690 [ 1733.189249] r4:c137b69c [ 1733.189252] [] (rcu_gp_kthread) from [] (kthread+0x16c/0x1a0) [ 1733.189267] r7:00000000 [ 1733.189270] [] (kthread) from [] (ret_from_fork+0x14/0x38) [ 1733.189281] Exception stack(0xc190ffb0 to 0xc190fff8) [ 1733.189287] ffa0: ???????? ???????? ???????? ???????? [ 1733.189292] ffc0: ???????? ???????? ???????? ???????? ???????? ???????? ???????? ???????? [ 1733.189297] ffe0: ???????? ???????? ???????? ???????? ???????? ???????? [ 1733.189304] r10:00000000 r9:00000000 r8:00000000 r7:00000000 r6:00000000 r5:c01539a8 [ 1733.189309] r4:c19320c0 r3:00000000 [ 1733.189312] rcu: Stack dump where RCU GP kthread last ran: [ 1733.189315] Sending NMI from CPU 0 to CPUs 1: [ 1733.189327] NMI backtrace for cpu 1 [ 1733.189335] CPU: 1 PID: 7755 Comm: stress-ng-cpu Tainted: G W 5.15.0-rc2+ #5 [ 1733.189343] Hardware name: Marvell Armada 370/XP (Device Tree) [ 1733.189346] PC is at 0x4bdee0 [ 1733.189354] LR is at 0x4bdf21 [ 1733.189358] pc : [<004bdee0>] lr : [<004bdf21>] psr: 20030030 [ 1733.189363] sp : beb8270c ip : 00004650 fp : beb8289c [ 1733.189367] r10: 00e5e800 r9 : 00514760 r8 : 00000358 [ 1733.189370] r7 : beb828a8 r6 : 00000047 r5 : 0000004d r4 : 000b2ab7 [ 1733.189375] r3 : 004bde10 r2 : 00001217 r1 : 0000004f r0 : 00000085 [ 1733.189379] Flags: nzCv IRQs on FIQs on Mode USER_32 ISA Thumb Segment user [ 1733.189385] Control: 10c5387d Table: 0734006a DAC: 00000055 [ 1733.189389] CPU: 1 PID: 7755 Comm: stress-ng-cpu Tainted: G W 5.15.0-rc2+ #5 [ 1733.189395] Hardware name: Marvell Armada 370/XP (Device Tree) [ 1733.189397] Backtrace: [ 1733.189402] [] (dump_backtrace) from [] (show_stack+0x20/0x24) [ 1733.189417] r7:c18db400 r6:c7375fb0 r5:60030193 r4:c1099c7c [ 1733.189420] [] (show_stack) from [] (dump_stack_lvl+0x48/0x54) [ 1733.189432] [] (dump_stack_lvl) from [] (dump_stack+0x18/0x1c) [ 1733.189444] r5:00000001 r4:20030193 [ 1733.189446] [] (dump_stack) from [] (show_regs+0x1c/0x20) [ 1733.189460] [] (show_regs) from [] (nmi_cpu_backtrace+0xc0/0x10c) [ 1733.189473] [] (nmi_cpu_backtrace) from [] (do_handle_IPI+0x54/0x3b8) [ 1733.189488] r7:c18db400 r6:00000017 r5:00000001 r4:00000007 [ 1733.189490] [] (do_handle_IPI) from [] (ipi_handler+0x28/0x30) [ 1733.189504] r10:c7375f58 r9:c7375fb0 r8:c7375f30 r7:c18db400 r6:00000017 r5:c13ecadc [ 1733.189508] r4:c18d9300 r3:00000010 [ 1733.189510] [] (ipi_handler) from [] (handle_percpu_devid_irq+0xb4/0x288) [ 1733.189523] [] (handle_percpu_devid_irq) from [] (handle_domain_irq+0x8c/0xc0) [ 1733.189538] r9:c7375fb0 r8:00000007 r7:00000000 r6:c1863d80 r5:00000000 r4:c12781e0 [ 1733.189540] [] (handle_domain_irq) from [] (armada_370_xp_handle_irq+0xdc/0x124) [ 1733.189555] r10:00e5e800 r9:00514760 r8:10c5387d r7:c147d604 r6:c7375fb0 r5:000003fe [ 1733.189559] r4:00000007 r3:00000007 [ 1733.189561] [] (armada_370_xp_handle_irq) from [] (__irq_usr+0x58/0x80) [ 1733.189570] Exception stack(0xc7375fb0 to 0xc7375ff8) [ 1733.189575] 5fa0: ???????? ???????? ???????? ???????? [ 1733.189579] 5fc0: ???????? ???????? ???????? ???????? ???????? ???????? ???????? ???????? [ 1733.189583] 5fe0: ???????? ???????? ???????? ???????? ???????? ???????? [ 1733.189589] r7:10c5387d r6:ffffffff r5:20030030 r4:004bdee0 [ 1752.029102] BUG: workqueue lockup - pool cpus=1 node=0 flags=0x0 nice=0 stuck for 100s! [ 1752.029137] Showing busy workqueues and worker pools: [ 1752.029141] workqueue events_unbound: flags=0x2 [ 1752.029146] pwq 4: cpus=0-1 flags=0x4 nice=0 active=3/512 refcnt=5 [ 1752.029161] in-flight: 7:call_usermodehelper_exec_work [ 1752.029180] pending: flush_memcg_stats_work, flush_memcg_stats_dwork [ 1752.029200] workqueue events_power_efficient: flags=0x80 [ 1752.029205] pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=5/256 refcnt=6 [ 1752.029221] in-flight: 53:fb_flashcursor fb_flashcursor [ 1752.029239] pending: neigh_periodic_work, neigh_periodic_work, do_cache_clean [ 1752.029269] workqueue mm_percpu_wq: flags=0x8 [ 1752.029272] pwq 2: cpus=1 node=0 flags=0x0 nice=0 active=1/256 refcnt=2 [ 1752.029288] pending: vmstat_update [ 1752.029306] workqueue edac-poller: flags=0xa000a [ 1752.029310] pwq 4: cpus=0-1 flags=0x4 nice=0 active=1/1 refcnt=4 [ 1752.029323] pending: edac_mc_workq_function [ 1752.029337] inactive: edac_device_workq_function [ 1752.029353] pool 2: cpus=1 node=0 flags=0x0 nice=0 hung=100s workers=3 idle: 7621 6478 [ 1752.029378] pool 4: cpus=0-1 flags=0x4 nice=0 hung=102s workers=3 idle: 6967 5672 stress-ng: info: [7740] successful run completed in 125.31s (2 mins, 5.31 secs) Earlier kernels (i.e v5.13.9) completely froze the machine resulting in the watchdog triggering and rebooting the machine. So, $something was already fixed here. Bisecting leads to the mentioned commit, reverting of the commit results in a BUG-less run of the stress-ng test. Any idea what might cause this and how to fix it? Best regards, Steffen Trumtrar -- Pengutronix e.K. | Dipl.-Inform. Steffen Trumtrar | Steuerwalder Str. 21 | https://www.pengutronix.de/ | 31137 Hildesheim, Germany | Phone: +49-5121-206917-0 | Amtsgericht Hildesheim, HRA 2686| Fax: +49-5121-206917-5555 | _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel