From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-13.1 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED,USER_AGENT_GIT autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 3E68BC433E0 for ; Thu, 23 Jul 2020 01:01:40 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 1689C20771 for ; Thu, 23 Jul 2020 01:01:40 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=chromium.org header.i=@chromium.org header.b="K42b5yLk" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1733289AbgGWBBj (ORCPT ); Wed, 22 Jul 2020 21:01:39 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:37284 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728607AbgGWBBj (ORCPT ); Wed, 22 Jul 2020 21:01:39 -0400 Received: from mail-pg1-x541.google.com (mail-pg1-x541.google.com [IPv6:2607:f8b0:4864:20::541]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1F215C0619DC for ; Wed, 22 Jul 2020 18:01:39 -0700 (PDT) Received: by mail-pg1-x541.google.com with SMTP id n5so2199995pgf.7 for ; Wed, 22 Jul 2020 18:01:39 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; h=from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=ABRvnCiTNBT/lbqlUb0TjhpOBcWmEJ9QDMFsMPZdCjM=; b=K42b5yLk5vCsxvswOvi5PtPNaMYFL8qGQ1KvvaS2BBRHyPFXOlKvglsZHr+gwAV3Qv JZrib+oR50EbeTMgPQQ8NbGANyUyTlXyJEnifLe4vYYjO43qt/XK3YPX9RwZjqJ0hnqY P632O98wTN8ImgrEos2g7IGA750UFKRwSax68= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=ABRvnCiTNBT/lbqlUb0TjhpOBcWmEJ9QDMFsMPZdCjM=; b=shLvCf8ZeW5ipJ0mpuaO6DlaxywYyahGFcUAOeVmNDnTRPtZqRLdGc0d4Nfhx95+xf dluVttK2/K4qc/yy5O95E8GEWMCjBQ4R47rzEJEppswqRD5LQnq3atbSRX9zZAnPL13e IsfjVvN7sRbiyzVZGeeyeiLx1HXdC/qEzXTcdlhFC4nLfiRtg/X1T6CXYHRWNMHrvRSu YAuObFTMTlNeumrc27MrdPASO0MG1niRW9rNY3NTS9hYOdA4xggXsuYvnftbBRlk2ZcV X91M6wfLkpMIbCGuoTvCjRNh3N0GFdAA3xSvvkn5Lay8RmH0KU5xJ+MUbs6HMmZVmvFn qaKw== X-Gm-Message-State: AOAM532j6wzXyoNaGh2LvLqtbc1OplJ9pICejSfWTyjOvLmn29Hsl83Y yMyflFRDF+df5UGoTKcKGHxJBA== X-Google-Smtp-Source: ABdhPJw45K/rtDvfqdHnxBAbawBUhv3wG/7f5jGXXvn1rAnchjtXoQWMz47euk5JLbfnygRqFAOX9A== X-Received: by 2002:a62:834c:: with SMTP id h73mr2053097pfe.221.1595466098551; Wed, 22 Jul 2020 18:01:38 -0700 (PDT) Received: from smtp.gmail.com ([2620:15c:202:1:3e52:82ff:fe6c:83ab]) by smtp.gmail.com with ESMTPSA id y7sm825351pjy.54.2020.07.22.18.01.37 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 22 Jul 2020 18:01:38 -0700 (PDT) From: Stephen Boyd To: Andy Gross , Bjorn Andersson Cc: linux-kernel@vger.kernel.org, linux-arm-msm@vger.kernel.org, Douglas Anderson , Maulik Shah , Lina Iyer Subject: [PATCH] soc: qcom: rpmh-rsc: Sleep waiting for tcs slots to be free Date: Wed, 22 Jul 2020 18:01:37 -0700 Message-Id: <20200723010137.3127584-1-swboyd@chromium.org> X-Mailer: git-send-email 2.28.0.rc0.105.gf9edc3c819-goog MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Sender: linux-arm-msm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-arm-msm@vger.kernel.org The busy loop in rpmh_rsc_send_data() is written with the assumption that the udelay will be preempted by the tcs_tx_done() irq handler when the TCS slots are all full. This doesn't hold true when the calling thread is an irqthread and the tcs_tx_done() irq is also an irqthread. That's because kernel irqthreads are SCHED_FIFO and thus need to voluntarily give up priority by calling into the scheduler so that other threads can run. I see RCU stalls when I boot with irqthreads on the kernel commandline because the modem remoteproc driver is trying to send an rpmh async message from an irqthread that needs to give up the CPU for the rpmh irqthread to run and clear out tcs slots. rcu: INFO: rcu_preempt self-detected stall on CPU rcu: 0-....: (1 GPs behind) idle=402/1/0x4000000000000002 softirq=2108/2109 fqs=4920 (t=21016 jiffies g=2933 q=590) Task dump for CPU 0: irq/11-smp2p R running task 0 148 2 0x00000028 Call trace: dump_backtrace+0x0/0x154 show_stack+0x20/0x2c sched_show_task+0xfc/0x108 dump_cpu_task+0x44/0x50 rcu_dump_cpu_stacks+0xa4/0xf8 rcu_sched_clock_irq+0x7dc/0xaa8 update_process_times+0x30/0x54 tick_sched_handle+0x50/0x64 tick_sched_timer+0x4c/0x8c __hrtimer_run_queues+0x21c/0x36c hrtimer_interrupt+0xf0/0x22c arch_timer_handler_phys+0x40/0x50 handle_percpu_devid_irq+0x114/0x25c __handle_domain_irq+0x84/0xc4 gic_handle_irq+0xd0/0x178 el1_irq+0xbc/0x180 save_return_addr+0x18/0x28 return_address+0x54/0x88 preempt_count_sub+0x40/0x88 _raw_spin_unlock_irqrestore+0x4c/0x6c ___ratelimit+0xd0/0x128 rpmh_rsc_send_data+0x24c/0x378 __rpmh_write+0x1b0/0x208 rpmh_write_async+0x90/0xbc rpmhpd_send_corner+0x60/0x8c rpmhpd_aggregate_corner+0x8c/0x124 rpmhpd_set_performance_state+0x8c/0xbc _genpd_set_performance_state+0xdc/0x1b8 dev_pm_genpd_set_performance_state+0xb8/0xf8 q6v5_pds_disable+0x34/0x60 [qcom_q6v5_mss] qcom_msa_handover+0x38/0x44 [qcom_q6v5_mss] q6v5_handover_interrupt+0x24/0x3c [qcom_q6v5] handle_nested_irq+0xd0/0x138 qcom_smp2p_intr+0x188/0x200 irq_thread_fn+0x2c/0x70 irq_thread+0xfc/0x14c kthread+0x11c/0x12c ret_from_fork+0x10/0x18 This busy loop naturally lends itself to using a wait queue so that each thread that tries to send a message will sleep waiting on the waitqueue and only be woken up when a free slot is available. This should make things more predictable too because the scheduler will be able to sleep tasks that are waiting on a free tcs instead of the busy loop we currently have today. Cc: Douglas Anderson Cc: Maulik Shah Cc: Lina Iyer Signed-off-by: Stephen Boyd --- drivers/soc/qcom/rpmh-internal.h | 2 + drivers/soc/qcom/rpmh-rsc.c | 101 ++++++++++++------------------- 2 files changed, 41 insertions(+), 62 deletions(-) diff --git a/drivers/soc/qcom/rpmh-internal.h b/drivers/soc/qcom/rpmh-internal.h index ef60e790a750..9a325bac58fe 100644 --- a/drivers/soc/qcom/rpmh-internal.h +++ b/drivers/soc/qcom/rpmh-internal.h @@ -8,6 +8,7 @@ #define __RPM_INTERNAL_H__ #include +#include #include #define TCS_TYPE_NR 4 @@ -118,6 +119,7 @@ struct rsc_drv { struct tcs_group tcs[TCS_TYPE_NR]; DECLARE_BITMAP(tcs_in_use, MAX_TCS_NR); spinlock_t lock; + wait_queue_head_t tcs_wait; struct rpmh_ctrlr client; }; diff --git a/drivers/soc/qcom/rpmh-rsc.c b/drivers/soc/qcom/rpmh-rsc.c index 076fd27f3081..6c758b052c95 100644 --- a/drivers/soc/qcom/rpmh-rsc.c +++ b/drivers/soc/qcom/rpmh-rsc.c @@ -19,6 +19,7 @@ #include #include #include +#include #include #include @@ -444,6 +445,7 @@ static irqreturn_t tcs_tx_done(int irq, void *p) */ if (!drv->tcs[ACTIVE_TCS].num_tcs) enable_tcs_irq(drv, i, false); + wake_up(&drv->tcs_wait); spin_unlock(&drv->lock); if (req) rpmh_tx_done(req, err); @@ -562,44 +564,59 @@ static int find_free_tcs(struct tcs_group *tcs) return -EBUSY; } +static int claim_tcs_for_req(struct rsc_drv *drv, struct tcs_group *tcs, + const struct tcs_request *msg) +{ + int ret; + + /* + * The h/w does not like if we send a request to the same address, + * when one is already in-flight or being processed. + */ + ret = check_for_req_inflight(drv, tcs, msg); + if (ret) + return ret; + + return find_free_tcs(tcs); +} + /** - * tcs_write() - Store messages into a TCS right now, or return -EBUSY. + * rpmh_rsc_send_data() - Write / trigger active-only message. * @drv: The controller. * @msg: The data to be sent. * - * Grabs a TCS for ACTIVE_ONLY transfers and writes the messages to it. - * - * If there are no free TCSes for ACTIVE_ONLY transfers or if a command for - * the same address is already transferring returns -EBUSY which means the - * client should retry shortly. + * NOTES: + * - This is only used for "ACTIVE_ONLY" since the limitations of this + * function don't make sense for sleep/wake cases. + * - To do the transfer, we will grab a whole TCS for ourselves--we don't + * try to share. If there are none available we'll wait indefinitely + * for a free one. + * - This function will not wait for the commands to be finished, only for + * data to be programmed into the RPMh. See rpmh_tx_done() which will + * be called when the transfer is fully complete. + * - This function must be called with interrupts enabled. If the hardware + * is busy doing someone else's transfer we need that transfer to fully + * finish so that we can have the hardware, and to fully finish it needs + * the interrupt handler to run. If the interrupts is set to run on the + * active CPU this can never happen if interrupts are disabled. * - * Return: 0 on success, -EBUSY if client should retry, or an error. - * Client should have interrupts enabled for a bit before retrying. + * Return: 0 on success, -EINVAL on error. */ -static int tcs_write(struct rsc_drv *drv, const struct tcs_request *msg) +int rpmh_rsc_send_data(struct rsc_drv *drv, const struct tcs_request *msg) { struct tcs_group *tcs; int tcs_id; unsigned long flags; - int ret; tcs = get_tcs_for_msg(drv, msg); if (IS_ERR(tcs)) return PTR_ERR(tcs); spin_lock_irqsave(&drv->lock, flags); - /* - * The h/w does not like if we send a request to the same address, - * when one is already in-flight or being processed. - */ - ret = check_for_req_inflight(drv, tcs, msg); - if (ret) - goto unlock; - ret = find_free_tcs(tcs); - if (ret < 0) - goto unlock; - tcs_id = ret; + wait_event_lock_irq(drv->tcs_wait, + (tcs_id = claim_tcs_for_req(drv, tcs, msg)) >= 0, + drv->lock); tcs->req[tcs_id - tcs->offset] = msg; set_bit(tcs_id, drv->tcs_in_use); @@ -627,47 +644,6 @@ static int tcs_write(struct rsc_drv *drv, const struct tcs_request *msg) __tcs_set_trigger(drv, tcs_id, true); return 0; -unlock: - spin_unlock_irqrestore(&drv->lock, flags); - return ret; -} - -/** - * rpmh_rsc_send_data() - Write / trigger active-only message. - * @drv: The controller. - * @msg: The data to be sent. - * - * NOTES: - * - This is only used for "ACTIVE_ONLY" since the limitations of this - * function don't make sense for sleep/wake cases. - * - To do the transfer, we will grab a whole TCS for ourselves--we don't - * try to share. If there are none available we'll wait indefinitely - * for a free one. - * - This function will not wait for the commands to be finished, only for - * data to be programmed into the RPMh. See rpmh_tx_done() which will - * be called when the transfer is fully complete. - * - This function must be called with interrupts enabled. If the hardware - * is busy doing someone else's transfer we need that transfer to fully - * finish so that we can have the hardware, and to fully finish it needs - * the interrupt handler to run. If the interrupts is set to run on the - * active CPU this can never happen if interrupts are disabled. - * - * Return: 0 on success, -EINVAL on error. - */ -int rpmh_rsc_send_data(struct rsc_drv *drv, const struct tcs_request *msg) -{ - int ret; - - do { - ret = tcs_write(drv, msg); - if (ret == -EBUSY) { - pr_info_ratelimited("TCS Busy, retrying RPMH message send: addr=%#x\n", - msg->cmds[0].addr); - udelay(10); - } - } while (ret == -EBUSY); - - return ret; } /** @@ -975,6 +951,7 @@ static int rpmh_rsc_probe(struct platform_device *pdev) return ret; spin_lock_init(&drv->lock); + init_waitqueue_head(&drv->tcs_wait); bitmap_zero(drv->tcs_in_use, MAX_TCS_NR); irq = platform_get_irq(pdev, drv->id); base-commit: b3a9e3b9622ae10064826dccb4f7a52bd88c7407 -- Sent by a computer, using git, on the internet