From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.2 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 41998C12002 for ; Mon, 19 Jul 2021 09:15:02 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 1A0CB61026 for ; Mon, 19 Jul 2021 09:15:02 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235812AbhGSIeU (ORCPT ); Mon, 19 Jul 2021 04:34:20 -0400 Received: from foss.arm.com ([217.140.110.172]:53332 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235290AbhGSIeT (ORCPT ); Mon, 19 Jul 2021 04:34:19 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id D051C6D; Mon, 19 Jul 2021 02:14:58 -0700 (PDT) Received: from e120937-lin (unknown [172.31.20.19]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 9F0933F73D; Mon, 19 Jul 2021 02:14:54 -0700 (PDT) Date: Mon, 19 Jul 2021 10:14:51 +0100 From: Cristian Marussi To: Peter Hilber Cc: linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org, virtualization@lists.linux-foundation.org, virtio-dev@lists.oasis-open.org, sudeep.holla@arm.com, james.quinlan@broadcom.com, Jonathan.Cameron@Huawei.com, f.fainelli@gmail.com, etienne.carriere@linaro.org, vincent.guittot@linaro.org, souvik.chakravarty@arm.com, igor.skalkin@opensynergy.com, alex.bennee@linaro.org, jean-philippe@linaro.org, mikhail.golubev@opensynergy.com, anton.yakovlev@opensynergy.com, Vasyl.Vavrychuk@opensynergy.com, Andriy.Tryshnivskyy@opensynergy.com Subject: Re: [PATCH v6 07/17] firmware: arm_scmi: Handle concurrent and out-of-order messages Message-ID: <20210719091451.GF49078@e120937-lin> References: <20210712141833.6628-1-cristian.marussi@arm.com> <20210712141833.6628-8-cristian.marussi@arm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: User-Agent: Mutt/1.9.4 (2018-02-28) Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Jul 15, 2021 at 06:36:03PM +0200, Peter Hilber wrote: > On 12.07.21 16:18, Cristian Marussi wrote: > > Even though in case of asynchronous commands an SCMI platform server is > > constrained to emit the delayed response message only after the related > > message response has been sent, the configured underlying transport could > > still deliver such messages together or in inverted order, causing races > > due to the concurrent or out-of-order access to the underlying xfer. > > > > Introduce a mechanism to grant exclusive access to an xfer in order to > > properly serialize concurrent accesses to the same xfer originating from > > multiple correlated messages. > > > > Add additional state information to xfer descriptors so as to be able to > > identify out-of-order message deliveries and act accordingly: > > > > - when a delayed response is expected but delivered before the related > > response, the synchronous response is considered as successfully > > received and the delayed response processing is carried on as usual. > > > > - when/if the missing synchronous response is subsequently received, it > > is discarded as not congruent with the current state of the xfer, or > > simply, because the xfer has been already released and so, now, the > > monotonically increasing sequence number carried by the late response > > is stale. > > > > Signed-off-by: Cristian Marussi > > --- Hi Peter, thanks for the review, I replied online down below. > > v5 --> v6 > > - added spinlock comment > > --- > > drivers/firmware/arm_scmi/common.h | 18 ++- > > drivers/firmware/arm_scmi/driver.c | 229 ++++++++++++++++++++++++----- > > 2 files changed, 212 insertions(+), 35 deletions(-) > > > > diff --git a/drivers/firmware/arm_scmi/common.h b/drivers/firmware/arm_scmi/common.h > > index 2233d0a188fc..9efebe1406d2 100644 > > --- a/drivers/firmware/arm_scmi/common.h > > +++ b/drivers/firmware/arm_scmi/common.h > > @@ -19,6 +19,7 @@ > > #include > > #include > > #include > > +#include > > #include > > #include > > @@ -145,6 +146,13 @@ struct scmi_msg { > > * @pending: True for xfers added to @pending_xfers hashtable > > * @node: An hlist_node reference used to store this xfer, alternatively, on > > * the free list @free_xfers or in the @pending_xfers hashtable > > + * @busy: An atomic flag to ensure exclusive write access to this xfer > > + * @state: The current state of this transfer, with states transitions deemed > > + * valid being: > > + * - SCMI_XFER_SENT_OK -> SCMI_XFER_RESP_OK [ -> SCMI_XFER_DRESP_OK ] > > + * - SCMI_XFER_SENT_OK -> SCMI_XFER_DRESP_OK > > + * (Missing synchronous response is assumed OK and ignored) > > + * @lock: A spinlock to protect state and busy fields. > > */ > > struct scmi_xfer { > > int transfer_id; > > @@ -156,6 +164,15 @@ struct scmi_xfer { > > refcount_t users; > > bool pending; > > struct hlist_node node; > > +#define SCMI_XFER_FREE 0 > > +#define SCMI_XFER_BUSY 1 > > + atomic_t busy; > > +#define SCMI_XFER_SENT_OK 0 > > +#define SCMI_XFER_RESP_OK 1 > > +#define SCMI_XFER_DRESP_OK 2 > > + int state; > > + /* A lock to protect state and busy fields */ > > + spinlock_t lock; > > }; > > /* > > @@ -392,5 +409,4 @@ bool shmem_poll_done(struct scmi_shared_mem __iomem *shmem, > > void scmi_notification_instance_data_set(const struct scmi_handle *handle, > > void *priv); > > void *scmi_notification_instance_data_get(const struct scmi_handle *handle); > > - > > #endif /* _SCMI_COMMON_H */ > > diff --git a/drivers/firmware/arm_scmi/driver.c b/drivers/firmware/arm_scmi/driver.c > > index 245ede223302..5ef33d692670 100644 > > --- a/drivers/firmware/arm_scmi/driver.c > > +++ b/drivers/firmware/arm_scmi/driver.c > > @@ -369,6 +369,7 @@ static struct scmi_xfer *scmi_xfer_get(const struct scmi_handle *handle, > > if (!IS_ERR(xfer)) { > > refcount_set(&xfer->users, 1); > > + atomic_set(&xfer->busy, SCMI_XFER_FREE); > > xfer->transfer_id = atomic_inc_return(&transfer_last_id); > > } > > spin_unlock_irqrestore(&minfo->xfer_lock, flags); > > @@ -430,6 +431,168 @@ scmi_xfer_lookup_unlocked(struct scmi_xfers_info *minfo, u16 xfer_id) > > return xfer ?: ERR_PTR(-EINVAL); > > } > > +/** > > + * scmi_msg_response_validate - Validate message type against state of related > > + * xfer > > + * > > + * @cinfo: A reference to the channel descriptor. > > + * @msg_type: Message type to check > > + * @xfer: A reference to the xfer to validate against @msg_type > > + * > > + * This function checks if @msg_type is congruent with the current state of > > + * a pending @xfer; if an asynchronous delayed response is received before the > > + * related synchronous response (Out-of-Order Delayed Response) the missing > > + * synchronous response is assumed to be OK and completed, carrying on with the > > + * Delayed Response: this is done to address the case in which the underlying > > + * SCMI transport can deliver such out-of-order responses. > > + * > > + * Context: Assumes to be called with xfer->lock already acquired. > > + * > > + * Return: 0 on Success, error otherwise > > + */ > > +static inline int scmi_msg_response_validate(struct scmi_chan_info *cinfo, > > + u8 msg_type, > > + struct scmi_xfer *xfer) > > +{ > > + /* > > + * Even if a response was indeed expected on this slot at this point, > > + * a buggy platform could wrongly reply feeding us an unexpected > > + * delayed response we're not prepared to handle: bail-out safely > > + * blaming firmware. > > + */ > > + if (msg_type == MSG_TYPE_DELAYED_RESP && !xfer->async_done) { > > + dev_err(cinfo->dev, > > + "Delayed Response for %d not expected! Buggy F/W ?\n", > > + xfer->hdr.seq); > > + return -EINVAL; > > + } > > + > > + switch (xfer->state) { > > + case SCMI_XFER_SENT_OK: > > + if (msg_type == MSG_TYPE_DELAYED_RESP) { > > + /* > > + * Delayed Response expected but delivered earlier. > > + * Assume message RESPONSE was OK and skip state. > > + */ > > + xfer->hdr.status = SCMI_SUCCESS; > > + xfer->state = SCMI_XFER_RESP_OK; > > + complete(&xfer->done); > > + dev_warn(cinfo->dev, > > + "Received valid OoO Delayed Response for %d\n", > > + xfer->hdr.seq); > > + } > > + break; > > + case SCMI_XFER_RESP_OK: > > + if (msg_type != MSG_TYPE_DELAYED_RESP) > > + return -EINVAL; > > + break; > > + case SCMI_XFER_DRESP_OK: > > + /* No further message expected once in SCMI_XFER_DRESP_OK */ > > + return -EINVAL; > > + } > > + > > + return 0; > > +} > > + > > +static bool scmi_xfer_is_free(struct scmi_xfer *xfer) > > +{ > > + int ret; > > + > > + ret = atomic_cmpxchg(&xfer->busy, SCMI_XFER_FREE, SCMI_XFER_BUSY); > > Naming: Rather unusual to change state in an _is_free() function, looking at > other _is_free() functions in the kernel. > Indeed, but this naming was meant to reflect more the fact that is called from spin_until_cond(is_free). The use of the atomic cmpxchg trick on the busy flag is an attempt to limit as much as possible the size of the spinlocked section. > > + > > + return ret == SCMI_XFER_FREE; > > +} > > + > > +/** > > + * scmi_xfer_command_acquire - Helper to lookup and acquire a command xfer > > + * > > + * @cinfo: A reference to the channel descriptor. > > + * @msg_hdr: A message header to use as lookup key > > + * > > + * When a valid xfer is found for the sequence number embedded in the provided > > + * msg_hdr, reference counting is properly updated and exclusive access to this > > + * xfer is granted till released with @scmi_xfer_command_release. > > + * > > + * Return: A valid @xfer on Success or error otherwise. > > + */ > > +static inline struct scmi_xfer * > > +scmi_xfer_command_acquire(struct scmi_chan_info *cinfo, u32 msg_hdr) > > +{ > > + int ret; > > + unsigned long flags; > > + struct scmi_xfer *xfer; > > + struct scmi_info *info = handle_to_scmi_info(cinfo->handle); > > + struct scmi_xfers_info *minfo = &info->tx_minfo; > > + u8 msg_type = MSG_XTRACT_TYPE(msg_hdr); > > + u16 xfer_id = MSG_XTRACT_TOKEN(msg_hdr); > > + > > + /* Are we even expecting this? */ > > + spin_lock_irqsave(&minfo->xfer_lock, flags); > > + xfer = scmi_xfer_lookup_unlocked(minfo, xfer_id); > > + if (IS_ERR(xfer)) { > > + dev_err(cinfo->dev, > > + "Message for %d type %d is not expected!\n", > > + xfer_id, msg_type); > > + spin_unlock_irqrestore(&minfo->xfer_lock, flags); > > + return xfer; > > + } > > + refcount_inc(&xfer->users); > > + spin_unlock_irqrestore(&minfo->xfer_lock, flags); > > + > > + spin_lock_irqsave(&xfer->lock, flags); > > + ret = scmi_msg_response_validate(cinfo, msg_type, xfer); > > + /* > > + * If a pending xfer was found which was also in a congruent state with > > + * the received message, acquire exclusive access to it setting the busy > > + * flag. > > + * Spins only on the rare limit condition of concurrent reception of > > + * RESP and DRESP for the same xfer. > > + */ > > + if (!ret) { > > + spin_until_cond(scmi_xfer_is_free(xfer)); > > Maybe there should be an additional comment to indicate that the xfer->lock > cannot be reaquired later during response processing, to avoid a deadlock in > conjunction with the xfer->busy flag. > I'll add that. Thanks. > > + xfer->hdr.type = msg_type; > > + } > > + spin_unlock_irqrestore(&xfer->lock, flags); > > + > > + if (ret) { > > + dev_err(cinfo->dev, > > + "Invalid message type:%d for %d - HDR:0x%X state:%d\n", > > + msg_type, xfer_id, msg_hdr, xfer->state); > > + /* On error the refcount incremented above has to be dropped */ > > + __scmi_xfer_put(minfo, xfer); > > + xfer = ERR_PTR(-EINVAL); > > + } > > + > > + return xfer; > > +} > > + > > +static inline void scmi_xfer_command_release(struct scmi_info *info, > > + struct scmi_xfer *xfer) > > +{ > > + atomic_set(&xfer->busy, SCMI_XFER_FREE); > > + __scmi_xfer_put(&info->tx_minfo, xfer); > > +} > > + > > +/** > > + * scmi_xfer_state_update - Update xfer state > > + * > > + * @xfer: A reference to the xfer to update > > + * > > + * Context: Assumes to be called on an xfer exclusively acquired using the > > + * busy flag. > > + */ > > +static inline void scmi_xfer_state_update(struct scmi_xfer *xfer) > > +{ > > + switch (xfer->hdr.type) { > > + case MSG_TYPE_COMMAND: > > + xfer->state = SCMI_XFER_RESP_OK; > > + break; > > + case MSG_TYPE_DELAYED_RESP: > > + xfer->state = SCMI_XFER_DRESP_OK; > > + break; > > + } > > +} > > + > > static void scmi_handle_notification(struct scmi_chan_info *cinfo, u32 msg_hdr) > > { > > struct scmi_xfer *xfer; > > @@ -462,57 +625,37 @@ static void scmi_handle_notification(struct scmi_chan_info *cinfo, u32 msg_hdr) > > info->desc->ops->clear_channel(cinfo); > > } > > -static void scmi_handle_response(struct scmi_chan_info *cinfo, > > - u16 xfer_id, u8 msg_type) > > +static void scmi_handle_response(struct scmi_chan_info *cinfo, u32 msg_hdr) > > { > > - unsigned long flags; > > struct scmi_xfer *xfer; > > - struct device *dev = cinfo->dev; > > struct scmi_info *info = handle_to_scmi_info(cinfo->handle); > > - struct scmi_xfers_info *minfo = &info->tx_minfo; > > - /* Are we even expecting this? */ > > - spin_lock_irqsave(&minfo->xfer_lock, flags); > > - xfer = scmi_xfer_lookup_unlocked(minfo, xfer_id); > > - spin_unlock_irqrestore(&minfo->xfer_lock, flags); > > + xfer = scmi_xfer_command_acquire(cinfo, msg_hdr); > > if (IS_ERR(xfer)) { > > - dev_err(dev, "message for %d is not expected!\n", xfer_id); > > info->desc->ops->clear_channel(cinfo); > > return; > > } > > - /* > > - * Even if a response was indeed expected on this slot at this point, > > - * a buggy platform could wrongly reply feeding us an unexpected > > - * delayed response we're not prepared to handle: bail-out safely > > - * blaming firmware. > > - */ > > - if (unlikely(msg_type == MSG_TYPE_DELAYED_RESP && !xfer->async_done)) { > > - dev_err(dev, > > - "Delayed Response for %d not expected! Buggy F/W ?\n", > > - xfer_id); > > - info->desc->ops->clear_channel(cinfo); > > - /* It was unexpected, so nobody will clear the xfer if not us */ > > - __scmi_xfer_put(minfo, xfer); > > - return; > > - } > > + scmi_xfer_state_update(xfer); > > Since this update is not protected by the xfer->lock any more, it may not > become visible in time to a concurrent response which is checking and > possibly updating state in scmi_msg_response_validate(). I think this should > be avoided, even if it might not cause practical problems ATM. > Right. I missed this, and looking back at this code it's also awkward indeed to have a status update disjoint from the msg state validation. (I think is a left over from using delegated xfers...my bad) For V7 I moved scmi_xfer_state_update() inside the spinlocked section in scmi_xfer_command_acquire() right after the busy flag is set. > > /* rx.len could be shrunk in the sync do_xfer, so reset to maxsz */ > > - if (msg_type == MSG_TYPE_DELAYED_RESP) > > + if (xfer->hdr.type == MSG_TYPE_DELAYED_RESP) > > xfer->rx.len = info->desc->max_msg_size; > > info->desc->ops->fetch_response(cinfo, xfer); > > trace_scmi_rx_done(xfer->transfer_id, xfer->hdr.id, > > xfer->hdr.protocol_id, xfer->hdr.seq, > > - msg_type); > > + xfer->hdr.type); > > - if (msg_type == MSG_TYPE_DELAYED_RESP) { > > + if (xfer->hdr.type == MSG_TYPE_DELAYED_RESP) { > > info->desc->ops->clear_channel(cinfo); > > complete(xfer->async_done); > > } else { > > complete(&xfer->done); > > } > > + > > + scmi_xfer_command_release(info, xfer); > > } > > /** > > @@ -529,7 +672,6 @@ static void scmi_handle_response(struct scmi_chan_info *cinfo, > > */ > > void scmi_rx_callback(struct scmi_chan_info *cinfo, u32 msg_hdr) > > { > > - u16 xfer_id = MSG_XTRACT_TOKEN(msg_hdr); > > u8 msg_type = MSG_XTRACT_TYPE(msg_hdr); > > switch (msg_type) { > > @@ -538,7 +680,7 @@ void scmi_rx_callback(struct scmi_chan_info *cinfo, u32 msg_hdr) > > break; > > case MSG_TYPE_COMMAND: > > case MSG_TYPE_DELAYED_RESP: > > - scmi_handle_response(cinfo, xfer_id, msg_type); > > + scmi_handle_response(cinfo, msg_hdr); > > break; > > default: > > WARN_ONCE(1, "received unknown msg_type:%d\n", msg_type); > > @@ -550,7 +692,7 @@ void scmi_rx_callback(struct scmi_chan_info *cinfo, u32 msg_hdr) > > * xfer_put() - Release a transmit message > > * > > * @ph: Pointer to SCMI protocol handle > > - * @xfer: message that was reserved by scmi_xfer_get > > + * @xfer: message that was reserved by xfer_get_init > > */ > > static void xfer_put(const struct scmi_protocol_handle *ph, > > struct scmi_xfer *xfer) > > @@ -568,7 +710,12 @@ static bool scmi_xfer_done_no_timeout(struct scmi_chan_info *cinfo, > > { > > struct scmi_info *info = handle_to_scmi_info(cinfo->handle); > > + /* > > + * Poll also on xfer->done so that polling can be forcibly terminated > > + * in case of out-of-order receptions of delayed responses > > + */ > > return info->desc->ops->poll_done(cinfo, xfer) || > > + try_wait_for_completion(&xfer->done) || > > ktime_after(ktime_get(), stop); > > } > > @@ -608,6 +755,7 @@ static int do_xfer(const struct scmi_protocol_handle *ph, > > xfer->hdr.protocol_id, xfer->hdr.seq, > > xfer->hdr.poll_completion); > > + xfer->state = SCMI_XFER_SENT_OK; > > To be completely safe, this assignment could also be protected by the > xfer->lock. > In fact this would be true being xfer->lock meant to protect the state but it seemed to me unnecessary here given that this is a brand new xfer with a brand new (monotonic) seq number so that any possibly late-received msg will carry an old stale seq number certainly different from this such that cannot be possibly mapped to this same xfer. (but just discarded on xfer lookup in xfer_command_acquire) The issue indeed could still exist only for do_xfer loops (as you pointed out already early on) where the seq_num is used, but in that case on a timeout we would have already bailed out of the loop and reported an error so any timed-out late received response would have been anyway discarded; so at the end I thought I could avoid spinlocking here. Thanks, Cristian