From: Sean Paul <sean@poorly.run>
To: "Lin, Wayne" <Wayne.Lin@amd.com>
Cc: David Airlie <airlied@linux.ie>,
Sean Paul <seanpaul@chromium.org>,
"dri-devel@lists.freedesktop.org"
<dri-devel@lists.freedesktop.org>, Sean Paul <sean@poorly.run>
Subject: Re: [PATCH 3/3] drm/dp_mst: Remove single tx msg restriction.
Date: Tue, 18 Feb 2020 12:15:22 -0500 [thread overview]
Message-ID: <20200218171522.GF253734@art_vandelay> (raw)
In-Reply-To: <20200218155206.GD253734@art_vandelay>
On Tue, Feb 18, 2020 at 10:52:06AM -0500, Sean Paul wrote:
> On Mon, Feb 17, 2020 at 07:08:37AM +0000, Lin, Wayne wrote:
> > [AMD Public Use]
> >
> >
> >
> > > -----Original Message-----
> > > From: Sean Paul <sean@poorly.run>
> > > Sent: Saturday, February 15, 2020 12:09 AM
> > > To: Lin, Wayne <Wayne.Lin@amd.com>
> > > Cc: dri-devel@lists.freedesktop.org; lyude@redhat.com; Sean Paul
> > > <seanpaul@chromium.org>; Maarten Lankhorst
> > > <maarten.lankhorst@linux.intel.com>; Maxime Ripard <mripard@kernel.org>;
> > > David Airlie <airlied@linux.ie>
> > > Subject: Re: [PATCH 3/3] drm/dp_mst: Remove single tx msg restriction.
> > >
> > > On Fri, Feb 14, 2020 at 12:58 AM Lin, Wayne <Wayne.Lin@amd.com> wrote:
> > > >
> > > > [AMD Public Use]
> > > >
> > > > Hi Paul,
> > > >
> > > > Thanks for the mail!
> > > >
> > > > I tried to solve this problem by having restriction on sending one msg at a
> > > time due to hub/dock compatibility problems.
> > > > From my experience, some branch devices don't handle well on
> > > > interleaved replies (Dock from HP I think)
> > >
> > > Hi Wayne,
> > > Hmm, that's interesting, do you have a part number of the failing dock so I can
> > > test it?
> > >
> > Hi Paul,
> >
> > Sorry but it's been quite a while. I can't exactly tell the part number.
> > If I remember correctly, when the specific branch device receives interleaved replies,
> > it just doesn't reply to any requests.
> >
> > > > As the result of that, correct me if I'm wrong, I remember most gpu vendors
> > > just send one down request at a time now in windows environment.
> > > > I would suggest the original solution :)
> > >
> > > I can't really say what happens on the Windows side of the world, but I suppose
> > > that makes sense if this is a widespread issue with docks. I do worry about the
> > > performance hit.
> > >
> > > If indeed this is a problem, could we ratelimit per branch device instead of
> > > globally? Even that would be better than serializing everything.
> > >
> > Since the problem was because some branch devices can't simultaneously handle
> > two replies, I'm afraid that we might still encounter the same problem?
> >
>
> Hi Wayne,
> Thanks for clarifying. I'm a bit hesitant to scrap this idea without solid
> evidence that this is a common problem. Do you have any hubs around AMD that
> you could try my patchset with?
Oh, if you can test the set, you'll also need this patch as well :-)
diff --git a/drivers/gpu/drm/drm_dp_mst_topology.c b/drivers/gpu/drm/drm_dp_mst_topology.c
--- a/drivers/gpu/drm/drm_dp_mst_topology.c
+++ b/drivers/gpu/drm/drm_dp_mst_topology.c
@@ -3814,7 +3814,9 @@ static bool drm_dp_get_one_sb_msg(struct drm_dp_mst_topology_mgr *mgr, bool up,
int basereg = up ? DP_SIDEBAND_MSG_UP_REQ_BASE :
DP_SIDEBAND_MSG_DOWN_REP_BASE;
- *mstb = NULL;
+ if (mstb)
+ *mstb = NULL;
+
*seqno = -1;
len = min(mgr->max_dpcd_transaction_bytes, 16);
> Perhaps we could get some hard data? I'm happy
> to test things on my end, but I probably shouldn't just start buying a bunch of
> random HP docks in hopes one of them exhibits the issue :)
>
> To add anecdote to anecdote, everything on my desk seems to work with
> interleaved replies.
>
> Since HDCP spec requires the host to verify each channel's encryption every <2s,
> we're going to be increasing the amount of sideband traffic a fair amount, so I
> would like to ensure we're doing everything possible to maximize our sideband
> bandwidth.
>
> Sean
>
> > Thanks!
> > > Sean
> > >
> > > >
> > > > Thanks!
> > > > > -----Original Message-----
> > > > > From: Sean Paul <sean@poorly.run>
> > > > > Sent: Friday, February 14, 2020 5:15 AM
> > > > > To: dri-devel@lists.freedesktop.org
> > > > > Cc: lyude@redhat.com; Lin, Wayne <Wayne.Lin@amd.com>; Sean Paul
> > > > > <seanpaul@chromium.org>; Maarten Lankhorst
> > > > > <maarten.lankhorst@linux.intel.com>; Maxime Ripard
> > > > > <mripard@kernel.org>; David Airlie <airlied@linux.ie>
> > > > > Subject: [PATCH 3/3] drm/dp_mst: Remove single tx msg restriction.
> > > > >
> > > > > From: Sean Paul <seanpaul@chromium.org>
> > > > >
> > > > > Now that we can support multiple simultaneous replies, remove the
> > > > > restrictions placed on sending new tx msgs.
> > > > >
> > > > > This patch essentially just reverts commit
> > > > > 5a64967a2f3b ("drm/dp_mst: Have DP_Tx send one msg at a time")
> > > now
> > > > > that the problem is solved in a different way.
> > > > >
> > > > > Cc: Wayne Lin <Wayne.Lin@amd.com>
> > > > > Signed-off-by: Sean Paul <seanpaul@chromium.org>
> > > > > ---
> > > > > drivers/gpu/drm/drm_dp_mst_topology.c | 14 ++------------
> > > > > include/drm/drm_dp_mst_helper.h | 6 ------
> > > > > 2 files changed, 2 insertions(+), 18 deletions(-)
> > > > >
> > > > > diff --git a/drivers/gpu/drm/drm_dp_mst_topology.c
> > > > > b/drivers/gpu/drm/drm_dp_mst_topology.c
> > > > > index 7e6a82efdfc02..cbf0bb0ddeb84 100644
> > > > > --- a/drivers/gpu/drm/drm_dp_mst_topology.c
> > > > > +++ b/drivers/gpu/drm/drm_dp_mst_topology.c
> > > > > @@ -1203,8 +1203,6 @@ static int drm_dp_mst_wait_tx_reply(struct
> > > > > drm_dp_mst_branch *mstb,
> > > > > txmsg->state == DRM_DP_SIDEBAND_TX_SENT) {
> > > > > mstb->tx_slots[txmsg->seqno] = NULL;
> > > > > }
> > > > > - mgr->is_waiting_for_dwn_reply = false;
> > > > > -
> > > > > }
> > > > > out:
> > > > > if (unlikely(ret == -EIO) && drm_debug_enabled(DRM_UT_DP)) {
> > > > > @@
> > > > > -1214,7 +1212,6 @@ static int drm_dp_mst_wait_tx_reply(struct
> > > > > drm_dp_mst_branch *mstb,
> > > > > }
> > > > > mutex_unlock(&mgr->qlock);
> > > > >
> > > > > - drm_dp_mst_kick_tx(mgr);
> > > > > return ret;
> > > > > }
> > > > >
> > > > > @@ -2797,11 +2794,9 @@ static void
> > > > > process_single_down_tx_qlock(struct
> > > > > drm_dp_mst_topology_mgr *mgr)
> > > > > ret = process_single_tx_qlock(mgr, txmsg, false);
> > > > > if (ret == 1) {
> > > > > /* txmsg is sent it should be in the slots now */
> > > > > - mgr->is_waiting_for_dwn_reply = true;
> > > > > list_del(&txmsg->next);
> > > > > } else if (ret) {
> > > > > DRM_DEBUG_KMS("failed to send msg in q %d\n", ret);
> > > > > - mgr->is_waiting_for_dwn_reply = false;
> > > > > list_del(&txmsg->next);
> > > > > if (txmsg->seqno != -1)
> > > > > txmsg->dst->tx_slots[txmsg->seqno] = NULL;
> > > @@
> > > > > -2841,8
> > > > > +2836,7 @@ static void drm_dp_queue_down_tx(struct
> > > > > drm_dp_mst_topology_mgr *mgr,
> > > > > drm_dp_mst_dump_sideband_msg_tx(&p, txmsg);
> > > > > }
> > > > >
> > > > > - if (list_is_singular(&mgr->tx_msg_downq) &&
> > > > > - !mgr->is_waiting_for_dwn_reply)
> > > > > + if (list_is_singular(&mgr->tx_msg_downq))
> > > > > process_single_down_tx_qlock(mgr);
> > > > > mutex_unlock(&mgr->qlock);
> > > > > }
> > > > > @@ -3822,7 +3816,6 @@ static int drm_dp_mst_handle_down_rep(struct
> > > > > drm_dp_mst_topology_mgr *mgr)
> > > > > mutex_lock(&mgr->qlock);
> > > > > txmsg->state = DRM_DP_SIDEBAND_TX_RX;
> > > > > mstb->tx_slots[seqno] = NULL;
> > > > > - mgr->is_waiting_for_dwn_reply = false;
> > > > > mutex_unlock(&mgr->qlock);
> > > > >
> > > > > wake_up_all(&mgr->tx_waitq);
> > > > > @@ -3830,9 +3823,6 @@ static int drm_dp_mst_handle_down_rep(struct
> > > > > drm_dp_mst_topology_mgr *mgr)
> > > > > return 0;
> > > > >
> > > > > out_clear_reply:
> > > > > - mutex_lock(&mgr->qlock);
> > > > > - mgr->is_waiting_for_dwn_reply = false;
> > > > > - mutex_unlock(&mgr->qlock);
> > > > > if (msg)
> > > > > memset(msg, 0, sizeof(struct
> > > drm_dp_sideband_msg_rx));
> > > > > out:
> > > > > @@ -4670,7 +4660,7 @@ static void drm_dp_tx_work(struct work_struct
> > > > > *work)
> > > > > struct drm_dp_mst_topology_mgr *mgr = container_of(work,
> > > > > struct drm_dp_mst_topology_mgr, tx_work);
> > > > >
> > > > > mutex_lock(&mgr->qlock);
> > > > > - if (!list_empty(&mgr->tx_msg_downq)
> > > > > && !mgr->is_waiting_for_dwn_reply)
> > > > > + if (!list_empty(&mgr->tx_msg_downq))
> > > > > process_single_down_tx_qlock(mgr);
> > > > > mutex_unlock(&mgr->qlock);
> > > > > }
> > > > > diff --git a/include/drm/drm_dp_mst_helper.h
> > > > > b/include/drm/drm_dp_mst_helper.h index
> > > 7d0341f94ce1b..fcc30e64c8e7e
> > > > > 100644
> > > > > --- a/include/drm/drm_dp_mst_helper.h
> > > > > +++ b/include/drm/drm_dp_mst_helper.h
> > > > > @@ -619,12 +619,6 @@ struct drm_dp_mst_topology_mgr {
> > > > > * &drm_dp_sideband_msg_tx.state once they are queued
> > > > > */
> > > > > struct mutex qlock;
> > > > > -
> > > > > - /**
> > > > > - * @is_waiting_for_dwn_reply: indicate whether is waiting for
> > > down
> > > > > reply
> > > > > - */
> > > > > - bool is_waiting_for_dwn_reply;
> > > > > -
> > > > > /**
> > > > > * @tx_msg_downq: List of pending down replies.
> > > > > */
> > > > > --
> > > > > Sean Paul, Software Engineer, Google / Chromium OS
> > > > --
> > > > Wayne Lin
> > --
> > Best regards,
> > Wayne Lin
>
> --
> Sean Paul, Software Engineer, Google / Chromium OS
--
Sean Paul, Software Engineer, Google / Chromium OS
_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel
next prev parent reply other threads:[~2020-02-18 17:15 UTC|newest]
Thread overview: 14+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-02-13 21:15 [PATCH 0/3] drm/mst: Add support for simultaneous down replies Sean Paul
2020-02-13 21:15 ` [PATCH 1/3] drm/mst: Separate sideband packet header parsing from message building Sean Paul
2020-02-13 21:15 ` [PATCH 2/3] drm/mst: Support simultaneous down replies Sean Paul
2020-02-13 21:15 ` [PATCH 3/3] drm/dp_mst: Remove single tx msg restriction Sean Paul
2020-02-14 5:57 ` Lin, Wayne
2020-02-14 16:08 ` Sean Paul
2020-02-17 7:08 ` Lin, Wayne
2020-02-18 15:52 ` Sean Paul
2020-02-18 17:15 ` Sean Paul [this message]
2020-02-19 11:00 ` Lin, Wayne
2020-03-05 17:19 ` Sean Paul
2020-03-06 13:41 ` Lin, Wayne
2020-02-14 23:33 ` [PATCH 0/3] drm/mst: Add support for simultaneous down replies Lyude Paul
2020-02-18 16:20 ` Sean Paul
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200218171522.GF253734@art_vandelay \
--to=sean@poorly.run \
--cc=Wayne.Lin@amd.com \
--cc=airlied@linux.ie \
--cc=dri-devel@lists.freedesktop.org \
--cc=seanpaul@chromium.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).