All of lore.kernel.org
 help / color / mirror / Atom feed
From: Chris Wilson <chris@chris-wilson.co.uk>
To: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>,
	intel-gfx@lists.freedesktop.org
Subject: Re: [PATCH v2 1/6] drm/i915: Stop tracking timeline->inflight_seqnos
Date: Tue, 24 Apr 2018 15:04:47 +0100	[thread overview]
Message-ID: <152457868721.12387.3278773617317596065@mail.alporthouse.com> (raw)
In-Reply-To: <2225828e-0189-5967-cd5d-de2cf05b8fd1@linux.intel.com>

Quoting Tvrtko Ursulin (2018-04-24 14:55:51)
> 
> On 24/04/2018 12:28, Chris Wilson wrote:
> > Quoting Tvrtko Ursulin (2018-04-24 12:17:15)
> >>
> >> On 24/04/2018 11:40, Chris Wilson wrote:
> >>> Quoting Tvrtko Ursulin (2018-04-24 11:14:21)
> >>>>
> >>>> On 23/04/2018 19:08, Chris Wilson wrote:
> >>>>> -static int reserve_engine(struct intel_engine_cs *engine)
> >>>>> +static int reserve_gt(struct drm_i915_private *i915)
> >>>>>     {
> >>>>> -     struct drm_i915_private *i915 = engine->i915;
> >>>>> -     u32 active = ++engine->timeline->inflight_seqnos;
> >>>>> -     u32 seqno = engine->timeline->seqno;
> >>>>>         int ret;
> >>>>>     
> >>>>> -     /* Reservation is fine until we need to wrap around */
> >>>>> -     if (unlikely(add_overflows(seqno, active))) {
> >>>>> +     /*
> >>>>> +      * Reservation is fine until we may need to wrap around
> >>>>> +      *
> >>>>> +      * By incrementing the serial for every request, we know that no
> >>>>> +      * individual engine may exceed that serial (as each is reset to 0
> >>>>> +      * on any wrap). This protects even the most pessimistic of migrations
> >>>>> +      * of every request from all engines onto just one.
> >>>>> +      */
> >>>>
> >>>> I didn't really figure out what was wrong with v1? Neither could handle
> >>>> more than four billion of simultaneously active requests - but I thought
> >>>> that should not concern us. :)
> >>>
> >>> It was still using the local engine->timeline.seqno as it's base. If we
> >>> swapped from one at 0 to another at U32_MAX, we would overflow much
> >>> later in submission; after the point of no return.
> >>
> >> By swapped you already refer to engine change? Ok, I can see that yes.
> >> In this case global counter does prevent that.
> >>
> >> In the light of that, what is your current thinking with regards to
> >> mixing engine classes?
> > 
> > That classes are a hw limitation that doesn't impact on balancing itself,
> > just which engines the user is allowed to put into the same group.
> > 
> >> If the thinking is still to only allow within a class then per-class
> >> seqno counter would be an option.
> > 
> > The goal of localising the seqno here was to try and reduce the locking
> > requirements (or at least make it easier to reduce them in future).
> > Whether it's one u32 across all engines, or one u32 across a few isn't
> > enough for me to worry. The breadcrumb tracking should be happy enough
> > (sorted by i915_seqno_passed rather than absolute u32) so the only
> > limitation in wrapping should be gen7 HW semaphores. Hmm, with a bit of
> > thought, I believe we can reduce the wrap logic to simply skip semaphore
> > sync inside the danger zone. Would be worth the effort.
> 
> I was thinking about reducing the number of global seqno resets as much 
> as we can in general. For instance would it be possible to keep using 
> the gt.active_requests together with a new gt.max_engine_seqno? The 
> latter would be the maximum last allocated seqno from the engine 
> timelines. This way reset would be much less frequent if the load is 
> distributed over engines (divided by num engines less frequent).

I win with a divide by 0 with removing the global seqno and wrap. :-p

The frequency we are talking about is a short wrap (will take as long as
the active request takes to sync) approximately every 47 days divided by
N engines. The cost of contention on struct_mutex must surely outweigh
that during those 47/N days...
-Chris
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

  reply	other threads:[~2018-04-24 14:04 UTC|newest]

Thread overview: 26+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-04-23 18:08 [PATCH v2 1/6] drm/i915: Stop tracking timeline->inflight_seqnos Chris Wilson
2018-04-23 18:08 ` [PATCH v2 2/6] drm/i915: Retire requests along rings Chris Wilson
2018-04-23 18:08 ` [PATCH v2 3/6] drm/i915: Only track live rings for retiring Chris Wilson
2018-04-24  9:37   ` Tvrtko Ursulin
2018-04-24  9:43     ` Chris Wilson
2018-04-23 18:08 ` [PATCH v2 4/6] drm/i915: Move timeline from GTT to ring Chris Wilson
2018-04-23 18:08 ` [PATCH v2 5/6] drm/i915: Split i915_gem_timeline into individual timelines Chris Wilson
2018-04-23 23:11   ` [PATCH] " Chris Wilson
2018-04-23 18:08 ` [PATCH v2 6/6] drm/i915: Lazily unbind vma on close Chris Wilson
2018-04-23 18:17 ` ✗ Fi.CI.CHECKPATCH: warning for series starting with [v2,1/6] drm/i915: Stop tracking timeline->inflight_seqnos Patchwork
2018-04-23 18:19 ` ✗ Fi.CI.SPARSE: " Patchwork
2018-04-23 18:32 ` ✗ Fi.CI.BAT: failure " Patchwork
2018-04-23 20:29 ` ✗ Fi.CI.CHECKPATCH: warning " Patchwork
2018-04-23 20:32 ` ✗ Fi.CI.SPARSE: " Patchwork
2018-04-23 20:46 ` ✗ Fi.CI.BAT: failure " Patchwork
2018-04-23 23:53 ` ✗ Fi.CI.CHECKPATCH: warning for series starting with [v2,1/6] drm/i915: Stop tracking timeline->inflight_seqnos (rev2) Patchwork
2018-04-23 23:55 ` ✗ Fi.CI.SPARSE: " Patchwork
2018-04-24  0:08 ` ✗ Fi.CI.BAT: failure " Patchwork
2018-04-24 10:14 ` [PATCH v2 1/6] drm/i915: Stop tracking timeline->inflight_seqnos Tvrtko Ursulin
2018-04-24 10:40   ` Chris Wilson
2018-04-24 11:17     ` Tvrtko Ursulin
2018-04-24 11:28       ` Chris Wilson
2018-04-24 13:55         ` Tvrtko Ursulin
2018-04-24 14:04           ` Chris Wilson [this message]
2018-04-24 14:48             ` Tvrtko Ursulin
2018-04-24 15:58               ` Chris Wilson

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=152457868721.12387.3278773617317596065@mail.alporthouse.com \
    --to=chris@chris-wilson.co.uk \
    --cc=intel-gfx@lists.freedesktop.org \
    --cc=tvrtko.ursulin@linux.intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.