All of lore.kernel.org
 help / color / mirror / Atom feed
From: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
To: Chris Wilson <chris@chris-wilson.co.uk>, intel-gfx@lists.freedesktop.org
Subject: Re: [Intel-gfx] [PATCH 11/20] drm/i915/gem: Separate the ww_mutex walker into its own list
Date: Mon, 13 Jul 2020 15:53:56 +0100	[thread overview]
Message-ID: <9e293a7a-0592-f102-bcc3-3d0d36c4da4f@linux.intel.com> (raw)
In-Reply-To: <20200706061926.6687-12-chris@chris-wilson.co.uk>


On 06/07/2020 07:19, Chris Wilson wrote:
> In preparation for making eb_vma bigger and heavy to run inn parallel,

in

> we need to stop apply an in-place swap() to reorder around ww_mutex

applying

> deadlocks. Keep the array intact and reorder the locks using a dedicated
> list.
> 
> Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> ---
>   .../gpu/drm/i915/gem/i915_gem_execbuffer.c    | 83 ++++++++++++-------
>   1 file changed, 54 insertions(+), 29 deletions(-)
> 
> diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> index 6d20be29ff3c..4d8ac89c56fc 100644
> --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
> @@ -37,6 +37,7 @@ struct eb_vma {
>   	struct list_head bind_link;
>   	struct list_head unbound_link;
>   	struct list_head reloc_link;
> +	struct list_head submit_link;
>   
>   	struct hlist_node node;
>   	u32 handle;
> @@ -248,6 +249,8 @@ struct i915_execbuffer {
>   	/** list of vma that have execobj.relocation_count */
>   	struct list_head relocs;
>   
> +	struct list_head submit_list;
> +
>   	/**
>   	 * Track the most recently used object for relocations, as we
>   	 * frequently have to perform multiple relocations within the same
> @@ -341,6 +344,42 @@ static void eb_vma_array_put(struct eb_vma_array *arr)
>   	kref_put(&arr->kref, eb_vma_array_destroy);
>   }
>   
> +static int
> +eb_lock_vma(struct i915_execbuffer *eb, struct ww_acquire_ctx *acquire)

vmas plural?

> +{
> +	struct eb_vma *ev;
> +	int err = 0;
> +
> +	list_for_each_entry(ev, &eb->submit_list, submit_link) {
> +		struct i915_vma *vma = ev->vma;
> +
> +		err = ww_mutex_lock_interruptible(&vma->resv->lock, acquire);
> +		if (err == -EDEADLK) {
> +			struct eb_vma *unlock = ev, *en;
> +
> +			list_for_each_entry_safe_continue_reverse(unlock, en,
> +								  &eb->submit_list,
> +								  submit_link) {
> +				ww_mutex_unlock(&unlock->vma->resv->lock);
> +				list_move_tail(&unlock->submit_link, &eb->submit_list);
> +			}
> +
> +			GEM_BUG_ON(!list_is_first(&ev->submit_link, &eb->submit_list));
> +			err = ww_mutex_lock_slow_interruptible(&vma->resv->lock,
> +							       acquire);
> +		}
> +		if (err) {
> +			list_for_each_entry_continue_reverse(ev,
> +							     &eb->submit_list,
> +							     submit_link)
> +				ww_mutex_unlock(&ev->vma->resv->lock);
> +			break;
> +		}
> +	}
> +
> +	return err;
> +}
> +
>   static int eb_create(struct i915_execbuffer *eb)
>   {
>   	/* Allocate an extra slot for use by the command parser + sentinel */
> @@ -393,6 +432,10 @@ static int eb_create(struct i915_execbuffer *eb)
>   		eb->lut_size = -eb->buffer_count;
>   	}
>   
> +	INIT_LIST_HEAD(&eb->bind_list);
> +	INIT_LIST_HEAD(&eb->submit_list);
> +	INIT_LIST_HEAD(&eb->relocs);
> +
>   	return 0;
>   }
>   
> @@ -574,6 +617,7 @@ eb_add_vma(struct i915_execbuffer *eb,
>   	}
>   
>   	list_add_tail(&ev->bind_link, &eb->bind_list);
> +	list_add_tail(&ev->submit_link, &eb->submit_list);
>   
>   	if (entry->relocation_count)
>   		list_add_tail(&ev->reloc_link, &eb->relocs);
> @@ -910,9 +954,6 @@ static int eb_lookup_vmas(struct i915_execbuffer *eb)
>   	unsigned int i;
>   	int err = 0;
>   
> -	INIT_LIST_HEAD(&eb->bind_list);
> -	INIT_LIST_HEAD(&eb->relocs);
> -
>   	for (i = 0; i < eb->buffer_count; i++) {
>   		struct i915_vma *vma;
>   
> @@ -1583,38 +1624,19 @@ static int eb_relocate(struct i915_execbuffer *eb)
>   
>   static int eb_move_to_gpu(struct i915_execbuffer *eb)
>   {
> -	const unsigned int count = eb->buffer_count;
>   	struct ww_acquire_ctx acquire;
> -	unsigned int i;
> +	struct eb_vma *ev;
>   	int err = 0;
>   
>   	ww_acquire_init(&acquire, &reservation_ww_class);
>   
> -	for (i = 0; i < count; i++) {
> -		struct eb_vma *ev = &eb->vma[i];
> -		struct i915_vma *vma = ev->vma;
> -
> -		err = ww_mutex_lock_interruptible(&vma->resv->lock, &acquire);
> -		if (err == -EDEADLK) {
> -			GEM_BUG_ON(i == 0);
> -			do {
> -				int j = i - 1;
> -
> -				ww_mutex_unlock(&eb->vma[j].vma->resv->lock);
> -
> -				swap(eb->vma[i],  eb->vma[j]);
> -			} while (--i);
> +	err = eb_lock_vma(eb, &acquire);
> +	if (err)
> +		goto err_fini;
>   
> -			err = ww_mutex_lock_slow_interruptible(&vma->resv->lock,
> -							       &acquire);
> -		}
> -		if (err)
> -			break;
> -	}
>   	ww_acquire_done(&acquire);
>   
> -	while (i--) {
> -		struct eb_vma *ev = &eb->vma[i];
> +	list_for_each_entry(ev, &eb->submit_list, submit_link) {
>   		struct i915_vma *vma = ev->vma;
>   		unsigned int flags = ev->flags;
>   		struct drm_i915_gem_object *obj = vma->obj;
> @@ -1671,6 +1693,8 @@ static int eb_move_to_gpu(struct i915_execbuffer *eb)
>   	intel_gt_chipset_flush(eb->engine->gt);
>   	return 0;
>   
> +err_fini:
> +	ww_acquire_fini(&acquire);
>   err_skip:
>   	i915_request_set_error_once(eb->request, err);
>   	return err;
> @@ -1952,9 +1976,10 @@ static int eb_parse(struct i915_execbuffer *eb)
>   	if (err)
>   		goto err_trampoline;
>   
> -	eb->vma[eb->buffer_count].vma = i915_vma_get(shadow);
> -	eb->vma[eb->buffer_count].flags = __EXEC_OBJECT_HAS_PIN;
>   	eb->batch = &eb->vma[eb->buffer_count++];
> +	eb->batch->vma = i915_vma_get(shadow);
> +	eb->batch->flags = __EXEC_OBJECT_HAS_PIN;
> +	list_add_tail(&eb->batch->submit_link, &eb->submit_list);
>   	eb->vma[eb->buffer_count].vma = NULL;
>   
>   	eb->trampoline = trampoline;
> 

Just a temporary stage... are we reviewing those? Best if they can be 
avoided.

Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>

Regards,

Tvrtko
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx

  reply	other threads:[~2020-07-13 14:54 UTC|newest]

Thread overview: 65+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-07-06  6:19 [Intel-gfx] s/obj->mm.lock// Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 01/20] drm/i915: Preallocate stashes for vma page-directories Chris Wilson
2020-07-06 18:15   ` Matthew Auld
2020-07-06 18:20     ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 02/20] drm/i915: Switch to object allocations for page directories Chris Wilson
2020-07-06 19:06   ` Matthew Auld
2020-07-06 19:31     ` Chris Wilson
2020-07-06 20:01     ` Chris Wilson
2020-07-06 21:08       ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 03/20] drm/i915/gem: Don't drop the timeline lock during execbuf Chris Wilson
2020-07-08 16:54   ` Tvrtko Ursulin
2020-07-08 18:08     ` Chris Wilson
2020-07-09 10:52       ` Tvrtko Ursulin
2020-07-09 10:57         ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 04/20] drm/i915/gem: Rename execbuf.bind_link to unbound_link Chris Wilson
2020-07-10 11:26   ` Tvrtko Ursulin
2020-07-06  6:19 ` [Intel-gfx] [PATCH 05/20] drm/i915/gem: Break apart the early i915_vma_pin from execbuf object lookup Chris Wilson
2020-07-10 11:27   ` Tvrtko Ursulin
2020-07-06  6:19 ` [Intel-gfx] [PATCH 06/20] drm/i915/gem: Remove the call for no-evict i915_vma_pin Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 07/20] drm/i915: Add list_for_each_entry_safe_continue_reverse Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 08/20] drm/i915: Always defer fenced work to the worker Chris Wilson
2020-07-08 12:18   ` Tvrtko Ursulin
2020-07-08 12:25     ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 09/20] drm/i915/gem: Assign context id for async work Chris Wilson
2020-07-08 12:26   ` Tvrtko Ursulin
2020-07-08 12:42     ` Chris Wilson
2020-07-08 14:24       ` Tvrtko Ursulin
2020-07-08 15:36         ` Chris Wilson
2020-07-09 11:01           ` Tvrtko Ursulin
2020-07-09 11:07             ` Chris Wilson
2020-07-09 11:59               ` Tvrtko Ursulin
2020-07-09 12:07                 ` Chris Wilson
2020-07-13 12:22                   ` Tvrtko Ursulin
2020-07-14 14:01                     ` Chris Wilson
2020-07-08 12:45     ` Tvrtko Ursulin
2020-07-06  6:19 ` [Intel-gfx] [PATCH 10/20] drm/i915: Export a preallocate variant of i915_active_acquire() Chris Wilson
2020-07-09 14:36   ` Maarten Lankhorst
2020-07-10 12:24     ` Tvrtko Ursulin
2020-07-10 12:32       ` Maarten Lankhorst
2020-07-13 14:29   ` Tvrtko Ursulin
2020-07-06  6:19 ` [Intel-gfx] [PATCH 11/20] drm/i915/gem: Separate the ww_mutex walker into its own list Chris Wilson
2020-07-13 14:53   ` Tvrtko Ursulin [this message]
2020-07-14 14:10     ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 12/20] drm/i915/gem: Asynchronous GTT unbinding Chris Wilson
2020-07-14  9:02   ` Tvrtko Ursulin
2020-07-14 15:05     ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 13/20] drm/i915/gem: Bind the fence async for execbuf Chris Wilson
2020-07-14 12:19   ` Tvrtko Ursulin
2020-07-14 15:21     ` Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 14/20] drm/i915/gem: Include cmdparser in common execbuf pinning Chris Wilson
2020-07-14 12:48   ` Tvrtko Ursulin
2020-07-06  6:19 ` [Intel-gfx] [PATCH 15/20] drm/i915/gem: Include secure batch " Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 16/20] drm/i915/gem: Reintroduce multiple passes for reloc processing Chris Wilson
2020-07-09 15:39   ` Tvrtko Ursulin
2020-07-06  6:19 ` [Intel-gfx] [PATCH 17/20] drm/i915: Add an implementation for i915_gem_ww_ctx locking, v2 Chris Wilson
2020-07-06 17:21   ` kernel test robot
2020-07-06  6:19 ` [Intel-gfx] [PATCH 18/20] drm/i915/gem: Pull execbuf dma resv under a single critical section Chris Wilson
2020-07-06  6:19 ` [Intel-gfx] [PATCH 19/20] drm/i915/gem: Replace i915_gem_object.mm.mutex with reservation_ww_class Chris Wilson
2020-07-09 14:06   ` Maarten Lankhorst
2020-07-06  6:19 ` [Intel-gfx] [PATCH 20/20] drm/i915: Track i915_vma with its own reference counter Chris Wilson
2020-07-06  6:28 ` [Intel-gfx] ✗ Fi.CI.CHECKPATCH: warning for series starting with [01/20] drm/i915: Preallocate stashes for vma page-directories Patchwork
2020-07-06  6:29 ` [Intel-gfx] ✗ Fi.CI.SPARSE: " Patchwork
2020-07-06  6:51 ` [Intel-gfx] ✓ Fi.CI.BAT: success " Patchwork
2020-07-06  7:55 ` [Intel-gfx] ✗ Fi.CI.IGT: failure " Patchwork
2020-07-27 18:53 ` [Intel-gfx] s/obj->mm.lock// Thomas Hellström (Intel)

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=9e293a7a-0592-f102-bcc3-3d0d36c4da4f@linux.intel.com \
    --to=tvrtko.ursulin@linux.intel.com \
    --cc=chris@chris-wilson.co.uk \
    --cc=intel-gfx@lists.freedesktop.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.