From: Jason Ekstrand <jason@jlekstrand.net>
To: intel-gfx@lists.freedesktop.org, dri-devel@lists.freedesktop.org
Cc: Daniel Vetter <daniel.vetter@ffwll.ch>,
Jason Ekstrand <jason@jlekstrand.net>
Subject: [PATCH 11/29] drm/i915/request: Remove the hook from await_execution
Date: Thu, 27 May 2021 11:26:32 -0500 [thread overview]
Message-ID: <20210527162650.1182544-12-jason@jlekstrand.net> (raw)
In-Reply-To: <20210527162650.1182544-1-jason@jlekstrand.net>
This was only ever used for FENCE_SUBMIT automatic engine selection
which was removed in the previous commit.
Signed-off-by: Jason Ekstrand <jason@jlekstrand.net>
Reviewed-by: Daniel Vetter <daniel.vetter@ffwll.ch>
---
.../gpu/drm/i915/gem/i915_gem_execbuffer.c | 3 +-
drivers/gpu/drm/i915/i915_request.c | 42 ++++---------------
drivers/gpu/drm/i915/i915_request.h | 4 +-
3 files changed, 9 insertions(+), 40 deletions(-)
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
index efb2fa3522a42..7024adcd5cf15 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
@@ -3473,8 +3473,7 @@ i915_gem_do_execbuffer(struct drm_device *dev,
if (in_fence) {
if (args->flags & I915_EXEC_FENCE_SUBMIT)
err = i915_request_await_execution(eb.request,
- in_fence,
- NULL);
+ in_fence);
else
err = i915_request_await_dma_fence(eb.request,
in_fence);
diff --git a/drivers/gpu/drm/i915/i915_request.c b/drivers/gpu/drm/i915/i915_request.c
index 970d8f4986bbe..53f23ce40dd63 100644
--- a/drivers/gpu/drm/i915/i915_request.c
+++ b/drivers/gpu/drm/i915/i915_request.c
@@ -49,7 +49,6 @@
struct execute_cb {
struct irq_work work;
struct i915_sw_fence *fence;
- void (*hook)(struct i915_request *rq, struct dma_fence *signal);
struct i915_request *signal;
};
@@ -180,17 +179,6 @@ static void irq_execute_cb(struct irq_work *wrk)
kmem_cache_free(global.slab_execute_cbs, cb);
}
-static void irq_execute_cb_hook(struct irq_work *wrk)
-{
- struct execute_cb *cb = container_of(wrk, typeof(*cb), work);
-
- cb->hook(container_of(cb->fence, struct i915_request, submit),
- &cb->signal->fence);
- i915_request_put(cb->signal);
-
- irq_execute_cb(wrk);
-}
-
static __always_inline void
__notify_execute_cb(struct i915_request *rq, bool (*fn)(struct irq_work *wrk))
{
@@ -517,17 +505,12 @@ static bool __request_in_flight(const struct i915_request *signal)
static int
__await_execution(struct i915_request *rq,
struct i915_request *signal,
- void (*hook)(struct i915_request *rq,
- struct dma_fence *signal),
gfp_t gfp)
{
struct execute_cb *cb;
- if (i915_request_is_active(signal)) {
- if (hook)
- hook(rq, &signal->fence);
+ if (i915_request_is_active(signal))
return 0;
- }
cb = kmem_cache_alloc(global.slab_execute_cbs, gfp);
if (!cb)
@@ -537,12 +520,6 @@ __await_execution(struct i915_request *rq,
i915_sw_fence_await(cb->fence);
init_irq_work(&cb->work, irq_execute_cb);
- if (hook) {
- cb->hook = hook;
- cb->signal = i915_request_get(signal);
- cb->work.func = irq_execute_cb_hook;
- }
-
/*
* Register the callback first, then see if the signaler is already
* active. This ensures that if we race with the
@@ -1253,7 +1230,7 @@ emit_semaphore_wait(struct i915_request *to,
goto await_fence;
/* Only submit our spinner after the signaler is running! */
- if (__await_execution(to, from, NULL, gfp))
+ if (__await_execution(to, from, gfp))
goto await_fence;
if (__emit_semaphore_wait(to, from, from->fence.seqno))
@@ -1284,16 +1261,14 @@ static int intel_timeline_sync_set_start(struct intel_timeline *tl,
static int
__i915_request_await_execution(struct i915_request *to,
- struct i915_request *from,
- void (*hook)(struct i915_request *rq,
- struct dma_fence *signal))
+ struct i915_request *from)
{
int err;
GEM_BUG_ON(intel_context_is_barrier(from->context));
/* Submit both requests at the same time */
- err = __await_execution(to, from, hook, I915_FENCE_GFP);
+ err = __await_execution(to, from, I915_FENCE_GFP);
if (err)
return err;
@@ -1406,9 +1381,7 @@ i915_request_await_external(struct i915_request *rq, struct dma_fence *fence)
int
i915_request_await_execution(struct i915_request *rq,
- struct dma_fence *fence,
- void (*hook)(struct i915_request *rq,
- struct dma_fence *signal))
+ struct dma_fence *fence)
{
struct dma_fence **child = &fence;
unsigned int nchild = 1;
@@ -1441,8 +1414,7 @@ i915_request_await_execution(struct i915_request *rq,
if (dma_fence_is_i915(fence))
ret = __i915_request_await_execution(rq,
- to_request(fence),
- hook);
+ to_request(fence));
else
ret = i915_request_await_external(rq, fence);
if (ret < 0)
@@ -1468,7 +1440,7 @@ await_request_submit(struct i915_request *to, struct i915_request *from)
&from->submit,
I915_FENCE_GFP);
else
- return __i915_request_await_execution(to, from, NULL);
+ return __i915_request_await_execution(to, from);
}
static int
diff --git a/drivers/gpu/drm/i915/i915_request.h b/drivers/gpu/drm/i915/i915_request.h
index 270f6cd37650c..63b087a7f5707 100644
--- a/drivers/gpu/drm/i915/i915_request.h
+++ b/drivers/gpu/drm/i915/i915_request.h
@@ -352,9 +352,7 @@ int i915_request_await_object(struct i915_request *to,
int i915_request_await_dma_fence(struct i915_request *rq,
struct dma_fence *fence);
int i915_request_await_execution(struct i915_request *rq,
- struct dma_fence *fence,
- void (*hook)(struct i915_request *rq,
- struct dma_fence *signal));
+ struct dma_fence *fence);
void i915_request_add(struct i915_request *rq);
--
2.31.1
next prev parent reply other threads:[~2021-05-27 16:27 UTC|newest]
Thread overview: 47+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-05-27 16:26 [PATCH 00/29] drm/i915/gem: ioctl clean-ups (v6) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 01/29] drm/i915: Drop I915_CONTEXT_PARAM_RINGSIZE Jason Ekstrand
2021-05-27 16:26 ` [PATCH 02/29] drm/i915: Stop storing the ring size in the ring pointer (v2) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 03/29] drm/i915: Drop I915_CONTEXT_PARAM_NO_ZEROMAP Jason Ekstrand
2021-05-27 16:26 ` [PATCH 04/29] drm/i915/gem: Set the watchdog timeout directly in intel_context_set_gem (v2) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 05/29] drm/i915/gem: Return void from context_apply_all Jason Ekstrand
2021-05-27 16:26 ` [PATCH 06/29] drm/i915: Drop the CONTEXT_CLONE API (v2) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 07/29] drm/i915: Implement SINGLE_TIMELINE with a syncobj (v4) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 08/29] drm/i915: Drop getparam support for I915_CONTEXT_PARAM_ENGINES Jason Ekstrand
2021-05-27 16:26 ` [PATCH 09/29] drm/i915/gem: Disallow bonding of virtual engines (v3) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 10/29] drm/i915/gem: Remove engine auto-magic with FENCE_SUBMIT (v2) Jason Ekstrand
2021-05-27 16:26 ` Jason Ekstrand [this message]
2021-05-27 16:26 ` [PATCH 12/29] drm/i915/gem: Disallow creating contexts with too many engines Jason Ekstrand
2021-05-27 16:26 ` [PATCH 13/29] drm/i915: Stop manually RCU banging in reset_stats_ioctl (v2) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 14/29] drm/i915/gem: Add a separate validate_priority helper Jason Ekstrand
2021-05-27 16:26 ` [PATCH 15/29] drm/i915: Add gem/i915_gem_context.h to the docs Jason Ekstrand
2021-05-27 16:26 ` [PATCH 16/29] drm/i915/gem: Add an intermediate proto_context struct Jason Ekstrand
2021-05-31 8:47 ` [Intel-gfx] " Daniel Vetter
2021-06-02 21:24 ` Jason Ekstrand
2021-06-03 7:09 ` Daniel Vetter
2021-05-27 16:26 ` [PATCH 17/29] drm/i915/gem: Rework error handling in default_engines Jason Ekstrand
2021-05-27 16:26 ` [PATCH 18/29] drm/i915/gem: Optionally set SSEU in intel_context_set_gem Jason Ekstrand
2021-05-31 8:48 ` Daniel Vetter
2021-06-02 22:05 ` Jason Ekstrand
2021-05-27 16:26 ` [PATCH 19/29] drm/i915: Add an i915_gem_vm_lookup helper Jason Ekstrand
2021-05-27 18:30 ` Daniel Vetter
2021-05-27 16:26 ` [PATCH 20/29] drm/i915/gem: Make an alignment check more sensible Jason Ekstrand
2021-05-27 18:31 ` Daniel Vetter
2021-05-27 16:26 ` [PATCH 21/29] drm/i915/gem: Use the proto-context to handle create parameters (v2) Jason Ekstrand
2021-05-31 9:11 ` Daniel Vetter
2021-06-02 22:23 ` Jason Ekstrand
2021-06-03 7:32 ` Daniel Vetter
2021-06-03 16:29 ` Jason Ekstrand
2021-05-27 16:26 ` [PATCH 22/29] drm/i915/gem: Return an error ptr from context_lookup Jason Ekstrand
2021-05-27 16:26 ` [PATCH 23/29] drm/i915/gt: Drop i915_address_space::file (v2) Jason Ekstrand
2021-05-27 16:26 ` [PATCH 24/29] drm/i915/gem: Delay context creation Jason Ekstrand
2021-05-31 9:50 ` Daniel Vetter
2021-06-02 22:55 ` Jason Ekstrand
2021-06-03 7:26 ` Daniel Vetter
2021-05-27 16:26 ` [PATCH 25/29] drm/i915/gem: Don't allow changing the VM on running contexts (v2) Jason Ekstrand
2021-05-31 14:08 ` Daniel Vetter
2021-05-27 16:26 ` [PATCH 26/29] drm/i915/gem: Don't allow changing the engine set " Jason Ekstrand
2021-05-31 13:58 ` [Intel-gfx] " Daniel Vetter
2021-05-27 16:26 ` [PATCH 27/29] drm/i915/selftests: Take a VM in kernel_context() Jason Ekstrand
2021-05-27 16:26 ` [PATCH 28/29] i915/gem/selftests: Assign the VM at context creation in igt_shared_ctx_exec Jason Ekstrand
2021-05-27 16:26 ` [PATCH 29/29] drm/i915/gem: Roll all of context creation together Jason Ekstrand
2021-05-31 14:18 ` [Intel-gfx] " Daniel Vetter
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210527162650.1182544-12-jason@jlekstrand.net \
--to=jason@jlekstrand.net \
--cc=daniel.vetter@ffwll.ch \
--cc=dri-devel@lists.freedesktop.org \
--cc=intel-gfx@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).