From: Daniele Ceraolo Spurio <daniele.ceraolospurio@intel.com> To: Matthew Brost <matthew.brost@intel.com>, <igt-dev@lists.freedesktop.org> Cc: <intel-gfx@lists.freedesktop.org> Subject: Re: [Intel-gfx] [PATCH i-g-t 1/3] i915/gem_exec_schedule: Make gem_exec_schedule understand static priority mapping Date: Fri, 13 Aug 2021 16:24:37 -0700 [thread overview] Message-ID: <0f42bc7d-2a3a-ffcc-a674-8d028093987a@intel.com> (raw) In-Reply-To: <20210804012303.158392-2-matthew.brost@intel.com> On 8/3/2021 6:23 PM, Matthew Brost wrote: > The i915 currently has 2k visible priority levels which are currently > unique. This is changing to statically map these 2k levels into 3 > buckets: > > low: < 0 > mid: 0 > high: > 0 > > Update gem_exec_schedule to understand this. This entails updating > promotion test to use 3 levels that will map into different buckets and > also add bit of delay after releasing a cork beforing completing the > spinners. This needs a line about why we add the delay, something like "to give time to the i915 scheduler to process the fence release and queue the requests" or something. BTW, any reason not to just add the delay unconditionally in unplug_show_queue, instead of only in one test? Other tests might suffer from the same problem even if they're not hitting it at the moment. Daniele > > Also skip any tests that rely on having more than 3 priority levels. > > v2: Add a delay between starting releasing spinner and cork in > promotion, add local define for static mapping engine info > > Signed-off-by: Matthew Brost <matthew.brost@intel.com> > --- > lib/i915/gem_scheduler.c | 14 ++++++++ > lib/i915/gem_scheduler.h | 1 + > lib/i915/i915_drm_local.h | 10 ++++++ > tests/i915/gem_exec_schedule.c | 62 +++++++++++++++++++++------------- > 4 files changed, 63 insertions(+), 24 deletions(-) > > diff --git a/lib/i915/gem_scheduler.c b/lib/i915/gem_scheduler.c > index cdddf42ad..d006b8676 100644 > --- a/lib/i915/gem_scheduler.c > +++ b/lib/i915/gem_scheduler.c > @@ -28,6 +28,7 @@ > #include "igt_core.h" > #include "ioctl_wrappers.h" > > +#include "i915/i915_drm_local.h" > #include "i915/gem_scheduler.h" > #include "i915/gem_submission.h" > > @@ -90,6 +91,19 @@ bool gem_scheduler_has_ctx_priority(int fd) > I915_SCHEDULER_CAP_PRIORITY; > } > > +/** > + * gem_scheduler_has_static_priority: > + * @fd: open i915 drm file descriptor > + * > + * Feature test macro to query whether the driver supports priority assigned > + * from user space are statically mapping into 3 buckets. > + */ > +bool gem_scheduler_has_static_priority(int fd) > +{ > + return gem_scheduler_capability(fd) & > + I915_SCHEDULER_CAP_STATIC_PRIORITY_MAP; > +} > + > /** > * gem_scheduler_has_preemption: > * @fd: open i915 drm file descriptor > diff --git a/lib/i915/gem_scheduler.h b/lib/i915/gem_scheduler.h > index d43e84bd2..b00804f70 100644 > --- a/lib/i915/gem_scheduler.h > +++ b/lib/i915/gem_scheduler.h > @@ -29,6 +29,7 @@ > unsigned gem_scheduler_capability(int fd); > bool gem_scheduler_enabled(int fd); > bool gem_scheduler_has_ctx_priority(int fd); > +bool gem_scheduler_has_static_priority(int fd); > bool gem_scheduler_has_preemption(int fd); > bool gem_scheduler_has_semaphores(int fd); > bool gem_scheduler_has_engine_busy_stats(int fd); > diff --git a/lib/i915/i915_drm_local.h b/lib/i915/i915_drm_local.h > index dd646aedf..a1527ff21 100644 > --- a/lib/i915/i915_drm_local.h > +++ b/lib/i915/i915_drm_local.h > @@ -20,6 +20,16 @@ extern "C" { > * clean these up when kernel uapi headers are sync'd. > */ > > +/* > + * Indicates the 2k user priority levels are statically mapped into 3 buckets as > + * follows: > + * > + * -1k to -1 Low priority > + * 0 Normal priority > + * 1 to 1k Highest priority > + */ > +#define I915_SCHEDULER_CAP_STATIC_PRIORITY_MAP (1ul << 5) > + > #if defined(__cplusplus) > } > #endif > diff --git a/tests/i915/gem_exec_schedule.c b/tests/i915/gem_exec_schedule.c > index e5fb45982..bb9fb6c14 100644 > --- a/tests/i915/gem_exec_schedule.c > +++ b/tests/i915/gem_exec_schedule.c > @@ -199,7 +199,8 @@ create_highest_priority(int fd, const intel_ctx_cfg_t *cfg) > > static void unplug_show_queue(int fd, struct igt_cork *c, > const intel_ctx_cfg_t *cfg, > - unsigned int engine) > + unsigned int engine, > + unsigned usec_delay) > { > igt_spin_t *spin[MAX_ELSP_QLEN]; > int max = MAX_ELSP_QLEN; > @@ -216,6 +217,7 @@ static void unplug_show_queue(int fd, struct igt_cork *c, > > igt_cork_unplug(c); /* batches will now be queued on the engine */ > igt_debugfs_dump(fd, "i915_engine_info"); > + usleep(usec_delay); > > for (int n = 0; n < max; n++) > igt_spin_free(fd, spin[n]); > @@ -237,7 +239,7 @@ static void fifo(int fd, const intel_ctx_t *ctx, unsigned ring) > store_dword_fenced(fd, ctx, ring, scratch, 0, 1, fence, 0); > store_dword_fenced(fd, ctx, ring, scratch, 0, 2, fence, 0); > > - unplug_show_queue(fd, &cork, &ctx->cfg, ring); > + unplug_show_queue(fd, &cork, &ctx->cfg, ring, 0); > close(fence); > > result = __sync_read_u32(fd, scratch, 0); > @@ -298,7 +300,7 @@ static void implicit_rw(int i915, const intel_ctx_t *ctx, unsigned int ring, > ring, scratch, 0, ring, > fence, I915_GEM_DOMAIN_RENDER); > > - unplug_show_queue(i915, &cork, &ctx->cfg, ring); > + unplug_show_queue(i915, &cork, &ctx->cfg, ring, 0); > close(fence); > > result = __sync_read_u32(i915, scratch, 0); > @@ -355,7 +357,7 @@ static void independent(int fd, const intel_ctx_t *ctx, unsigned int engine, > /* Same priority, but different timeline (as different engine) */ > batch = __store_dword(fd, ctx, engine, scratch, 0, engine, 0, fence, 0); > > - unplug_show_queue(fd, &cork, &ctx->cfg, engine); > + unplug_show_queue(fd, &cork, &ctx->cfg, engine, 0); > close(fence); > > gem_sync(fd, batch); > @@ -1326,7 +1328,7 @@ static void reorder(int fd, const intel_ctx_cfg_t *cfg, > store_dword_fenced(fd, ctx[LO], ring, scratch, 0, ctx[LO]->id, fence, 0); > store_dword_fenced(fd, ctx[HI], ring, scratch, 0, ctx[HI]->id, fence, 0); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > close(fence); > > result = __sync_read_u32(fd, scratch, 0); > @@ -1353,10 +1355,10 @@ static void promotion(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > gem_context_set_priority(fd, ctx[LO]->id, MIN_PRIO); > > ctx[HI] = intel_ctx_create(fd, cfg); > - gem_context_set_priority(fd, ctx[HI]->id, 0); > + gem_context_set_priority(fd, ctx[HI]->id, MAX_PRIO); > > ctx[NOISE] = intel_ctx_create(fd, cfg); > - gem_context_set_priority(fd, ctx[NOISE]->id, MIN_PRIO/2); > + gem_context_set_priority(fd, ctx[NOISE]->id, 0); > > result = gem_create(fd, 4096); > dep = gem_create(fd, 4096); > @@ -1377,7 +1379,7 @@ static void promotion(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > > store_dword(fd, ctx[HI], ring, result, 0, ctx[HI]->id, 0); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 250000); > close(fence); > > dep_read = __sync_read_u32(fd, dep, 0); > @@ -1893,7 +1895,7 @@ static void deep(int fd, const intel_ctx_cfg_t *cfg, > igt_info("Second deptree: %d requests [%.3fs]\n", > n * XS, 1e-9*igt_nsec_elapsed(&tv)); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > gem_close(fd, plug); > igt_require(expected); /* too slow */ > > @@ -1962,7 +1964,7 @@ static void wide(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > igt_info("Submitted %d requests over %d contexts in %.1fms\n", > count, MAX_CONTEXTS, igt_nsec_elapsed(&tv) * 1e-6); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > close(fence); > > __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); > @@ -2067,7 +2069,7 @@ static void reorder_wide(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > intel_ctx_destroy(fd, tmp_ctx); > } > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > close(fence); > > __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); > @@ -2963,19 +2965,25 @@ igt_main > test_each_engine_store("preempt-other-chain", fd, ctx, e) > preempt_other(fd, &ctx->cfg, e->flags, CHAIN); > > - test_each_engine_store("preempt-queue", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, 0); > + test_each_engine_store("preempt-engines", fd, ctx, e) > + preempt_engines(fd, e, 0); > > - test_each_engine_store("preempt-queue-chain", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, CHAIN); > - test_each_engine_store("preempt-queue-contexts", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS); > + igt_subtest_group { > + igt_fixture { > + igt_require(!gem_scheduler_has_static_priority(fd)); > + } > > - test_each_engine_store("preempt-queue-contexts-chain", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS | CHAIN); > + test_each_engine_store("preempt-queue", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, 0); > > - test_each_engine_store("preempt-engines", fd, ctx, e) > - preempt_engines(fd, e, 0); > + test_each_engine_store("preempt-queue-chain", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, CHAIN); > + test_each_engine_store("preempt-queue-contexts", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS); > + > + test_each_engine_store("preempt-queue-contexts-chain", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS | CHAIN); > + } > > igt_subtest_group { > igt_hang_t hang; > @@ -3017,11 +3025,17 @@ igt_main > test_each_engine_store("wide", fd, ctx, e) > wide(fd, &ctx->cfg, e->flags); > > - test_each_engine_store("reorder-wide", fd, ctx, e) > - reorder_wide(fd, &ctx->cfg, e->flags); > - > test_each_engine_store("smoketest", fd, ctx, e) > smoketest(fd, &ctx->cfg, e->flags, 5); > + > + igt_subtest_group { > + igt_fixture { > + igt_require(!gem_scheduler_has_static_priority(fd)); > + } > + > + test_each_engine_store("reorder-wide", fd, ctx, e) > + reorder_wide(fd, &ctx->cfg, e->flags); > + } > } > > igt_subtest_group {
WARNING: multiple messages have this Message-ID (diff)
From: Daniele Ceraolo Spurio <daniele.ceraolospurio@intel.com> To: Matthew Brost <matthew.brost@intel.com>, igt-dev@lists.freedesktop.org Cc: intel-gfx@lists.freedesktop.org Subject: Re: [igt-dev] [Intel-gfx] [PATCH i-g-t 1/3] i915/gem_exec_schedule: Make gem_exec_schedule understand static priority mapping Date: Fri, 13 Aug 2021 16:24:37 -0700 [thread overview] Message-ID: <0f42bc7d-2a3a-ffcc-a674-8d028093987a@intel.com> (raw) In-Reply-To: <20210804012303.158392-2-matthew.brost@intel.com> On 8/3/2021 6:23 PM, Matthew Brost wrote: > The i915 currently has 2k visible priority levels which are currently > unique. This is changing to statically map these 2k levels into 3 > buckets: > > low: < 0 > mid: 0 > high: > 0 > > Update gem_exec_schedule to understand this. This entails updating > promotion test to use 3 levels that will map into different buckets and > also add bit of delay after releasing a cork beforing completing the > spinners. This needs a line about why we add the delay, something like "to give time to the i915 scheduler to process the fence release and queue the requests" or something. BTW, any reason not to just add the delay unconditionally in unplug_show_queue, instead of only in one test? Other tests might suffer from the same problem even if they're not hitting it at the moment. Daniele > > Also skip any tests that rely on having more than 3 priority levels. > > v2: Add a delay between starting releasing spinner and cork in > promotion, add local define for static mapping engine info > > Signed-off-by: Matthew Brost <matthew.brost@intel.com> > --- > lib/i915/gem_scheduler.c | 14 ++++++++ > lib/i915/gem_scheduler.h | 1 + > lib/i915/i915_drm_local.h | 10 ++++++ > tests/i915/gem_exec_schedule.c | 62 +++++++++++++++++++++------------- > 4 files changed, 63 insertions(+), 24 deletions(-) > > diff --git a/lib/i915/gem_scheduler.c b/lib/i915/gem_scheduler.c > index cdddf42ad..d006b8676 100644 > --- a/lib/i915/gem_scheduler.c > +++ b/lib/i915/gem_scheduler.c > @@ -28,6 +28,7 @@ > #include "igt_core.h" > #include "ioctl_wrappers.h" > > +#include "i915/i915_drm_local.h" > #include "i915/gem_scheduler.h" > #include "i915/gem_submission.h" > > @@ -90,6 +91,19 @@ bool gem_scheduler_has_ctx_priority(int fd) > I915_SCHEDULER_CAP_PRIORITY; > } > > +/** > + * gem_scheduler_has_static_priority: > + * @fd: open i915 drm file descriptor > + * > + * Feature test macro to query whether the driver supports priority assigned > + * from user space are statically mapping into 3 buckets. > + */ > +bool gem_scheduler_has_static_priority(int fd) > +{ > + return gem_scheduler_capability(fd) & > + I915_SCHEDULER_CAP_STATIC_PRIORITY_MAP; > +} > + > /** > * gem_scheduler_has_preemption: > * @fd: open i915 drm file descriptor > diff --git a/lib/i915/gem_scheduler.h b/lib/i915/gem_scheduler.h > index d43e84bd2..b00804f70 100644 > --- a/lib/i915/gem_scheduler.h > +++ b/lib/i915/gem_scheduler.h > @@ -29,6 +29,7 @@ > unsigned gem_scheduler_capability(int fd); > bool gem_scheduler_enabled(int fd); > bool gem_scheduler_has_ctx_priority(int fd); > +bool gem_scheduler_has_static_priority(int fd); > bool gem_scheduler_has_preemption(int fd); > bool gem_scheduler_has_semaphores(int fd); > bool gem_scheduler_has_engine_busy_stats(int fd); > diff --git a/lib/i915/i915_drm_local.h b/lib/i915/i915_drm_local.h > index dd646aedf..a1527ff21 100644 > --- a/lib/i915/i915_drm_local.h > +++ b/lib/i915/i915_drm_local.h > @@ -20,6 +20,16 @@ extern "C" { > * clean these up when kernel uapi headers are sync'd. > */ > > +/* > + * Indicates the 2k user priority levels are statically mapped into 3 buckets as > + * follows: > + * > + * -1k to -1 Low priority > + * 0 Normal priority > + * 1 to 1k Highest priority > + */ > +#define I915_SCHEDULER_CAP_STATIC_PRIORITY_MAP (1ul << 5) > + > #if defined(__cplusplus) > } > #endif > diff --git a/tests/i915/gem_exec_schedule.c b/tests/i915/gem_exec_schedule.c > index e5fb45982..bb9fb6c14 100644 > --- a/tests/i915/gem_exec_schedule.c > +++ b/tests/i915/gem_exec_schedule.c > @@ -199,7 +199,8 @@ create_highest_priority(int fd, const intel_ctx_cfg_t *cfg) > > static void unplug_show_queue(int fd, struct igt_cork *c, > const intel_ctx_cfg_t *cfg, > - unsigned int engine) > + unsigned int engine, > + unsigned usec_delay) > { > igt_spin_t *spin[MAX_ELSP_QLEN]; > int max = MAX_ELSP_QLEN; > @@ -216,6 +217,7 @@ static void unplug_show_queue(int fd, struct igt_cork *c, > > igt_cork_unplug(c); /* batches will now be queued on the engine */ > igt_debugfs_dump(fd, "i915_engine_info"); > + usleep(usec_delay); > > for (int n = 0; n < max; n++) > igt_spin_free(fd, spin[n]); > @@ -237,7 +239,7 @@ static void fifo(int fd, const intel_ctx_t *ctx, unsigned ring) > store_dword_fenced(fd, ctx, ring, scratch, 0, 1, fence, 0); > store_dword_fenced(fd, ctx, ring, scratch, 0, 2, fence, 0); > > - unplug_show_queue(fd, &cork, &ctx->cfg, ring); > + unplug_show_queue(fd, &cork, &ctx->cfg, ring, 0); > close(fence); > > result = __sync_read_u32(fd, scratch, 0); > @@ -298,7 +300,7 @@ static void implicit_rw(int i915, const intel_ctx_t *ctx, unsigned int ring, > ring, scratch, 0, ring, > fence, I915_GEM_DOMAIN_RENDER); > > - unplug_show_queue(i915, &cork, &ctx->cfg, ring); > + unplug_show_queue(i915, &cork, &ctx->cfg, ring, 0); > close(fence); > > result = __sync_read_u32(i915, scratch, 0); > @@ -355,7 +357,7 @@ static void independent(int fd, const intel_ctx_t *ctx, unsigned int engine, > /* Same priority, but different timeline (as different engine) */ > batch = __store_dword(fd, ctx, engine, scratch, 0, engine, 0, fence, 0); > > - unplug_show_queue(fd, &cork, &ctx->cfg, engine); > + unplug_show_queue(fd, &cork, &ctx->cfg, engine, 0); > close(fence); > > gem_sync(fd, batch); > @@ -1326,7 +1328,7 @@ static void reorder(int fd, const intel_ctx_cfg_t *cfg, > store_dword_fenced(fd, ctx[LO], ring, scratch, 0, ctx[LO]->id, fence, 0); > store_dword_fenced(fd, ctx[HI], ring, scratch, 0, ctx[HI]->id, fence, 0); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > close(fence); > > result = __sync_read_u32(fd, scratch, 0); > @@ -1353,10 +1355,10 @@ static void promotion(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > gem_context_set_priority(fd, ctx[LO]->id, MIN_PRIO); > > ctx[HI] = intel_ctx_create(fd, cfg); > - gem_context_set_priority(fd, ctx[HI]->id, 0); > + gem_context_set_priority(fd, ctx[HI]->id, MAX_PRIO); > > ctx[NOISE] = intel_ctx_create(fd, cfg); > - gem_context_set_priority(fd, ctx[NOISE]->id, MIN_PRIO/2); > + gem_context_set_priority(fd, ctx[NOISE]->id, 0); > > result = gem_create(fd, 4096); > dep = gem_create(fd, 4096); > @@ -1377,7 +1379,7 @@ static void promotion(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > > store_dword(fd, ctx[HI], ring, result, 0, ctx[HI]->id, 0); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 250000); > close(fence); > > dep_read = __sync_read_u32(fd, dep, 0); > @@ -1893,7 +1895,7 @@ static void deep(int fd, const intel_ctx_cfg_t *cfg, > igt_info("Second deptree: %d requests [%.3fs]\n", > n * XS, 1e-9*igt_nsec_elapsed(&tv)); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > gem_close(fd, plug); > igt_require(expected); /* too slow */ > > @@ -1962,7 +1964,7 @@ static void wide(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > igt_info("Submitted %d requests over %d contexts in %.1fms\n", > count, MAX_CONTEXTS, igt_nsec_elapsed(&tv) * 1e-6); > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > close(fence); > > __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); > @@ -2067,7 +2069,7 @@ static void reorder_wide(int fd, const intel_ctx_cfg_t *cfg, unsigned ring) > intel_ctx_destroy(fd, tmp_ctx); > } > > - unplug_show_queue(fd, &cork, cfg, ring); > + unplug_show_queue(fd, &cork, cfg, ring, 0); > close(fence); > > __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); > @@ -2963,19 +2965,25 @@ igt_main > test_each_engine_store("preempt-other-chain", fd, ctx, e) > preempt_other(fd, &ctx->cfg, e->flags, CHAIN); > > - test_each_engine_store("preempt-queue", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, 0); > + test_each_engine_store("preempt-engines", fd, ctx, e) > + preempt_engines(fd, e, 0); > > - test_each_engine_store("preempt-queue-chain", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, CHAIN); > - test_each_engine_store("preempt-queue-contexts", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS); > + igt_subtest_group { > + igt_fixture { > + igt_require(!gem_scheduler_has_static_priority(fd)); > + } > > - test_each_engine_store("preempt-queue-contexts-chain", fd, ctx, e) > - preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS | CHAIN); > + test_each_engine_store("preempt-queue", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, 0); > > - test_each_engine_store("preempt-engines", fd, ctx, e) > - preempt_engines(fd, e, 0); > + test_each_engine_store("preempt-queue-chain", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, CHAIN); > + test_each_engine_store("preempt-queue-contexts", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS); > + > + test_each_engine_store("preempt-queue-contexts-chain", fd, ctx, e) > + preempt_queue(fd, &ctx->cfg, e->flags, CONTEXTS | CHAIN); > + } > > igt_subtest_group { > igt_hang_t hang; > @@ -3017,11 +3025,17 @@ igt_main > test_each_engine_store("wide", fd, ctx, e) > wide(fd, &ctx->cfg, e->flags); > > - test_each_engine_store("reorder-wide", fd, ctx, e) > - reorder_wide(fd, &ctx->cfg, e->flags); > - > test_each_engine_store("smoketest", fd, ctx, e) > smoketest(fd, &ctx->cfg, e->flags, 5); > + > + igt_subtest_group { > + igt_fixture { > + igt_require(!gem_scheduler_has_static_priority(fd)); > + } > + > + test_each_engine_store("reorder-wide", fd, ctx, e) > + reorder_wide(fd, &ctx->cfg, e->flags); > + } > } > > igt_subtest_group {
next prev parent reply other threads:[~2021-08-13 23:24 UTC|newest] Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top 2021-08-04 1:23 [Intel-gfx] [PATCH i-g-t 0/3] IGT fixes for priority management + capture with GuC submission Matthew Brost 2021-08-04 1:23 ` [igt-dev] " Matthew Brost 2021-08-04 1:23 ` [Intel-gfx] [PATCH i-g-t 1/3] i915/gem_exec_schedule: Make gem_exec_schedule understand static priority mapping Matthew Brost 2021-08-04 1:23 ` Matthew Brost 2021-08-13 23:24 ` Daniele Ceraolo Spurio [this message] 2021-08-13 23:24 ` [igt-dev] " Daniele Ceraolo Spurio 2021-08-16 16:39 ` Matthew Brost 2021-08-16 16:39 ` [igt-dev] " Matthew Brost 2021-08-04 1:23 ` [Intel-gfx] [PATCH i-g-t 2/3] i915/gem_ctx_shared: Make gem_ctx_shared " Matthew Brost 2021-08-04 1:23 ` [igt-dev] " Matthew Brost 2021-08-04 1:23 ` [Intel-gfx] [PATCH i-g-t 3/3] i915/gem_exec_capture: Update to support GuC based resets Matthew Brost 2021-08-04 1:23 ` Matthew Brost 2021-08-14 0:28 ` Daniele Ceraolo Spurio 2021-08-14 0:28 ` [igt-dev] " Daniele Ceraolo Spurio 2021-08-04 1:47 ` [igt-dev] ✓ Fi.CI.BAT: success for IGT fixes for priority management + capture with GuC submission Patchwork 2021-08-05 11:46 ` [igt-dev] ✓ Fi.CI.IGT: " Patchwork
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=0f42bc7d-2a3a-ffcc-a674-8d028093987a@intel.com \ --to=daniele.ceraolospurio@intel.com \ --cc=igt-dev@lists.freedesktop.org \ --cc=intel-gfx@lists.freedesktop.org \ --cc=matthew.brost@intel.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.