* [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks
@ 2020-03-03 9:20 Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 2/4] drm/i915/gem: Consolidate ctx->engines[] release Chris Wilson
` (6 more replies)
0 siblings, 7 replies; 8+ messages in thread
From: Chris Wilson @ 2020-03-03 9:20 UTC (permalink / raw)
To: intel-gfx
Update locations for
./drivers/gpu/drm/i915/i915_vma.h:1: warning: 'Virtual Memory Address' not found
./drivers/gpu/drm/i915/i915_gem_gtt.c:1: warning: 'Global GTT views' not found
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
---
Documentation/gpu/i915.rst | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/Documentation/gpu/i915.rst b/Documentation/gpu/i915.rst
index cc74e24ca3b5..f6d363b6756e 100644
--- a/Documentation/gpu/i915.rst
+++ b/Documentation/gpu/i915.rst
@@ -332,7 +332,7 @@ This process is dubbed relocation.
GEM BO Management Implementation Details
----------------------------------------
-.. kernel-doc:: drivers/gpu/drm/i915/i915_vma.h
+.. kernel-doc:: drivers/gpu/drm/i915/i915_vma_types.h
:doc: Virtual Memory Address
Buffer Object Eviction
@@ -382,7 +382,7 @@ Logical Rings, Logical Ring Contexts and Execlists
Global GTT views
----------------
-.. kernel-doc:: drivers/gpu/drm/i915/i915_gem_gtt.c
+.. kernel-doc:: drivers/gpu/drm/i915/i915_vma_types.h
:doc: Global GTT views
.. kernel-doc:: drivers/gpu/drm/i915/i915_gem_gtt.c
--
2.25.1
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Intel-gfx] [PATCH 2/4] drm/i915/gem: Consolidate ctx->engines[] release
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
@ 2020-03-03 9:20 ` Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 3/4] drm/i915/gt: Prevent allocation on a banned context Chris Wilson
` (5 subsequent siblings)
6 siblings, 0 replies; 8+ messages in thread
From: Chris Wilson @ 2020-03-03 9:20 UTC (permalink / raw)
To: intel-gfx
Use the same engine_idle_release() routine for cleaning all old
ctx->engine[] state, closing any potential races with concurrent execbuf
submission.
v2ish: Use the ce->pin_count to close the execbuf gap.
Closes: https://gitlab.freedesktop.org/drm/intel/issues/1241
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
---
drivers/gpu/drm/i915/gem/i915_gem_context.c | 193 +++++++++---------
drivers/gpu/drm/i915/gem/i915_gem_context.h | 1 -
.../gpu/drm/i915/gem/selftests/mock_context.c | 3 +
3 files changed, 105 insertions(+), 92 deletions(-)
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c b/drivers/gpu/drm/i915/gem/i915_gem_context.c
index e525ead073f7..cb6b6be48978 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_context.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c
@@ -242,7 +242,6 @@ static void __free_engines(struct i915_gem_engines *e, unsigned int count)
if (!e->engines[count])
continue;
- RCU_INIT_POINTER(e->engines[count]->gem_context, NULL);
intel_context_put(e->engines[count]);
}
kfree(e);
@@ -255,7 +254,11 @@ static void free_engines(struct i915_gem_engines *e)
static void free_engines_rcu(struct rcu_head *rcu)
{
- free_engines(container_of(rcu, struct i915_gem_engines, rcu));
+ struct i915_gem_engines *engines =
+ container_of(rcu, struct i915_gem_engines, rcu);
+
+ i915_sw_fence_fini(&engines->fence);
+ free_engines(engines);
}
static struct i915_gem_engines *default_engines(struct i915_gem_context *ctx)
@@ -269,8 +272,6 @@ static struct i915_gem_engines *default_engines(struct i915_gem_context *ctx)
if (!e)
return ERR_PTR(-ENOMEM);
- e->ctx = ctx;
-
for_each_engine(engine, gt, id) {
struct intel_context *ce;
@@ -304,7 +305,6 @@ static void i915_gem_context_free(struct i915_gem_context *ctx)
list_del(&ctx->link);
spin_unlock(&ctx->i915->gem.contexts.lock);
- free_engines(rcu_access_pointer(ctx->engines));
mutex_destroy(&ctx->engines_mutex);
if (ctx->timeline)
@@ -491,30 +491,104 @@ static void kill_engines(struct i915_gem_engines *engines)
static void kill_stale_engines(struct i915_gem_context *ctx)
{
struct i915_gem_engines *pos, *next;
- unsigned long flags;
- spin_lock_irqsave(&ctx->stale.lock, flags);
+ spin_lock_irq(&ctx->stale.lock);
+ GEM_BUG_ON(!i915_gem_context_is_closed(ctx));
list_for_each_entry_safe(pos, next, &ctx->stale.engines, link) {
- if (!i915_sw_fence_await(&pos->fence))
+ if (!i915_sw_fence_await(&pos->fence)) {
+ list_del_init(&pos->link);
continue;
+ }
- spin_unlock_irqrestore(&ctx->stale.lock, flags);
+ spin_unlock_irq(&ctx->stale.lock);
kill_engines(pos);
- spin_lock_irqsave(&ctx->stale.lock, flags);
+ spin_lock_irq(&ctx->stale.lock);
+ GEM_BUG_ON(i915_sw_fence_signaled(&pos->fence));
list_safe_reset_next(pos, next, link);
list_del_init(&pos->link); /* decouple from FENCE_COMPLETE */
i915_sw_fence_complete(&pos->fence);
}
- spin_unlock_irqrestore(&ctx->stale.lock, flags);
+ spin_unlock_irq(&ctx->stale.lock);
}
static void kill_context(struct i915_gem_context *ctx)
{
kill_stale_engines(ctx);
- kill_engines(__context_engines_static(ctx));
+}
+
+static int engines_notify(struct i915_sw_fence *fence,
+ enum i915_sw_fence_notify state)
+{
+ struct i915_gem_engines *engines =
+ container_of(fence, typeof(*engines), fence);
+
+ switch (state) {
+ case FENCE_COMPLETE:
+ if (!list_empty(&engines->link)) {
+ struct i915_gem_context *ctx = engines->ctx;
+ unsigned long flags;
+
+ spin_lock_irqsave(&ctx->stale.lock, flags);
+ list_del(&engines->link);
+ spin_unlock_irqrestore(&ctx->stale.lock, flags);
+ }
+ i915_gem_context_put(engines->ctx);
+ break;
+
+ case FENCE_FREE:
+ init_rcu_head(&engines->rcu);
+ call_rcu(&engines->rcu, free_engines_rcu);
+ break;
+ }
+
+ return NOTIFY_DONE;
+}
+
+static void engines_idle_release(struct i915_gem_context *ctx,
+ struct i915_gem_engines *engines)
+{
+ struct i915_gem_engines_iter it;
+ struct intel_context *ce;
+
+ i915_sw_fence_init(&engines->fence, engines_notify);
+ INIT_LIST_HEAD(&engines->link);
+
+ engines->ctx = i915_gem_context_get(ctx);
+
+ for_each_gem_engine(ce, engines, it) {
+ struct dma_fence *fence;
+ int err = 0;
+
+ /* serialises with execbuf */
+ RCU_INIT_POINTER(ce->gem_context, NULL);
+ if (!intel_context_pin_if_active(ce))
+ continue;
+
+ fence = i915_active_fence_get(&ce->timeline->last_request);
+ if (fence) {
+ err = i915_sw_fence_await_dma_fence(&engines->fence,
+ fence, 0,
+ GFP_KERNEL);
+ dma_fence_put(fence);
+ }
+ intel_context_unpin(ce);
+ if (err < 0)
+ goto kill;
+ }
+
+ spin_lock_irq(&ctx->stale.lock);
+ if (!i915_gem_context_is_closed(ctx))
+ list_add_tail(&engines->link, &ctx->stale.engines);
+ spin_unlock_irq(&ctx->stale.lock);
+
+kill:
+ if (list_empty(&engines->link)) /* raced, already closed */
+ kill_engines(engines);
+
+ i915_sw_fence_commit(&engines->fence);
}
static void set_closed_name(struct i915_gem_context *ctx)
@@ -538,11 +612,16 @@ static void context_close(struct i915_gem_context *ctx)
{
struct i915_address_space *vm;
+ /* Flush any concurrent set_engines() */
+ mutex_lock(&ctx->engines_mutex);
+ engines_idle_release(ctx, rcu_replace_pointer(ctx->engines, NULL, 1));
i915_gem_context_set_closed(ctx);
- set_closed_name(ctx);
+ mutex_unlock(&ctx->engines_mutex);
mutex_lock(&ctx->mutex);
+ set_closed_name(ctx);
+
vm = i915_gem_context_vm(ctx);
if (vm)
i915_vm_close(vm);
@@ -1626,77 +1705,6 @@ static const i915_user_extension_fn set_engines__extensions[] = {
[I915_CONTEXT_ENGINES_EXT_BOND] = set_engines__bond,
};
-static int engines_notify(struct i915_sw_fence *fence,
- enum i915_sw_fence_notify state)
-{
- struct i915_gem_engines *engines =
- container_of(fence, typeof(*engines), fence);
-
- switch (state) {
- case FENCE_COMPLETE:
- if (!list_empty(&engines->link)) {
- struct i915_gem_context *ctx = engines->ctx;
- unsigned long flags;
-
- spin_lock_irqsave(&ctx->stale.lock, flags);
- list_del(&engines->link);
- spin_unlock_irqrestore(&ctx->stale.lock, flags);
- }
- break;
-
- case FENCE_FREE:
- init_rcu_head(&engines->rcu);
- call_rcu(&engines->rcu, free_engines_rcu);
- break;
- }
-
- return NOTIFY_DONE;
-}
-
-static void engines_idle_release(struct i915_gem_engines *engines)
-{
- struct i915_gem_engines_iter it;
- struct intel_context *ce;
- unsigned long flags;
-
- GEM_BUG_ON(!engines);
- i915_sw_fence_init(&engines->fence, engines_notify);
-
- INIT_LIST_HEAD(&engines->link);
- spin_lock_irqsave(&engines->ctx->stale.lock, flags);
- if (!i915_gem_context_is_closed(engines->ctx))
- list_add(&engines->link, &engines->ctx->stale.engines);
- spin_unlock_irqrestore(&engines->ctx->stale.lock, flags);
- if (list_empty(&engines->link)) /* raced, already closed */
- goto kill;
-
- for_each_gem_engine(ce, engines, it) {
- struct dma_fence *fence;
- int err;
-
- if (!ce->timeline)
- continue;
-
- fence = i915_active_fence_get(&ce->timeline->last_request);
- if (!fence)
- continue;
-
- err = i915_sw_fence_await_dma_fence(&engines->fence,
- fence, 0,
- GFP_KERNEL);
-
- dma_fence_put(fence);
- if (err < 0)
- goto kill;
- }
- goto out;
-
-kill:
- kill_engines(engines);
-out:
- i915_sw_fence_commit(&engines->fence);
-}
-
static int
set_engines(struct i915_gem_context *ctx,
const struct drm_i915_gem_context_param *args)
@@ -1739,8 +1747,6 @@ set_engines(struct i915_gem_context *ctx,
if (!set.engines)
return -ENOMEM;
- set.engines->ctx = ctx;
-
for (n = 0; n < num_engines; n++) {
struct i915_engine_class_instance ci;
struct intel_engine_cs *engine;
@@ -1793,6 +1799,11 @@ set_engines(struct i915_gem_context *ctx,
replace:
mutex_lock(&ctx->engines_mutex);
+ if (i915_gem_context_is_closed(ctx)) {
+ mutex_unlock(&ctx->engines_mutex);
+ free_engines(set.engines);
+ return -ENOENT;
+ }
if (args->size)
i915_gem_context_set_user_engines(ctx);
else
@@ -1801,7 +1812,7 @@ set_engines(struct i915_gem_context *ctx,
mutex_unlock(&ctx->engines_mutex);
/* Keep track of old engine sets for kill_context() */
- engines_idle_release(set.engines);
+ engines_idle_release(ctx, set.engines);
return 0;
}
@@ -2077,8 +2088,6 @@ static int clone_engines(struct i915_gem_context *dst,
if (!clone)
goto err_unlock;
- clone->ctx = dst;
-
for (n = 0; n < e->num_engines; n++) {
struct intel_engine_cs *engine;
@@ -2121,8 +2130,7 @@ static int clone_engines(struct i915_gem_context *dst,
i915_gem_context_unlock_engines(src);
/* Serialised by constructor */
- free_engines(__context_engines_static(dst));
- RCU_INIT_POINTER(dst->engines, clone);
+ engines_idle_release(dst, rcu_replace_pointer(dst->engines, clone, 1));
if (user_engines)
i915_gem_context_set_user_engines(dst);
else
@@ -2553,6 +2561,9 @@ i915_gem_engines_iter_next(struct i915_gem_engines_iter *it)
const struct i915_gem_engines *e = it->engines;
struct intel_context *ctx;
+ if (unlikely(!e))
+ return NULL;
+
do {
if (it->idx >= e->num_engines)
return NULL;
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.h b/drivers/gpu/drm/i915/gem/i915_gem_context.h
index 3ae61a355d87..57b7ae2893e1 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_context.h
+++ b/drivers/gpu/drm/i915/gem/i915_gem_context.h
@@ -207,7 +207,6 @@ static inline void
i915_gem_engines_iter_init(struct i915_gem_engines_iter *it,
struct i915_gem_engines *engines)
{
- GEM_BUG_ON(!engines);
it->engines = engines;
it->idx = 0;
}
diff --git a/drivers/gpu/drm/i915/gem/selftests/mock_context.c b/drivers/gpu/drm/i915/gem/selftests/mock_context.c
index b12ea1daa29d..e7e3c620f542 100644
--- a/drivers/gpu/drm/i915/gem/selftests/mock_context.c
+++ b/drivers/gpu/drm/i915/gem/selftests/mock_context.c
@@ -23,6 +23,9 @@ mock_context(struct drm_i915_private *i915,
INIT_LIST_HEAD(&ctx->link);
ctx->i915 = i915;
+ spin_lock_init(&ctx->stale.lock);
+ INIT_LIST_HEAD(&ctx->stale.engines);
+
i915_gem_context_set_persistence(ctx);
mutex_init(&ctx->engines_mutex);
--
2.25.1
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Intel-gfx] [PATCH 3/4] drm/i915/gt: Prevent allocation on a banned context
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 2/4] drm/i915/gem: Consolidate ctx->engines[] release Chris Wilson
@ 2020-03-03 9:20 ` Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 4/4] drm/i915/gem: Check that the context wasn't closed during setup Chris Wilson
` (4 subsequent siblings)
6 siblings, 0 replies; 8+ messages in thread
From: Chris Wilson @ 2020-03-03 9:20 UTC (permalink / raw)
To: intel-gfx; +Cc: Matthew Auld
If a context is banned even before we submit our first request to it,
report the failure before we attempt to allocate any resources for the
context.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Reviewed-by: Matthew Auld <matthew.auld@intel.com>
---
drivers/gpu/drm/i915/gt/intel_context.c | 5 +++++
1 file changed, 5 insertions(+)
diff --git a/drivers/gpu/drm/i915/gt/intel_context.c b/drivers/gpu/drm/i915/gt/intel_context.c
index 8bb444cda14f..01474d3a558b 100644
--- a/drivers/gpu/drm/i915/gt/intel_context.c
+++ b/drivers/gpu/drm/i915/gt/intel_context.c
@@ -51,6 +51,11 @@ int intel_context_alloc_state(struct intel_context *ce)
return -EINTR;
if (!test_bit(CONTEXT_ALLOC_BIT, &ce->flags)) {
+ if (intel_context_is_banned(ce)) {
+ err = -EIO;
+ goto unlock;
+ }
+
err = ce->ops->alloc(ce);
if (unlikely(err))
goto unlock;
--
2.25.1
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Intel-gfx] [PATCH 4/4] drm/i915/gem: Check that the context wasn't closed during setup
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 2/4] drm/i915/gem: Consolidate ctx->engines[] release Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 3/4] drm/i915/gt: Prevent allocation on a banned context Chris Wilson
@ 2020-03-03 9:20 ` Chris Wilson
2020-03-03 14:09 ` [Intel-gfx] ✗ Fi.CI.DOCS: warning for series starting with [1/4] drm/i915: Fix doclinks Patchwork
` (3 subsequent siblings)
6 siblings, 0 replies; 8+ messages in thread
From: Chris Wilson @ 2020-03-03 9:20 UTC (permalink / raw)
To: intel-gfx; +Cc: Matthew Auld
As setup takes a long time, the user may close the context during the
construction of the execbuf. In order to make sure we correctly track
all outstanding work with non-persistent contexts, we need to serialise
the submission with the context closure and mop up any leaks.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Reviewed-by: Matthew Auld <matthew.auld@intel.com>
---
.../gpu/drm/i915/gem/i915_gem_execbuffer.c | 67 ++++++++++++++++++-
drivers/gpu/drm/i915/i915_request.c | 54 +++------------
2 files changed, 74 insertions(+), 47 deletions(-)
diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
index ac0e5fc5675e..84bc820a4608 100644
--- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
+++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
@@ -2566,6 +2566,71 @@ signal_fence_array(struct i915_execbuffer *eb,
}
}
+static void retire_requests(struct intel_timeline *tl, struct i915_request *end)
+{
+ struct i915_request *rq, *rn;
+
+ list_for_each_entry_safe(rq, rn, &tl->requests, link)
+ if (rq == end || !i915_request_retire(rq))
+ break;
+}
+
+static void eb_request_add(struct i915_execbuffer *eb)
+{
+ struct i915_request *rq = eb->request;
+ struct intel_timeline * const tl = i915_request_timeline(rq);
+ struct i915_sched_attr attr = {};
+ struct i915_request *prev;
+
+ lockdep_assert_held(&tl->mutex);
+ lockdep_unpin_lock(&tl->mutex, rq->cookie);
+
+ trace_i915_request_add(rq);
+
+ prev = __i915_request_commit(rq);
+
+ /* Check that the context wasn't destroyed before submission */
+ if (likely(rcu_access_pointer(eb->context->gem_context))) {
+ attr = eb->gem_context->sched;
+
+ /*
+ * Boost actual workloads past semaphores!
+ *
+ * With semaphores we spin on one engine waiting for another,
+ * simply to reduce the latency of starting our work when
+ * the signaler completes. However, if there is any other
+ * work that we could be doing on this engine instead, that
+ * is better utilisation and will reduce the overall duration
+ * of the current work. To avoid PI boosting a semaphore
+ * far in the distance past over useful work, we keep a history
+ * of any semaphore use along our dependency chain.
+ */
+ if (!(rq->sched.flags & I915_SCHED_HAS_SEMAPHORE_CHAIN))
+ attr.priority |= I915_PRIORITY_NOSEMAPHORE;
+
+ /*
+ * Boost priorities to new clients (new request flows).
+ *
+ * Allow interactive/synchronous clients to jump ahead of
+ * the bulk clients. (FQ_CODEL)
+ */
+ if (list_empty(&rq->sched.signalers_list))
+ attr.priority |= I915_PRIORITY_WAIT;
+ } else {
+ /* Serialise with context_close via the add_to_timeline */
+ i915_request_skip(rq, -ENOENT);
+ }
+
+ local_bh_disable();
+ __i915_request_queue(rq, &attr);
+ local_bh_enable(); /* Kick the execlists tasklet if just scheduled */
+
+ /* Try to clean up the client's timeline after submitting the request */
+ retire_requests(tl, prev);
+
+ mutex_unlock(&tl->mutex);
+}
+
static int
i915_gem_do_execbuffer(struct drm_device *dev,
struct drm_file *file,
@@ -2778,7 +2843,7 @@ i915_gem_do_execbuffer(struct drm_device *dev,
err_request:
add_to_client(eb.request, file);
i915_request_get(eb.request);
- i915_request_add(eb.request);
+ eb_request_add(&eb);
if (fences)
signal_fence_array(&eb, fences);
diff --git a/drivers/gpu/drm/i915/i915_request.c b/drivers/gpu/drm/i915/i915_request.c
index feccf29381aa..d837c1380015 100644
--- a/drivers/gpu/drm/i915/i915_request.c
+++ b/drivers/gpu/drm/i915/i915_request.c
@@ -1339,39 +1339,23 @@ void i915_request_add(struct i915_request *rq)
{
struct intel_timeline * const tl = i915_request_timeline(rq);
struct i915_sched_attr attr = {};
- struct i915_request *prev;
+ struct i915_gem_context *ctx;
lockdep_assert_held(&tl->mutex);
lockdep_unpin_lock(&tl->mutex, rq->cookie);
trace_i915_request_add(rq);
+ __i915_request_commit(rq);
- prev = __i915_request_commit(rq);
-
- if (rcu_access_pointer(rq->context->gem_context))
- attr = i915_request_gem_context(rq)->sched;
+ /* XXX placeholder for selftests */
+ rcu_read_lock();
+ ctx = rcu_dereference(rq->context->gem_context);
+ if (ctx)
+ attr = ctx->sched;
+ rcu_read_unlock();
- /*
- * Boost actual workloads past semaphores!
- *
- * With semaphores we spin on one engine waiting for another,
- * simply to reduce the latency of starting our work when
- * the signaler completes. However, if there is any other
- * work that we could be doing on this engine instead, that
- * is better utilisation and will reduce the overall duration
- * of the current work. To avoid PI boosting a semaphore
- * far in the distance past over useful work, we keep a history
- * of any semaphore use along our dependency chain.
- */
if (!(rq->sched.flags & I915_SCHED_HAS_SEMAPHORE_CHAIN))
attr.priority |= I915_PRIORITY_NOSEMAPHORE;
-
- /*
- * Boost priorities to new clients (new request flows).
- *
- * Allow interactive/synchronous clients to jump ahead of
- * the bulk clients. (FQ_CODEL)
- */
if (list_empty(&rq->sched.signalers_list))
attr.priority |= I915_PRIORITY_WAIT;
@@ -1379,28 +1363,6 @@ void i915_request_add(struct i915_request *rq)
__i915_request_queue(rq, &attr);
local_bh_enable(); /* Kick the execlists tasklet if just scheduled */
- /*
- * In typical scenarios, we do not expect the previous request on
- * the timeline to be still tracked by timeline->last_request if it
- * has been completed. If the completed request is still here, that
- * implies that request retirement is a long way behind submission,
- * suggesting that we haven't been retiring frequently enough from
- * the combination of retire-before-alloc, waiters and the background
- * retirement worker. So if the last request on this timeline was
- * already completed, do a catch up pass, flushing the retirement queue
- * up to this client. Since we have now moved the heaviest operations
- * during retirement onto secondary workers, such as freeing objects
- * or contexts, retiring a bunch of requests is mostly list management
- * (and cache misses), and so we should not be overly penalizing this
- * client by performing excess work, though we may still performing
- * work on behalf of others -- but instead we should benefit from
- * improved resource management. (Well, that's the theory at least.)
- */
- if (prev &&
- i915_request_completed(prev) &&
- rcu_access_pointer(prev->timeline) == tl)
- i915_request_retire_upto(prev);
-
mutex_unlock(&tl->mutex);
}
--
2.25.1
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply related [flat|nested] 8+ messages in thread
* [Intel-gfx] ✗ Fi.CI.DOCS: warning for series starting with [1/4] drm/i915: Fix doclinks
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
` (2 preceding siblings ...)
2020-03-03 9:20 ` [Intel-gfx] [PATCH 4/4] drm/i915/gem: Check that the context wasn't closed during setup Chris Wilson
@ 2020-03-03 14:09 ` Patchwork
2020-03-03 14:23 ` [Intel-gfx] ✓ Fi.CI.BAT: success " Patchwork
` (2 subsequent siblings)
6 siblings, 0 replies; 8+ messages in thread
From: Patchwork @ 2020-03-03 14:09 UTC (permalink / raw)
To: Chris Wilson; +Cc: intel-gfx
== Series Details ==
Series: series starting with [1/4] drm/i915: Fix doclinks
URL : https://patchwork.freedesktop.org/series/74194/
State : warning
== Summary ==
$ make htmldocs 2>&1 > /dev/null | grep i915
./drivers/gpu/drm/i915/display/intel_dpll_mgr.h:285: warning: Function parameter or member 'get_freq' not described in 'intel_shared_dpll_funcs'
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Intel-gfx] ✓ Fi.CI.BAT: success for series starting with [1/4] drm/i915: Fix doclinks
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
` (3 preceding siblings ...)
2020-03-03 14:09 ` [Intel-gfx] ✗ Fi.CI.DOCS: warning for series starting with [1/4] drm/i915: Fix doclinks Patchwork
@ 2020-03-03 14:23 ` Patchwork
2020-03-04 2:16 ` [Intel-gfx] ✗ Fi.CI.IGT: failure " Patchwork
2020-03-04 23:38 ` [Intel-gfx] [PATCH 1/4] " Souza, Jose
6 siblings, 0 replies; 8+ messages in thread
From: Patchwork @ 2020-03-03 14:23 UTC (permalink / raw)
To: Chris Wilson; +Cc: intel-gfx
== Series Details ==
Series: series starting with [1/4] drm/i915: Fix doclinks
URL : https://patchwork.freedesktop.org/series/74194/
State : success
== Summary ==
CI Bug Log - changes from CI_DRM_8055 -> Patchwork_16796
====================================================
Summary
-------
**WARNING**
Minor unknown changes coming with Patchwork_16796 need to be verified
manually.
If you think the reported changes have nothing to do with the changes
introduced in Patchwork_16796, please notify your bug team to allow them
to document this new failure mode, which will reduce false positives in CI.
External URL: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/index.html
Possible new issues
-------------------
Here are the unknown changes that may have been introduced in Patchwork_16796:
### IGT changes ###
#### Warnings ####
* igt@amdgpu/amd_basic@query-info:
- fi-bwr-2160: [SKIP][1] ([fdo#109271]) -> [FAIL][2]
[1]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/fi-bwr-2160/igt@amdgpu/amd_basic@query-info.html
[2]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/fi-bwr-2160/igt@amdgpu/amd_basic@query-info.html
Known issues
------------
Here are the changes found in Patchwork_16796 that come from known issues:
### IGT changes ###
#### Issues hit ####
* igt@gem_flink_basic@bad-open:
- fi-tgl-y: [PASS][3] -> [DMESG-WARN][4] ([CI#94] / [i915#402])
[3]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/fi-tgl-y/igt@gem_flink_basic@bad-open.html
[4]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/fi-tgl-y/igt@gem_flink_basic@bad-open.html
* igt@i915_pm_rpm@module-reload:
- fi-kbl-guc: [PASS][5] -> [FAIL][6] ([i915#579])
[5]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/fi-kbl-guc/igt@i915_pm_rpm@module-reload.html
[6]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/fi-kbl-guc/igt@i915_pm_rpm@module-reload.html
* igt@kms_chamelium@hdmi-hpd-fast:
- fi-kbl-7500u: [PASS][7] -> [FAIL][8] ([fdo#111096] / [i915#323])
[7]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/fi-kbl-7500u/igt@kms_chamelium@hdmi-hpd-fast.html
[8]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/fi-kbl-7500u/igt@kms_chamelium@hdmi-hpd-fast.html
#### Possible fixes ####
* igt@kms_addfb_basic@addfb25-framebuffer-vs-set-tiling:
- fi-tgl-y: [DMESG-WARN][9] ([CI#94] / [i915#402]) -> [PASS][10] +2 similar issues
[9]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/fi-tgl-y/igt@kms_addfb_basic@addfb25-framebuffer-vs-set-tiling.html
[10]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/fi-tgl-y/igt@kms_addfb_basic@addfb25-framebuffer-vs-set-tiling.html
[CI#94]: https://gitlab.freedesktop.org/gfx-ci/i915-infra/issues/94
[fdo#109271]: https://bugs.freedesktop.org/show_bug.cgi?id=109271
[fdo#111096]: https://bugs.freedesktop.org/show_bug.cgi?id=111096
[i915#323]: https://gitlab.freedesktop.org/drm/intel/issues/323
[i915#402]: https://gitlab.freedesktop.org/drm/intel/issues/402
[i915#579]: https://gitlab.freedesktop.org/drm/intel/issues/579
Participating hosts (46 -> 42)
------------------------------
Additional (4): fi-bsw-kefka fi-kbl-7560u fi-cfl-8109u fi-elk-e7500
Missing (8): fi-bdw-5557u fi-hsw-4200u fi-hsw-peppy fi-ctg-p8600 fi-skl-6600u fi-bdw-samus fi-byt-clapper fi-skl-6700k2
Build changes
-------------
* CI: CI-20190529 -> None
* Linux: CI_DRM_8055 -> Patchwork_16796
CI-20190529: 20190529
CI_DRM_8055: df5c59f81df63a983b7490968e16c3255adc196b @ git://anongit.freedesktop.org/gfx-ci/linux
IGT_5484: 91b36b61e76901a2bd09fe93ac7bf7b8a60f258c @ git://anongit.freedesktop.org/xorg/app/intel-gpu-tools
Patchwork_16796: 049a64981bce33df9b0fbaf7f0be135886a27d31 @ git://anongit.freedesktop.org/gfx-ci/linux
== Linux commits ==
049a64981bce drm/i915/gem: Check that the context wasn't closed during setup
a06923fc59b6 drm/i915/gt: Prevent allocation on a banned context
3115b9457cab drm/i915/gem: Consolidate ctx->engines[] release
== Logs ==
For more details see: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/index.html
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply [flat|nested] 8+ messages in thread
* [Intel-gfx] ✗ Fi.CI.IGT: failure for series starting with [1/4] drm/i915: Fix doclinks
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
` (4 preceding siblings ...)
2020-03-03 14:23 ` [Intel-gfx] ✓ Fi.CI.BAT: success " Patchwork
@ 2020-03-04 2:16 ` Patchwork
2020-03-04 23:38 ` [Intel-gfx] [PATCH 1/4] " Souza, Jose
6 siblings, 0 replies; 8+ messages in thread
From: Patchwork @ 2020-03-04 2:16 UTC (permalink / raw)
To: Chris Wilson; +Cc: intel-gfx
== Series Details ==
Series: series starting with [1/4] drm/i915: Fix doclinks
URL : https://patchwork.freedesktop.org/series/74194/
State : failure
== Summary ==
CI Bug Log - changes from CI_DRM_8055_full -> Patchwork_16796_full
====================================================
Summary
-------
**FAILURE**
Serious unknown changes coming with Patchwork_16796_full absolutely need to be
verified manually.
If you think the reported changes have nothing to do with the changes
introduced in Patchwork_16796_full, please notify your bug team to allow them
to document this new failure mode, which will reduce false positives in CI.
Possible new issues
-------------------
Here are the unknown changes that may have been introduced in Patchwork_16796_full:
### IGT changes ###
#### Possible regressions ####
* igt@gem_exec_schedule@pi-common-bsd:
- shard-skl: [PASS][1] -> [DMESG-WARN][2]
[1]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl1/igt@gem_exec_schedule@pi-common-bsd.html
[2]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl5/igt@gem_exec_schedule@pi-common-bsd.html
Known issues
------------
Here are the changes found in Patchwork_16796_full that come from known issues:
### IGT changes ###
#### Issues hit ####
* igt@gem_ctx_exec@basic-nohangcheck:
- shard-snb: [PASS][3] -> [FAIL][4] ([i915#1148])
[3]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-snb2/igt@gem_ctx_exec@basic-nohangcheck.html
[4]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-snb2/igt@gem_ctx_exec@basic-nohangcheck.html
- shard-hsw: [PASS][5] -> [FAIL][6] ([i915#1148])
[5]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-hsw8/igt@gem_ctx_exec@basic-nohangcheck.html
[6]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-hsw2/igt@gem_ctx_exec@basic-nohangcheck.html
* igt@gem_ctx_persistence@legacy-engines-mixed-process@blt:
- shard-kbl: [PASS][7] -> [INCOMPLETE][8] ([fdo#103665] / [i915#1197] / [i915#1239])
[7]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-kbl2/igt@gem_ctx_persistence@legacy-engines-mixed-process@blt.html
[8]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-kbl7/igt@gem_ctx_persistence@legacy-engines-mixed-process@blt.html
* igt@gem_ctx_persistence@legacy-engines-mixed-process@bsd2:
- shard-kbl: [PASS][9] -> [FAIL][10] ([i915#679])
[9]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-kbl2/igt@gem_ctx_persistence@legacy-engines-mixed-process@bsd2.html
[10]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-kbl7/igt@gem_ctx_persistence@legacy-engines-mixed-process@bsd2.html
* igt@gem_ctx_shared@exec-single-timeline-bsd:
- shard-iclb: [PASS][11] -> [SKIP][12] ([fdo#110841])
[11]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb5/igt@gem_ctx_shared@exec-single-timeline-bsd.html
[12]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb4/igt@gem_ctx_shared@exec-single-timeline-bsd.html
* igt@gem_eio@in-flight-suspend:
- shard-apl: [PASS][13] -> [DMESG-WARN][14] ([i915#180])
[13]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-apl3/igt@gem_eio@in-flight-suspend.html
[14]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-apl1/igt@gem_eio@in-flight-suspend.html
* igt@gem_exec_schedule@implicit-both-bsd:
- shard-iclb: [PASS][15] -> [SKIP][16] ([i915#677]) +1 similar issue
[15]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb5/igt@gem_exec_schedule@implicit-both-bsd.html
[16]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb4/igt@gem_exec_schedule@implicit-both-bsd.html
* igt@gem_exec_schedule@preemptive-hang-bsd:
- shard-iclb: [PASS][17] -> [SKIP][18] ([fdo#112146]) +4 similar issues
[17]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb8/igt@gem_exec_schedule@preemptive-hang-bsd.html
[18]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb1/igt@gem_exec_schedule@preemptive-hang-bsd.html
* igt@kms_cursor_crc@pipe-b-cursor-128x128-random:
- shard-skl: [PASS][19] -> [FAIL][20] ([i915#54])
[19]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl6/igt@kms_cursor_crc@pipe-b-cursor-128x128-random.html
[20]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl1/igt@kms_cursor_crc@pipe-b-cursor-128x128-random.html
* igt@kms_cursor_legacy@pipe-c-torture-bo:
- shard-glk: [PASS][21] -> [DMESG-WARN][22] ([i915#128])
[21]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-glk9/igt@kms_cursor_legacy@pipe-c-torture-bo.html
[22]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-glk3/igt@kms_cursor_legacy@pipe-c-torture-bo.html
* igt@kms_draw_crc@draw-method-xrgb2101010-mmap-wc-xtiled:
- shard-skl: [PASS][23] -> [FAIL][24] ([i915#52] / [i915#54])
[23]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl1/igt@kms_draw_crc@draw-method-xrgb2101010-mmap-wc-xtiled.html
[24]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl5/igt@kms_draw_crc@draw-method-xrgb2101010-mmap-wc-xtiled.html
* igt@kms_flip@flip-vs-suspend:
- shard-skl: [PASS][25] -> [INCOMPLETE][26] ([i915#221]) +1 similar issue
[25]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl6/igt@kms_flip@flip-vs-suspend.html
[26]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl1/igt@kms_flip@flip-vs-suspend.html
* igt@kms_psr@psr2_cursor_render:
- shard-iclb: [PASS][27] -> [SKIP][28] ([fdo#109441]) +2 similar issues
[27]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb2/igt@kms_psr@psr2_cursor_render.html
[28]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb6/igt@kms_psr@psr2_cursor_render.html
* igt@kms_setmode@basic:
- shard-kbl: [PASS][29] -> [FAIL][30] ([i915#31])
[29]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-kbl2/igt@kms_setmode@basic.html
[30]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-kbl1/igt@kms_setmode@basic.html
* igt@kms_vblank@pipe-a-ts-continuation-suspend:
- shard-kbl: [PASS][31] -> [DMESG-WARN][32] ([i915#180]) +4 similar issues
[31]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-kbl4/igt@kms_vblank@pipe-a-ts-continuation-suspend.html
[32]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-kbl6/igt@kms_vblank@pipe-a-ts-continuation-suspend.html
* igt@perf_pmu@busy-check-all-vcs1:
- shard-iclb: [PASS][33] -> [SKIP][34] ([fdo#112080]) +12 similar issues
[33]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb2/igt@perf_pmu@busy-check-all-vcs1.html
[34]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb6/igt@perf_pmu@busy-check-all-vcs1.html
* igt@prime_busy@hang-bsd2:
- shard-iclb: [PASS][35] -> [SKIP][36] ([fdo#109276]) +23 similar issues
[35]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb4/igt@prime_busy@hang-bsd2.html
[36]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb8/igt@prime_busy@hang-bsd2.html
#### Possible fixes ####
* igt@gem_ctx_persistence@close-replace-race:
- shard-tglb: [INCOMPLETE][37] ([i915#1291]) -> [PASS][38]
[37]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-tglb7/igt@gem_ctx_persistence@close-replace-race.html
[38]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-tglb3/igt@gem_ctx_persistence@close-replace-race.html
* igt@gem_ctx_persistence@engines-mixed-process@bcs0:
- shard-apl: [INCOMPLETE][39] ([fdo#103927] / [i915#1197] / [i915#1239]) -> [PASS][40]
[39]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-apl8/igt@gem_ctx_persistence@engines-mixed-process@bcs0.html
[40]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-apl8/igt@gem_ctx_persistence@engines-mixed-process@bcs0.html
* igt@gem_ctx_persistence@engines-mixed-process@rcs0:
- shard-apl: [FAIL][41] ([i915#679]) -> [PASS][42]
[41]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-apl8/igt@gem_ctx_persistence@engines-mixed-process@rcs0.html
[42]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-apl8/igt@gem_ctx_persistence@engines-mixed-process@rcs0.html
* igt@gem_ctx_persistence@engines-mixed-process@vcs0:
- shard-skl: [FAIL][43] ([i915#679]) -> [PASS][44]
[43]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl8/igt@gem_ctx_persistence@engines-mixed-process@vcs0.html
[44]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl5/igt@gem_ctx_persistence@engines-mixed-process@vcs0.html
* igt@gem_ctx_persistence@engines-mixed-process@vecs0:
- shard-skl: [INCOMPLETE][45] ([i915#1197] / [i915#1239]) -> [PASS][46]
[45]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl8/igt@gem_ctx_persistence@engines-mixed-process@vecs0.html
[46]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl5/igt@gem_ctx_persistence@engines-mixed-process@vecs0.html
* igt@gem_eio@in-flight-suspend:
- shard-skl: [INCOMPLETE][47] ([i915#69]) -> [PASS][48] +1 similar issue
[47]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl3/igt@gem_eio@in-flight-suspend.html
[48]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl9/igt@gem_eio@in-flight-suspend.html
* igt@gem_exec_balancer@smoke:
- shard-iclb: [SKIP][49] ([fdo#110854]) -> [PASS][50]
[49]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb5/igt@gem_exec_balancer@smoke.html
[50]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb4/igt@gem_exec_balancer@smoke.html
* igt@gem_exec_schedule@fifo-bsd1:
- shard-iclb: [SKIP][51] ([fdo#109276]) -> [PASS][52] +15 similar issues
[51]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb8/igt@gem_exec_schedule@fifo-bsd1.html
[52]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb4/igt@gem_exec_schedule@fifo-bsd1.html
* igt@gem_exec_schedule@implicit-both-bsd1:
- shard-iclb: [SKIP][53] ([fdo#109276] / [i915#677]) -> [PASS][54]
[53]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb3/igt@gem_exec_schedule@implicit-both-bsd1.html
[54]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb1/igt@gem_exec_schedule@implicit-both-bsd1.html
* igt@gem_exec_schedule@pi-common-bsd:
- shard-iclb: [SKIP][55] ([i915#677]) -> [PASS][56] +2 similar issues
[55]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb4/igt@gem_exec_schedule@pi-common-bsd.html
[56]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb8/igt@gem_exec_schedule@pi-common-bsd.html
* igt@gem_exec_schedule@preempt-other-chain-bsd:
- shard-iclb: [SKIP][57] ([fdo#112146]) -> [PASS][58] +7 similar issues
[57]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb2/igt@gem_exec_schedule@preempt-other-chain-bsd.html
[58]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb3/igt@gem_exec_schedule@preempt-other-chain-bsd.html
* igt@i915_pm_rps@waitboost:
- shard-iclb: [FAIL][59] ([i915#413]) -> [PASS][60]
[59]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb8/igt@i915_pm_rps@waitboost.html
[60]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb4/igt@i915_pm_rps@waitboost.html
* igt@kms_cursor_legacy@flip-vs-cursor-busy-crc-atomic:
- shard-skl: [FAIL][61] ([IGT#5]) -> [PASS][62]
[61]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl6/igt@kms_cursor_legacy@flip-vs-cursor-busy-crc-atomic.html
[62]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl2/igt@kms_cursor_legacy@flip-vs-cursor-busy-crc-atomic.html
* igt@kms_draw_crc@draw-method-xrgb2101010-mmap-cpu-untiled:
- shard-snb: [SKIP][63] ([fdo#109271]) -> [PASS][64] +1 similar issue
[63]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-snb6/igt@kms_draw_crc@draw-method-xrgb2101010-mmap-cpu-untiled.html
[64]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-snb5/igt@kms_draw_crc@draw-method-xrgb2101010-mmap-cpu-untiled.html
* igt@kms_flip@flip-vs-suspend-interruptible:
- shard-apl: [DMESG-WARN][65] ([i915#180]) -> [PASS][66]
[65]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-apl4/igt@kms_flip@flip-vs-suspend-interruptible.html
[66]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-apl2/igt@kms_flip@flip-vs-suspend-interruptible.html
* igt@kms_frontbuffer_tracking@fbc-suspend:
- shard-kbl: [DMESG-WARN][67] ([i915#180]) -> [PASS][68] +4 similar issues
[67]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-kbl3/igt@kms_frontbuffer_tracking@fbc-suspend.html
[68]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-kbl4/igt@kms_frontbuffer_tracking@fbc-suspend.html
* igt@kms_plane_alpha_blend@pipe-a-constant-alpha-min:
- shard-skl: [FAIL][69] ([fdo#108145]) -> [PASS][70]
[69]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-skl8/igt@kms_plane_alpha_blend@pipe-a-constant-alpha-min.html
[70]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-skl6/igt@kms_plane_alpha_blend@pipe-a-constant-alpha-min.html
* igt@kms_psr@psr2_basic:
- shard-iclb: [SKIP][71] ([fdo#109441]) -> [PASS][72] +1 similar issue
[71]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb1/igt@kms_psr@psr2_basic.html
[72]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb2/igt@kms_psr@psr2_basic.html
* igt@kms_setmode@basic:
- shard-apl: [FAIL][73] ([i915#31]) -> [PASS][74]
[73]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-apl3/igt@kms_setmode@basic.html
[74]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-apl8/igt@kms_setmode@basic.html
* igt@perf_pmu@busy-no-semaphores-vcs1:
- shard-iclb: [SKIP][75] ([fdo#112080]) -> [PASS][76] +12 similar issues
[75]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-iclb3/igt@perf_pmu@busy-no-semaphores-vcs1.html
[76]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-iclb1/igt@perf_pmu@busy-no-semaphores-vcs1.html
#### Warnings ####
* igt@i915_pm_dc@dc6-psr:
- shard-tglb: [SKIP][77] ([i915#468]) -> [FAIL][78] ([i915#454])
[77]: https://intel-gfx-ci.01.org/tree/drm-tip/CI_DRM_8055/shard-tglb2/igt@i915_pm_dc@dc6-psr.html
[78]: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/shard-tglb6/igt@i915_pm_dc@dc6-psr.html
[IGT#5]: https://gitlab.freedesktop.org/drm/igt-gpu-tools/issues/5
[fdo#103665]: https://bugs.freedesktop.org/show_bug.cgi?id=103665
[fdo#103927]: https://bugs.freedesktop.org/show_bug.cgi?id=103927
[fdo#108145]: https://bugs.freedesktop.org/show_bug.cgi?id=108145
[fdo#109271]: https://bugs.freedesktop.org/show_bug.cgi?id=109271
[fdo#109276]: https://bugs.freedesktop.org/show_bug.cgi?id=109276
[fdo#109441]: https://bugs.freedesktop.org/show_bug.cgi?id=109441
[fdo#110841]: https://bugs.freedesktop.org/show_bug.cgi?id=110841
[fdo#110854]: https://bugs.freedesktop.org/show_bug.cgi?id=110854
[fdo#112080]: https://bugs.freedesktop.org/show_bug.cgi?id=112080
[fdo#112146]: https://bugs.freedesktop.org/show_bug.cgi?id=112146
[i915#1148]: https://gitlab.freedesktop.org/drm/intel/issues/1148
[i915#1197]: https://gitlab.freedesktop.org/drm/intel/issues/1197
[i915#1239]: https://gitlab.freedesktop.org/drm/intel/issues/1239
[i915#128]: https://gitlab.freedesktop.org/drm/intel/issues/128
[i915#1291]: https://gitlab.freedesktop.org/drm/intel/issues/1291
[i915#180]: https://gitlab.freedesktop.org/drm/intel/issues/180
[i915#221]: https://gitlab.freedesktop.org/drm/intel/issues/221
[i915#31]: https://gitlab.freedesktop.org/drm/intel/issues/31
[i915#413]: https://gitlab.freedesktop.org/drm/intel/issues/413
[i915#454]: https://gitlab.freedesktop.org/drm/intel/issues/454
[i915#468]: https://gitlab.freedesktop.org/drm/intel/issues/468
[i915#52]: https://gitlab.freedesktop.org/drm/intel/issues/52
[i915#54]: https://gitlab.freedesktop.org/drm/intel/issues/54
[i915#677]: https://gitlab.freedesktop.org/drm/intel/issues/677
[i915#679]: https://gitlab.freedesktop.org/drm/intel/issues/679
[i915#69]: https://gitlab.freedesktop.org/drm/intel/issues/69
Participating hosts (10 -> 10)
------------------------------
No changes in participating hosts
Build changes
-------------
* CI: CI-20190529 -> None
* Linux: CI_DRM_8055 -> Patchwork_16796
CI-20190529: 20190529
CI_DRM_8055: df5c59f81df63a983b7490968e16c3255adc196b @ git://anongit.freedesktop.org/gfx-ci/linux
IGT_5484: 91b36b61e76901a2bd09fe93ac7bf7b8a60f258c @ git://anongit.freedesktop.org/xorg/app/intel-gpu-tools
Patchwork_16796: 049a64981bce33df9b0fbaf7f0be135886a27d31 @ git://anongit.freedesktop.org/gfx-ci/linux
piglit_4509: fdc5a4ca11124ab8413c7988896eec4c97336694 @ git://anongit.freedesktop.org/piglit
== Logs ==
For more details see: https://intel-gfx-ci.01.org/tree/drm-tip/Patchwork_16796/index.html
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
` (5 preceding siblings ...)
2020-03-04 2:16 ` [Intel-gfx] ✗ Fi.CI.IGT: failure " Patchwork
@ 2020-03-04 23:38 ` Souza, Jose
6 siblings, 0 replies; 8+ messages in thread
From: Souza, Jose @ 2020-03-04 23:38 UTC (permalink / raw)
To: intel-gfx, chris
On Tue, 2020-03-03 at 09:20 +0000, Chris Wilson wrote:
> Update locations for
>
> ./drivers/gpu/drm/i915/i915_vma.h:1: warning: 'Virtual Memory
> Address' not found
> ./drivers/gpu/drm/i915/i915_gem_gtt.c:1: warning: 'Global GTT views'
> not found
Reviewed-by: José Roberto de Souza <jose.souza@intel.com>
>
> Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
> ---
> Documentation/gpu/i915.rst | 4 ++--
> 1 file changed, 2 insertions(+), 2 deletions(-)
>
> diff --git a/Documentation/gpu/i915.rst b/Documentation/gpu/i915.rst
> index cc74e24ca3b5..f6d363b6756e 100644
> --- a/Documentation/gpu/i915.rst
> +++ b/Documentation/gpu/i915.rst
> @@ -332,7 +332,7 @@ This process is dubbed relocation.
> GEM BO Management Implementation Details
> ----------------------------------------
>
> -.. kernel-doc:: drivers/gpu/drm/i915/i915_vma.h
> +.. kernel-doc:: drivers/gpu/drm/i915/i915_vma_types.h
> :doc: Virtual Memory Address
>
> Buffer Object Eviction
> @@ -382,7 +382,7 @@ Logical Rings, Logical Ring Contexts and
> Execlists
> Global GTT views
> ----------------
>
> -.. kernel-doc:: drivers/gpu/drm/i915/i915_gem_gtt.c
> +.. kernel-doc:: drivers/gpu/drm/i915/i915_vma_types.h
> :doc: Global GTT views
>
> .. kernel-doc:: drivers/gpu/drm/i915/i915_gem_gtt.c
_______________________________________________
Intel-gfx mailing list
Intel-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/intel-gfx
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2020-03-04 23:38 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-03-03 9:20 [Intel-gfx] [PATCH 1/4] drm/i915: Fix doclinks Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 2/4] drm/i915/gem: Consolidate ctx->engines[] release Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 3/4] drm/i915/gt: Prevent allocation on a banned context Chris Wilson
2020-03-03 9:20 ` [Intel-gfx] [PATCH 4/4] drm/i915/gem: Check that the context wasn't closed during setup Chris Wilson
2020-03-03 14:09 ` [Intel-gfx] ✗ Fi.CI.DOCS: warning for series starting with [1/4] drm/i915: Fix doclinks Patchwork
2020-03-03 14:23 ` [Intel-gfx] ✓ Fi.CI.BAT: success " Patchwork
2020-03-04 2:16 ` [Intel-gfx] ✗ Fi.CI.IGT: failure " Patchwork
2020-03-04 23:38 ` [Intel-gfx] [PATCH 1/4] " Souza, Jose
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.