All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH] drm/sched: add run job trace
@ 2020-03-13  9:56 Lucas Stach
  2020-03-13 10:17 ` Christian König
  0 siblings, 1 reply; 3+ messages in thread
From: Lucas Stach @ 2020-03-13  9:56 UTC (permalink / raw)
  To: dri-devel; +Cc: Robert Beckett, Christian König, kernel, patchwork-lst

From: Robert Beckett <bob.beckett@collabora.com>

Add a new trace event to show when jobs are run on the HW.

Signed-off-by: Robert Beckett <bob.beckett@collabora.com>
Signed-off-by: Lucas Stach <l.stach@pengutronix.de>
---
 .../gpu/drm/scheduler/gpu_scheduler_trace.h   | 27 +++++++++++++++++++
 drivers/gpu/drm/scheduler/sched_main.c        |  1 +
 2 files changed, 28 insertions(+)

diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
index d79086498aff..877ce9b127f1 100644
--- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
+++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
@@ -59,6 +59,33 @@ TRACE_EVENT(drm_sched_job,
 		      __entry->job_count, __entry->hw_job_count)
 );
 
+TRACE_EVENT(drm_run_job,
+	    TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
+	    TP_ARGS(sched_job, entity),
+	    TP_STRUCT__entry(
+			     __field(struct drm_sched_entity *, entity)
+			     __field(struct dma_fence *, fence)
+			     __field(const char *, name)
+			     __field(uint64_t, id)
+			     __field(u32, job_count)
+			     __field(int, hw_job_count)
+			     ),
+
+	    TP_fast_assign(
+			   __entry->entity = entity;
+			   __entry->id = sched_job->id;
+			   __entry->fence = &sched_job->s_fence->finished;
+			   __entry->name = sched_job->sched->name;
+			   __entry->job_count = spsc_queue_count(&entity->job_queue);
+			   __entry->hw_job_count = atomic_read(
+				   &sched_job->sched->hw_rq_count);
+			   ),
+	    TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d",
+		      __entry->entity, __entry->id,
+		      __entry->fence, __entry->name,
+		      __entry->job_count, __entry->hw_job_count)
+);
+
 TRACE_EVENT(drm_sched_process_job,
 	    TP_PROTO(struct drm_sched_fence *fence),
 	    TP_ARGS(fence),
diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c
index 71ce6215956f..34231b7163cc 100644
--- a/drivers/gpu/drm/scheduler/sched_main.c
+++ b/drivers/gpu/drm/scheduler/sched_main.c
@@ -773,6 +773,7 @@ static int drm_sched_main(void *param)
 		atomic_inc(&sched->hw_rq_count);
 		drm_sched_job_begin(sched_job);
 
+		trace_drm_run_job(sched_job, entity);
 		fence = sched->ops->run_job(sched_job);
 		drm_sched_fence_scheduled(s_fence);
 
-- 
2.20.1

_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel

^ permalink raw reply related	[flat|nested] 3+ messages in thread

* Re: [PATCH] drm/sched: add run job trace
  2020-03-13  9:56 [PATCH] drm/sched: add run job trace Lucas Stach
@ 2020-03-13 10:17 ` Christian König
  2020-03-13 14:25   ` Alex Deucher
  0 siblings, 1 reply; 3+ messages in thread
From: Christian König @ 2020-03-13 10:17 UTC (permalink / raw)
  To: Lucas Stach, dri-devel; +Cc: Robert Beckett, kernel, patchwork-lst

Am 13.03.20 um 10:56 schrieb Lucas Stach:
> From: Robert Beckett <bob.beckett@collabora.com>
>
> Add a new trace event to show when jobs are run on the HW.
>
> Signed-off-by: Robert Beckett <bob.beckett@collabora.com>
> Signed-off-by: Lucas Stach <l.stach@pengutronix.de>

There is also the scheduled fence we could used for this, but this trace 
point adds a few extra fields which might be useful.

Acked-by: Christian König <christian.koenig@amd.com>

> ---
>   .../gpu/drm/scheduler/gpu_scheduler_trace.h   | 27 +++++++++++++++++++
>   drivers/gpu/drm/scheduler/sched_main.c        |  1 +
>   2 files changed, 28 insertions(+)
>
> diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> index d79086498aff..877ce9b127f1 100644
> --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> @@ -59,6 +59,33 @@ TRACE_EVENT(drm_sched_job,
>   		      __entry->job_count, __entry->hw_job_count)
>   );
>   
> +TRACE_EVENT(drm_run_job,
> +	    TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
> +	    TP_ARGS(sched_job, entity),
> +	    TP_STRUCT__entry(
> +			     __field(struct drm_sched_entity *, entity)
> +			     __field(struct dma_fence *, fence)
> +			     __field(const char *, name)
> +			     __field(uint64_t, id)
> +			     __field(u32, job_count)
> +			     __field(int, hw_job_count)
> +			     ),
> +
> +	    TP_fast_assign(
> +			   __entry->entity = entity;
> +			   __entry->id = sched_job->id;
> +			   __entry->fence = &sched_job->s_fence->finished;
> +			   __entry->name = sched_job->sched->name;
> +			   __entry->job_count = spsc_queue_count(&entity->job_queue);
> +			   __entry->hw_job_count = atomic_read(
> +				   &sched_job->sched->hw_rq_count);
> +			   ),
> +	    TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d",
> +		      __entry->entity, __entry->id,
> +		      __entry->fence, __entry->name,
> +		      __entry->job_count, __entry->hw_job_count)
> +);
> +
>   TRACE_EVENT(drm_sched_process_job,
>   	    TP_PROTO(struct drm_sched_fence *fence),
>   	    TP_ARGS(fence),
> diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c
> index 71ce6215956f..34231b7163cc 100644
> --- a/drivers/gpu/drm/scheduler/sched_main.c
> +++ b/drivers/gpu/drm/scheduler/sched_main.c
> @@ -773,6 +773,7 @@ static int drm_sched_main(void *param)
>   		atomic_inc(&sched->hw_rq_count);
>   		drm_sched_job_begin(sched_job);
>   
> +		trace_drm_run_job(sched_job, entity);
>   		fence = sched->ops->run_job(sched_job);
>   		drm_sched_fence_scheduled(s_fence);
>   

_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [PATCH] drm/sched: add run job trace
  2020-03-13 10:17 ` Christian König
@ 2020-03-13 14:25   ` Alex Deucher
  0 siblings, 0 replies; 3+ messages in thread
From: Alex Deucher @ 2020-03-13 14:25 UTC (permalink / raw)
  To: Christian König
  Cc: Robert Beckett, patchwork-lst, Sascha Hauer,
	Maling list - DRI developers

On Fri, Mar 13, 2020 at 6:18 AM Christian König
<christian.koenig@amd.com> wrote:
>
> Am 13.03.20 um 10:56 schrieb Lucas Stach:
> > From: Robert Beckett <bob.beckett@collabora.com>
> >
> > Add a new trace event to show when jobs are run on the HW.
> >
> > Signed-off-by: Robert Beckett <bob.beckett@collabora.com>
> > Signed-off-by: Lucas Stach <l.stach@pengutronix.de>
>
> There is also the scheduled fence we could used for this, but this trace
> point adds a few extra fields which might be useful.
>
> Acked-by: Christian König <christian.koenig@amd.com>

Applied.  thanks!

Alex

>
> > ---
> >   .../gpu/drm/scheduler/gpu_scheduler_trace.h   | 27 +++++++++++++++++++
> >   drivers/gpu/drm/scheduler/sched_main.c        |  1 +
> >   2 files changed, 28 insertions(+)
> >
> > diff --git a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> > index d79086498aff..877ce9b127f1 100644
> > --- a/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> > +++ b/drivers/gpu/drm/scheduler/gpu_scheduler_trace.h
> > @@ -59,6 +59,33 @@ TRACE_EVENT(drm_sched_job,
> >                     __entry->job_count, __entry->hw_job_count)
> >   );
> >
> > +TRACE_EVENT(drm_run_job,
> > +         TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
> > +         TP_ARGS(sched_job, entity),
> > +         TP_STRUCT__entry(
> > +                          __field(struct drm_sched_entity *, entity)
> > +                          __field(struct dma_fence *, fence)
> > +                          __field(const char *, name)
> > +                          __field(uint64_t, id)
> > +                          __field(u32, job_count)
> > +                          __field(int, hw_job_count)
> > +                          ),
> > +
> > +         TP_fast_assign(
> > +                        __entry->entity = entity;
> > +                        __entry->id = sched_job->id;
> > +                        __entry->fence = &sched_job->s_fence->finished;
> > +                        __entry->name = sched_job->sched->name;
> > +                        __entry->job_count = spsc_queue_count(&entity->job_queue);
> > +                        __entry->hw_job_count = atomic_read(
> > +                                &sched_job->sched->hw_rq_count);
> > +                        ),
> > +         TP_printk("entity=%p, id=%llu, fence=%p, ring=%s, job count:%u, hw job count:%d",
> > +                   __entry->entity, __entry->id,
> > +                   __entry->fence, __entry->name,
> > +                   __entry->job_count, __entry->hw_job_count)
> > +);
> > +
> >   TRACE_EVENT(drm_sched_process_job,
> >           TP_PROTO(struct drm_sched_fence *fence),
> >           TP_ARGS(fence),
> > diff --git a/drivers/gpu/drm/scheduler/sched_main.c b/drivers/gpu/drm/scheduler/sched_main.c
> > index 71ce6215956f..34231b7163cc 100644
> > --- a/drivers/gpu/drm/scheduler/sched_main.c
> > +++ b/drivers/gpu/drm/scheduler/sched_main.c
> > @@ -773,6 +773,7 @@ static int drm_sched_main(void *param)
> >               atomic_inc(&sched->hw_rq_count);
> >               drm_sched_job_begin(sched_job);
> >
> > +             trace_drm_run_job(sched_job, entity);
> >               fence = sched->ops->run_job(sched_job);
> >               drm_sched_fence_scheduled(s_fence);
> >
>
> _______________________________________________
> dri-devel mailing list
> dri-devel@lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/dri-devel
_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2020-03-13 14:25 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-03-13  9:56 [PATCH] drm/sched: add run job trace Lucas Stach
2020-03-13 10:17 ` Christian König
2020-03-13 14:25   ` Alex Deucher

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.