* [PATCH] drm/panfrost: Don't corrupt the queue mutex on open/close
@ 2020-10-29 17:00 Steven Price
2020-10-30 9:18 ` Boris Brezillon
0 siblings, 1 reply; 2+ messages in thread
From: Steven Price @ 2020-10-29 17:00 UTC (permalink / raw)
To: Daniel Vetter, David Airlie, Rob Herring, Tomeu Vizoso
Cc: Alyssa Rosenzweig, dri-devel, linux-kernel, Steven Price,
Boris Brezillon
The mutex within the panfrost_queue_state should have the lifetime of
the queue, however it was erroneously initialised/destroyed during
panfrost_job_{open,close} which is called every time a client
opens/closes the drm node.
Move the initialisation/destruction to panfrost_job_{init,fini} where it
belongs.
Fixes: 1a11a88cfd9a ("drm/panfrost: Fix job timeout handling")
Signed-off-by: Steven Price <steven.price@arm.com>
Reviewed-by: Boris Brezillon <boris.brezillon@collabora.com>
---
drivers/gpu/drm/panfrost/panfrost_job.c | 8 +++++---
1 file changed, 5 insertions(+), 3 deletions(-)
diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c
index cfb431624eea..145ad37eda6a 100644
--- a/drivers/gpu/drm/panfrost/panfrost_job.c
+++ b/drivers/gpu/drm/panfrost/panfrost_job.c
@@ -595,6 +595,8 @@ int panfrost_job_init(struct panfrost_device *pfdev)
}
for (j = 0; j < NUM_JOB_SLOTS; j++) {
+ mutex_init(&js->queue[j].lock);
+
js->queue[j].fence_context = dma_fence_context_alloc(1);
ret = drm_sched_init(&js->queue[j].sched,
@@ -625,8 +627,10 @@ void panfrost_job_fini(struct panfrost_device *pfdev)
job_write(pfdev, JOB_INT_MASK, 0);
- for (j = 0; j < NUM_JOB_SLOTS; j++)
+ for (j = 0; j < NUM_JOB_SLOTS; j++) {
drm_sched_fini(&js->queue[j].sched);
+ mutex_destroy(&js->queue[j].lock);
+ }
}
@@ -638,7 +642,6 @@ int panfrost_job_open(struct panfrost_file_priv *panfrost_priv)
int ret, i;
for (i = 0; i < NUM_JOB_SLOTS; i++) {
- mutex_init(&js->queue[i].lock);
sched = &js->queue[i].sched;
ret = drm_sched_entity_init(&panfrost_priv->sched_entity[i],
DRM_SCHED_PRIORITY_NORMAL, &sched,
@@ -657,7 +660,6 @@ void panfrost_job_close(struct panfrost_file_priv *panfrost_priv)
for (i = 0; i < NUM_JOB_SLOTS; i++) {
drm_sched_entity_destroy(&panfrost_priv->sched_entity[i]);
- mutex_destroy(&js->queue[i].lock);
/* Ensure any timeouts relating to this client have completed */
flush_delayed_work(&js->queue[i].sched.work_tdr);
}
--
2.20.1
^ permalink raw reply related [flat|nested] 2+ messages in thread
* Re: [PATCH] drm/panfrost: Don't corrupt the queue mutex on open/close
2020-10-29 17:00 [PATCH] drm/panfrost: Don't corrupt the queue mutex on open/close Steven Price
@ 2020-10-30 9:18 ` Boris Brezillon
0 siblings, 0 replies; 2+ messages in thread
From: Boris Brezillon @ 2020-10-30 9:18 UTC (permalink / raw)
To: Steven Price
Cc: Daniel Vetter, David Airlie, Rob Herring, Tomeu Vizoso,
Alyssa Rosenzweig, dri-devel, linux-kernel
On Thu, 29 Oct 2020 17:00:47 +0000
Steven Price <steven.price@arm.com> wrote:
> The mutex within the panfrost_queue_state should have the lifetime of
> the queue, however it was erroneously initialised/destroyed during
> panfrost_job_{open,close} which is called every time a client
> opens/closes the drm node.
>
> Move the initialisation/destruction to panfrost_job_{init,fini} where it
> belongs.
>
Queued to drm-misc-next.
Thanks,
Boris
> Fixes: 1a11a88cfd9a ("drm/panfrost: Fix job timeout handling")
> Signed-off-by: Steven Price <steven.price@arm.com>
> Reviewed-by: Boris Brezillon <boris.brezillon@collabora.com>
> ---
> drivers/gpu/drm/panfrost/panfrost_job.c | 8 +++++---
> 1 file changed, 5 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/gpu/drm/panfrost/panfrost_job.c b/drivers/gpu/drm/panfrost/panfrost_job.c
> index cfb431624eea..145ad37eda6a 100644
> --- a/drivers/gpu/drm/panfrost/panfrost_job.c
> +++ b/drivers/gpu/drm/panfrost/panfrost_job.c
> @@ -595,6 +595,8 @@ int panfrost_job_init(struct panfrost_device *pfdev)
> }
>
> for (j = 0; j < NUM_JOB_SLOTS; j++) {
> + mutex_init(&js->queue[j].lock);
> +
> js->queue[j].fence_context = dma_fence_context_alloc(1);
>
> ret = drm_sched_init(&js->queue[j].sched,
> @@ -625,8 +627,10 @@ void panfrost_job_fini(struct panfrost_device *pfdev)
>
> job_write(pfdev, JOB_INT_MASK, 0);
>
> - for (j = 0; j < NUM_JOB_SLOTS; j++)
> + for (j = 0; j < NUM_JOB_SLOTS; j++) {
> drm_sched_fini(&js->queue[j].sched);
> + mutex_destroy(&js->queue[j].lock);
> + }
>
> }
>
> @@ -638,7 +642,6 @@ int panfrost_job_open(struct panfrost_file_priv *panfrost_priv)
> int ret, i;
>
> for (i = 0; i < NUM_JOB_SLOTS; i++) {
> - mutex_init(&js->queue[i].lock);
> sched = &js->queue[i].sched;
> ret = drm_sched_entity_init(&panfrost_priv->sched_entity[i],
> DRM_SCHED_PRIORITY_NORMAL, &sched,
> @@ -657,7 +660,6 @@ void panfrost_job_close(struct panfrost_file_priv *panfrost_priv)
>
> for (i = 0; i < NUM_JOB_SLOTS; i++) {
> drm_sched_entity_destroy(&panfrost_priv->sched_entity[i]);
> - mutex_destroy(&js->queue[i].lock);
> /* Ensure any timeouts relating to this client have completed */
> flush_delayed_work(&js->queue[i].sched.work_tdr);
> }
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2020-10-30 9:18 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-10-29 17:00 [PATCH] drm/panfrost: Don't corrupt the queue mutex on open/close Steven Price
2020-10-30 9:18 ` Boris Brezillon
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).