All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH] dma-buf: fix and rework dma_buf_poll v3
@ 2021-06-22 13:04 Christian König
  2021-06-22 13:07 ` Christian König
  0 siblings, 1 reply; 10+ messages in thread
From: Christian König @ 2021-06-22 13:04 UTC (permalink / raw)
  To: daniel.vetter, dri-devel

Daniel pointed me towards this function and there are multiple obvious problems
in the implementation.

First of all the retry loop is not working as intended. In general the retry
makes only sense if you grab the reference first and then check the sequence
values.

It's also good practice to keep the reference around when installing callbacks
to fences you don't own.

And last the whole implementation was unnecessary complex and rather hard to
understand which could lead to probably unexpected behavior of the IOCTL.

Fix all this by reworking the implementation from scratch.

Only mildly tested and needs a thoughtful review of the code.

v2: fix the reference counting as well
v3: keep the excl fence handling as is for stable

Signed-off-by: Christian König <christian.koenig@amd.com>
CC: stable@vger.kernel.org
---
 drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
 include/linux/dma-buf.h   |   2 +-
 2 files changed, 55 insertions(+), 80 deletions(-)

diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
index eadd1eaa2fb5..e97c3a9d98d5 100644
--- a/drivers/dma-buf/dma-buf.c
+++ b/drivers/dma-buf/dma-buf.c
@@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
 	 * If you hit this BUG() it means someone dropped their ref to the
 	 * dma-buf while still having pending operation to the buffer.
 	 */
-	BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
+	BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
 
 	dmabuf->ops->release(dmabuf);
 
@@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
 	wake_up_locked_poll(dcb->poll, dcb->active);
 	dcb->active = 0;
 	spin_unlock_irqrestore(&dcb->poll->lock, flags);
+	dma_fence_put(fence);
 }
 
 static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
 {
+	struct dma_buf_poll_cb_t *dcb;
 	struct dma_buf *dmabuf;
 	struct dma_resv *resv;
 	struct dma_resv_list *fobj;
 	struct dma_fence *fence_excl;
-	__poll_t events;
 	unsigned shared_count, seq;
+	struct dma_fence *fence;
+	__poll_t events;
+	int r, i;
 
 	dmabuf = file->private_data;
 	if (!dmabuf || !dmabuf->resv)
@@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
 	if (!events)
 		return 0;
 
+	dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
+
+	/* Only queue a new one if we are not still waiting for the old one */
+	spin_lock_irq(&dmabuf->poll.lock);
+	if (dcb->active)
+		events = 0;
+	else
+		dcb->active = events;
+	spin_unlock_irq(&dmabuf->poll.lock);
+	if (!events)
+		return 0;
+
 retry:
 	seq = read_seqcount_begin(&resv->seq);
 	rcu_read_lock();
 
 	fobj = rcu_dereference(resv->fence);
-	if (fobj)
+	if (fobj && events & EPOLLOUT)
 		shared_count = fobj->shared_count;
 	else
 		shared_count = 0;
-	fence_excl = rcu_dereference(resv->fence_excl);
-	if (read_seqcount_retry(&resv->seq, seq)) {
-		rcu_read_unlock();
-		goto retry;
-	}
 
-	if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
-		struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
-		__poll_t pevents = EPOLLIN;
-
-		if (shared_count == 0)
-			pevents |= EPOLLOUT;
-
-		spin_lock_irq(&dmabuf->poll.lock);
-		if (dcb->active) {
-			dcb->active |= pevents;
-			events &= ~pevents;
-		} else
-			dcb->active = pevents;
-		spin_unlock_irq(&dmabuf->poll.lock);
-
-		if (events & pevents) {
-			if (!dma_fence_get_rcu(fence_excl)) {
-				/* force a recheck */
-				events &= ~pevents;
-				dma_buf_poll_cb(NULL, &dcb->cb);
-			} else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
-							   dma_buf_poll_cb)) {
-				events &= ~pevents;
-				dma_fence_put(fence_excl);
-			} else {
-				/*
-				 * No callback queued, wake up any additional
-				 * waiters.
-				 */
-				dma_fence_put(fence_excl);
-				dma_buf_poll_cb(NULL, &dcb->cb);
-			}
+	for (i = 0; i < shared_count; ++i) {
+		fence = rcu_dereference(fobj->shared[i]);
+		fence = dma_fence_get_rcu(fence);
+		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
+			/* Concurrent modify detected, force re-check */
+			dma_fence_put(fence);
+			rcu_read_unlock();
+			goto retry;
 		}
-	}
 
-	if ((events & EPOLLOUT) && shared_count > 0) {
-		struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
-		int i;
-
-		/* Only queue a new callback if no event has fired yet */
-		spin_lock_irq(&dmabuf->poll.lock);
-		if (dcb->active)
-			events &= ~EPOLLOUT;
-		else
-			dcb->active = EPOLLOUT;
-		spin_unlock_irq(&dmabuf->poll.lock);
-
-		if (!(events & EPOLLOUT))
+		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
+		if (!r) {
+			/* Callback queued */
+			events = 0;
 			goto out;
+		}
+		dma_fence_put(fence);
+	}
 
-		for (i = 0; i < shared_count; ++i) {
-			struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
-
-			if (!dma_fence_get_rcu(fence)) {
-				/*
-				 * fence refcount dropped to zero, this means
-				 * that fobj has been freed
-				 *
-				 * call dma_buf_poll_cb and force a recheck!
-				 */
-				events &= ~EPOLLOUT;
-				dma_buf_poll_cb(NULL, &dcb->cb);
-				break;
-			}
-			if (!dma_fence_add_callback(fence, &dcb->cb,
-						    dma_buf_poll_cb)) {
-				dma_fence_put(fence);
-				events &= ~EPOLLOUT;
-				break;
-			}
+	fence = dma_resv_get_excl(resv);
+	if (fence && shared_count == 0) {
+		fence = dma_fence_get_rcu(fence);
+		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
+			/* Concurrent modify detected, force re-check */
 			dma_fence_put(fence);
+			rcu_read_unlock();
+			goto retry;
+
 		}
 
-		/* No callback queued, wake up any additional waiters. */
-		if (i == shared_count)
-			dma_buf_poll_cb(NULL, &dcb->cb);
+		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
+		if (!r) {
+			/* Callback queued */
+			events = 0;
+			goto out;
+		}
+		dma_fence_put(fence_excl);
 	}
 
+	/* No callback queued, wake up any additional waiters. */
+	dma_buf_poll_cb(NULL, &dcb->cb);
+
 out:
 	rcu_read_unlock();
 	return events;
@@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
 	dmabuf->owner = exp_info->owner;
 	spin_lock_init(&dmabuf->name_lock);
 	init_waitqueue_head(&dmabuf->poll);
-	dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
-	dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
+	dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
+	dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
 
 	if (!resv) {
 		resv = (struct dma_resv *)&dmabuf[1];
diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
index efdc56b9d95f..7e747ad54c81 100644
--- a/include/linux/dma-buf.h
+++ b/include/linux/dma-buf.h
@@ -329,7 +329,7 @@ struct dma_buf {
 		wait_queue_head_t *poll;
 
 		__poll_t active;
-	} cb_excl, cb_shared;
+	} cb_in, cb_out;
 };
 
 /**
-- 
2.25.1


^ permalink raw reply related	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-22 13:04 [PATCH] dma-buf: fix and rework dma_buf_poll v3 Christian König
@ 2021-06-22 13:07 ` Christian König
  2021-06-22 17:02   ` Daniel Vetter
  0 siblings, 1 reply; 10+ messages in thread
From: Christian König @ 2021-06-22 13:07 UTC (permalink / raw)
  To: daniel.vetter, dri-devel

Crap, hit enter to early before adding a cover letter.

This is the same patch as before, but as requested I'm keeping the 
exclusive fence handling as it is for now.

Daniel can you double check this and/or make sure that it is tested?

I only smoke tested it and the code is so complicated that I'm not sure 
I catched all side effects.

Regards,
Christian.

Am 22.06.21 um 15:04 schrieb Christian König:
> Daniel pointed me towards this function and there are multiple obvious problems
> in the implementation.
>
> First of all the retry loop is not working as intended. In general the retry
> makes only sense if you grab the reference first and then check the sequence
> values.
>
> It's also good practice to keep the reference around when installing callbacks
> to fences you don't own.
>
> And last the whole implementation was unnecessary complex and rather hard to
> understand which could lead to probably unexpected behavior of the IOCTL.
>
> Fix all this by reworking the implementation from scratch.
>
> Only mildly tested and needs a thoughtful review of the code.
>
> v2: fix the reference counting as well
> v3: keep the excl fence handling as is for stable
>
> Signed-off-by: Christian König <christian.koenig@amd.com>
> CC: stable@vger.kernel.org
> ---
>   drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
>   include/linux/dma-buf.h   |   2 +-
>   2 files changed, 55 insertions(+), 80 deletions(-)
>
> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> index eadd1eaa2fb5..e97c3a9d98d5 100644
> --- a/drivers/dma-buf/dma-buf.c
> +++ b/drivers/dma-buf/dma-buf.c
> @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
>   	 * If you hit this BUG() it means someone dropped their ref to the
>   	 * dma-buf while still having pending operation to the buffer.
>   	 */
> -	BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
> +	BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
>   
>   	dmabuf->ops->release(dmabuf);
>   
> @@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>   	wake_up_locked_poll(dcb->poll, dcb->active);
>   	dcb->active = 0;
>   	spin_unlock_irqrestore(&dcb->poll->lock, flags);
> +	dma_fence_put(fence);
>   }
>   
>   static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
>   {
> +	struct dma_buf_poll_cb_t *dcb;
>   	struct dma_buf *dmabuf;
>   	struct dma_resv *resv;
>   	struct dma_resv_list *fobj;
>   	struct dma_fence *fence_excl;
> -	__poll_t events;
>   	unsigned shared_count, seq;
> +	struct dma_fence *fence;
> +	__poll_t events;
> +	int r, i;
>   
>   	dmabuf = file->private_data;
>   	if (!dmabuf || !dmabuf->resv)
> @@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
>   	if (!events)
>   		return 0;
>   
> +	dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
> +
> +	/* Only queue a new one if we are not still waiting for the old one */
> +	spin_lock_irq(&dmabuf->poll.lock);
> +	if (dcb->active)
> +		events = 0;
> +	else
> +		dcb->active = events;
> +	spin_unlock_irq(&dmabuf->poll.lock);
> +	if (!events)
> +		return 0;
> +
>   retry:
>   	seq = read_seqcount_begin(&resv->seq);
>   	rcu_read_lock();
>   
>   	fobj = rcu_dereference(resv->fence);
> -	if (fobj)
> +	if (fobj && events & EPOLLOUT)
>   		shared_count = fobj->shared_count;
>   	else
>   		shared_count = 0;
> -	fence_excl = rcu_dereference(resv->fence_excl);
> -	if (read_seqcount_retry(&resv->seq, seq)) {
> -		rcu_read_unlock();
> -		goto retry;
> -	}
>   
> -	if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
> -		struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
> -		__poll_t pevents = EPOLLIN;
> -
> -		if (shared_count == 0)
> -			pevents |= EPOLLOUT;
> -
> -		spin_lock_irq(&dmabuf->poll.lock);
> -		if (dcb->active) {
> -			dcb->active |= pevents;
> -			events &= ~pevents;
> -		} else
> -			dcb->active = pevents;
> -		spin_unlock_irq(&dmabuf->poll.lock);
> -
> -		if (events & pevents) {
> -			if (!dma_fence_get_rcu(fence_excl)) {
> -				/* force a recheck */
> -				events &= ~pevents;
> -				dma_buf_poll_cb(NULL, &dcb->cb);
> -			} else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
> -							   dma_buf_poll_cb)) {
> -				events &= ~pevents;
> -				dma_fence_put(fence_excl);
> -			} else {
> -				/*
> -				 * No callback queued, wake up any additional
> -				 * waiters.
> -				 */
> -				dma_fence_put(fence_excl);
> -				dma_buf_poll_cb(NULL, &dcb->cb);
> -			}
> +	for (i = 0; i < shared_count; ++i) {
> +		fence = rcu_dereference(fobj->shared[i]);
> +		fence = dma_fence_get_rcu(fence);
> +		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> +			/* Concurrent modify detected, force re-check */
> +			dma_fence_put(fence);
> +			rcu_read_unlock();
> +			goto retry;
>   		}
> -	}
>   
> -	if ((events & EPOLLOUT) && shared_count > 0) {
> -		struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
> -		int i;
> -
> -		/* Only queue a new callback if no event has fired yet */
> -		spin_lock_irq(&dmabuf->poll.lock);
> -		if (dcb->active)
> -			events &= ~EPOLLOUT;
> -		else
> -			dcb->active = EPOLLOUT;
> -		spin_unlock_irq(&dmabuf->poll.lock);
> -
> -		if (!(events & EPOLLOUT))
> +		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> +		if (!r) {
> +			/* Callback queued */
> +			events = 0;
>   			goto out;
> +		}
> +		dma_fence_put(fence);
> +	}
>   
> -		for (i = 0; i < shared_count; ++i) {
> -			struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
> -
> -			if (!dma_fence_get_rcu(fence)) {
> -				/*
> -				 * fence refcount dropped to zero, this means
> -				 * that fobj has been freed
> -				 *
> -				 * call dma_buf_poll_cb and force a recheck!
> -				 */
> -				events &= ~EPOLLOUT;
> -				dma_buf_poll_cb(NULL, &dcb->cb);
> -				break;
> -			}
> -			if (!dma_fence_add_callback(fence, &dcb->cb,
> -						    dma_buf_poll_cb)) {
> -				dma_fence_put(fence);
> -				events &= ~EPOLLOUT;
> -				break;
> -			}
> +	fence = dma_resv_get_excl(resv);
> +	if (fence && shared_count == 0) {
> +		fence = dma_fence_get_rcu(fence);
> +		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> +			/* Concurrent modify detected, force re-check */
>   			dma_fence_put(fence);
> +			rcu_read_unlock();
> +			goto retry;
> +
>   		}
>   
> -		/* No callback queued, wake up any additional waiters. */
> -		if (i == shared_count)
> -			dma_buf_poll_cb(NULL, &dcb->cb);
> +		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> +		if (!r) {
> +			/* Callback queued */
> +			events = 0;
> +			goto out;
> +		}
> +		dma_fence_put(fence_excl);
>   	}
>   
> +	/* No callback queued, wake up any additional waiters. */
> +	dma_buf_poll_cb(NULL, &dcb->cb);
> +
>   out:
>   	rcu_read_unlock();
>   	return events;
> @@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
>   	dmabuf->owner = exp_info->owner;
>   	spin_lock_init(&dmabuf->name_lock);
>   	init_waitqueue_head(&dmabuf->poll);
> -	dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
> -	dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
> +	dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
> +	dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
>   
>   	if (!resv) {
>   		resv = (struct dma_resv *)&dmabuf[1];
> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
> index efdc56b9d95f..7e747ad54c81 100644
> --- a/include/linux/dma-buf.h
> +++ b/include/linux/dma-buf.h
> @@ -329,7 +329,7 @@ struct dma_buf {
>   		wait_queue_head_t *poll;
>   
>   		__poll_t active;
> -	} cb_excl, cb_shared;
> +	} cb_in, cb_out;
>   };
>   
>   /**


^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-22 13:07 ` Christian König
@ 2021-06-22 17:02   ` Daniel Vetter
  2021-06-23 11:17     ` Christian König
  0 siblings, 1 reply; 10+ messages in thread
From: Daniel Vetter @ 2021-06-22 17:02 UTC (permalink / raw)
  To: Christian König, Michel Dänzer, Pekka Paalanen, Simon Ser
  Cc: dri-devel

On Tue, Jun 22, 2021 at 3:07 PM Christian König
<ckoenig.leichtzumerken@gmail.com> wrote:
>
> Crap, hit enter to early before adding a cover letter.
>
> This is the same patch as before, but as requested I'm keeping the
> exclusive fence handling as it is for now.
>
> Daniel can you double check this and/or make sure that it is tested?
>
> I only smoke tested it and the code is so complicated that I'm not sure
> I catched all side effects.

So I've thought about this some more, and we actually have docs for
how this is supposed to work:

https://dri.freedesktop.org/docs/drm/driver-api/dma-buf.html#implicit-fence-poll-support

Docs are pretty clear that we want both read and write for EPOLLOUT or
well both exclusive and shared fences. So I guess back to your actual
thing, but maybe we should get some acks from userspace people for it
(Michel, Pekka, Simon probably usual suspects).

The other thing I brought up and I haven't seen you reply to (maybe
missed it) is whether we shouldn't just use dma_resv_get_fences(). We
need to do the refcounting anyway, and this avoids us having to
open-code this very nasty code. Finally, and imo most important, this
is what's also used in drm_gem_fence_array_add_implicit(), which many
drivers use to handle their implicit in-fences. So would be nice to
have exactly matching code between that and what dma-buf poll does for
"can I read" and "can I write".

Thoughts?
-Daniel

>
> Regards,
> Christian.
>
> Am 22.06.21 um 15:04 schrieb Christian König:
> > Daniel pointed me towards this function and there are multiple obvious problems
> > in the implementation.
> >
> > First of all the retry loop is not working as intended. In general the retry
> > makes only sense if you grab the reference first and then check the sequence
> > values.
> >
> > It's also good practice to keep the reference around when installing callbacks
> > to fences you don't own.
> >
> > And last the whole implementation was unnecessary complex and rather hard to
> > understand which could lead to probably unexpected behavior of the IOCTL.
> >
> > Fix all this by reworking the implementation from scratch.
> >
> > Only mildly tested and needs a thoughtful review of the code.
> >
> > v2: fix the reference counting as well
> > v3: keep the excl fence handling as is for stable
> >
> > Signed-off-by: Christian König <christian.koenig@amd.com>
> > CC: stable@vger.kernel.org
> > ---
> >   drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
> >   include/linux/dma-buf.h   |   2 +-
> >   2 files changed, 55 insertions(+), 80 deletions(-)
> >
> > diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> > index eadd1eaa2fb5..e97c3a9d98d5 100644
> > --- a/drivers/dma-buf/dma-buf.c
> > +++ b/drivers/dma-buf/dma-buf.c
> > @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
> >        * If you hit this BUG() it means someone dropped their ref to the
> >        * dma-buf while still having pending operation to the buffer.
> >        */
> > -     BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
> > +     BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
> >
> >       dmabuf->ops->release(dmabuf);
> >
> > @@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
> >       wake_up_locked_poll(dcb->poll, dcb->active);
> >       dcb->active = 0;
> >       spin_unlock_irqrestore(&dcb->poll->lock, flags);
> > +     dma_fence_put(fence);
> >   }
> >
> >   static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
> >   {
> > +     struct dma_buf_poll_cb_t *dcb;
> >       struct dma_buf *dmabuf;
> >       struct dma_resv *resv;
> >       struct dma_resv_list *fobj;
> >       struct dma_fence *fence_excl;
> > -     __poll_t events;
> >       unsigned shared_count, seq;
> > +     struct dma_fence *fence;
> > +     __poll_t events;
> > +     int r, i;
> >
> >       dmabuf = file->private_data;
> >       if (!dmabuf || !dmabuf->resv)
> > @@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
> >       if (!events)
> >               return 0;
> >
> > +     dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
> > +
> > +     /* Only queue a new one if we are not still waiting for the old one */
> > +     spin_lock_irq(&dmabuf->poll.lock);
> > +     if (dcb->active)
> > +             events = 0;
> > +     else
> > +             dcb->active = events;
> > +     spin_unlock_irq(&dmabuf->poll.lock);
> > +     if (!events)
> > +             return 0;
> > +
> >   retry:
> >       seq = read_seqcount_begin(&resv->seq);
> >       rcu_read_lock();
> >
> >       fobj = rcu_dereference(resv->fence);
> > -     if (fobj)
> > +     if (fobj && events & EPOLLOUT)
> >               shared_count = fobj->shared_count;
> >       else
> >               shared_count = 0;
> > -     fence_excl = rcu_dereference(resv->fence_excl);
> > -     if (read_seqcount_retry(&resv->seq, seq)) {
> > -             rcu_read_unlock();
> > -             goto retry;
> > -     }
> >
> > -     if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
> > -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
> > -             __poll_t pevents = EPOLLIN;
> > -
> > -             if (shared_count == 0)
> > -                     pevents |= EPOLLOUT;
> > -
> > -             spin_lock_irq(&dmabuf->poll.lock);
> > -             if (dcb->active) {
> > -                     dcb->active |= pevents;
> > -                     events &= ~pevents;
> > -             } else
> > -                     dcb->active = pevents;
> > -             spin_unlock_irq(&dmabuf->poll.lock);
> > -
> > -             if (events & pevents) {
> > -                     if (!dma_fence_get_rcu(fence_excl)) {
> > -                             /* force a recheck */
> > -                             events &= ~pevents;
> > -                             dma_buf_poll_cb(NULL, &dcb->cb);
> > -                     } else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
> > -                                                        dma_buf_poll_cb)) {
> > -                             events &= ~pevents;
> > -                             dma_fence_put(fence_excl);
> > -                     } else {
> > -                             /*
> > -                              * No callback queued, wake up any additional
> > -                              * waiters.
> > -                              */
> > -                             dma_fence_put(fence_excl);
> > -                             dma_buf_poll_cb(NULL, &dcb->cb);
> > -                     }
> > +     for (i = 0; i < shared_count; ++i) {
> > +             fence = rcu_dereference(fobj->shared[i]);
> > +             fence = dma_fence_get_rcu(fence);
> > +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> > +                     /* Concurrent modify detected, force re-check */
> > +                     dma_fence_put(fence);
> > +                     rcu_read_unlock();
> > +                     goto retry;
> >               }
> > -     }
> >
> > -     if ((events & EPOLLOUT) && shared_count > 0) {
> > -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
> > -             int i;
> > -
> > -             /* Only queue a new callback if no event has fired yet */
> > -             spin_lock_irq(&dmabuf->poll.lock);
> > -             if (dcb->active)
> > -                     events &= ~EPOLLOUT;
> > -             else
> > -                     dcb->active = EPOLLOUT;
> > -             spin_unlock_irq(&dmabuf->poll.lock);
> > -
> > -             if (!(events & EPOLLOUT))
> > +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> > +             if (!r) {
> > +                     /* Callback queued */
> > +                     events = 0;
> >                       goto out;
> > +             }
> > +             dma_fence_put(fence);
> > +     }
> >
> > -             for (i = 0; i < shared_count; ++i) {
> > -                     struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
> > -
> > -                     if (!dma_fence_get_rcu(fence)) {
> > -                             /*
> > -                              * fence refcount dropped to zero, this means
> > -                              * that fobj has been freed
> > -                              *
> > -                              * call dma_buf_poll_cb and force a recheck!
> > -                              */
> > -                             events &= ~EPOLLOUT;
> > -                             dma_buf_poll_cb(NULL, &dcb->cb);
> > -                             break;
> > -                     }
> > -                     if (!dma_fence_add_callback(fence, &dcb->cb,
> > -                                                 dma_buf_poll_cb)) {
> > -                             dma_fence_put(fence);
> > -                             events &= ~EPOLLOUT;
> > -                             break;
> > -                     }
> > +     fence = dma_resv_get_excl(resv);
> > +     if (fence && shared_count == 0) {
> > +             fence = dma_fence_get_rcu(fence);
> > +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> > +                     /* Concurrent modify detected, force re-check */
> >                       dma_fence_put(fence);
> > +                     rcu_read_unlock();
> > +                     goto retry;
> > +
> >               }
> >
> > -             /* No callback queued, wake up any additional waiters. */
> > -             if (i == shared_count)
> > -                     dma_buf_poll_cb(NULL, &dcb->cb);
> > +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> > +             if (!r) {
> > +                     /* Callback queued */
> > +                     events = 0;
> > +                     goto out;
> > +             }
> > +             dma_fence_put(fence_excl);
> >       }
> >
> > +     /* No callback queued, wake up any additional waiters. */
> > +     dma_buf_poll_cb(NULL, &dcb->cb);
> > +
> >   out:
> >       rcu_read_unlock();
> >       return events;
> > @@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
> >       dmabuf->owner = exp_info->owner;
> >       spin_lock_init(&dmabuf->name_lock);
> >       init_waitqueue_head(&dmabuf->poll);
> > -     dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
> > -     dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
> > +     dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
> > +     dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
> >
> >       if (!resv) {
> >               resv = (struct dma_resv *)&dmabuf[1];
> > diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
> > index efdc56b9d95f..7e747ad54c81 100644
> > --- a/include/linux/dma-buf.h
> > +++ b/include/linux/dma-buf.h
> > @@ -329,7 +329,7 @@ struct dma_buf {
> >               wait_queue_head_t *poll;
> >
> >               __poll_t active;
> > -     } cb_excl, cb_shared;
> > +     } cb_in, cb_out;
> >   };
> >
> >   /**
>


-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-22 13:04 [PATCH] dma-buf: fix and rework dma_buf_poll v3 Christian König
  2021-06-22 13:07 ` Christian König
@ 2021-06-23  9:28 ` Dan Carpenter
  0 siblings, 0 replies; 10+ messages in thread
From: kernel test robot @ 2021-06-23  1:23 UTC (permalink / raw)
  To: kbuild

[-- Attachment #1: Type: text/plain, Size: 8414 bytes --]

CC: kbuild-all(a)lists.01.org
In-Reply-To: <20210622130459.122723-1-christian.koenig@amd.com>
References: <20210622130459.122723-1-christian.koenig@amd.com>
TO: "Christian König" <ckoenig.leichtzumerken@gmail.com>
TO: daniel.vetter(a)ffwll.ch
TO: dri-devel(a)lists.freedesktop.org

Hi "Christian,

I love your patch! Perhaps something to improve:

[auto build test WARNING on tegra-drm/drm/tegra/for-next]
[also build test WARNING on linus/master v5.13-rc7]
[cannot apply to next-20210622]
[If your patch is applied to the wrong git tree, kindly drop us a note.
And when submitting patch, we suggest to use '--base' as documented in
https://git-scm.com/docs/git-format-patch]

url:    https://github.com/0day-ci/linux/commits/Christian-K-nig/dma-buf-fix-and-rework-dma_buf_poll-v3/20210622-210643
base:   git://anongit.freedesktop.org/tegra/linux.git drm/tegra/for-next
:::::: branch date: 12 hours ago
:::::: commit date: 12 hours ago
config: i386-randconfig-m021-20210622 (attached as .config)
compiler: gcc-9 (Debian 9.3.0-22) 9.3.0

If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp@intel.com>
Reported-by: Dan Carpenter <dan.carpenter@oracle.com>

smatch warnings:
drivers/dma-buf/dma-buf.c:290 dma_buf_poll() error: uninitialized symbol 'fence_excl'.

vim +/fence_excl +290 drivers/dma-buf/dma-buf.c

9b495a5887994a Maarten Lankhorst 2014-07-01  207  
afc9a42b7464f7 Al Viro           2017-07-03  208  static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
9b495a5887994a Maarten Lankhorst 2014-07-01  209  {
d40fce0f010662 Christian König   2021-06-22  210  	struct dma_buf_poll_cb_t *dcb;
9b495a5887994a Maarten Lankhorst 2014-07-01  211  	struct dma_buf *dmabuf;
52791eeec1d9f4 Christian König   2019-08-11  212  	struct dma_resv *resv;
52791eeec1d9f4 Christian König   2019-08-11  213  	struct dma_resv_list *fobj;
f54d1867005c33 Chris Wilson      2016-10-25  214  	struct dma_fence *fence_excl;
b016cd6ed4b772 Chris Wilson      2019-08-14  215  	unsigned shared_count, seq;
d40fce0f010662 Christian König   2021-06-22  216  	struct dma_fence *fence;
d40fce0f010662 Christian König   2021-06-22  217  	__poll_t events;
d40fce0f010662 Christian König   2021-06-22  218  	int r, i;
9b495a5887994a Maarten Lankhorst 2014-07-01  219  
9b495a5887994a Maarten Lankhorst 2014-07-01  220  	dmabuf = file->private_data;
9b495a5887994a Maarten Lankhorst 2014-07-01  221  	if (!dmabuf || !dmabuf->resv)
a9a08845e9acbd Linus Torvalds    2018-02-11  222  		return EPOLLERR;
9b495a5887994a Maarten Lankhorst 2014-07-01  223  
9b495a5887994a Maarten Lankhorst 2014-07-01  224  	resv = dmabuf->resv;
9b495a5887994a Maarten Lankhorst 2014-07-01  225  
9b495a5887994a Maarten Lankhorst 2014-07-01  226  	poll_wait(file, &dmabuf->poll, poll);
9b495a5887994a Maarten Lankhorst 2014-07-01  227  
a9a08845e9acbd Linus Torvalds    2018-02-11  228  	events = poll_requested_events(poll) & (EPOLLIN | EPOLLOUT);
9b495a5887994a Maarten Lankhorst 2014-07-01  229  	if (!events)
9b495a5887994a Maarten Lankhorst 2014-07-01  230  		return 0;
9b495a5887994a Maarten Lankhorst 2014-07-01  231  
d40fce0f010662 Christian König   2021-06-22  232  	dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
d40fce0f010662 Christian König   2021-06-22  233  
d40fce0f010662 Christian König   2021-06-22  234  	/* Only queue a new one if we are not still waiting for the old one */
d40fce0f010662 Christian König   2021-06-22  235  	spin_lock_irq(&dmabuf->poll.lock);
d40fce0f010662 Christian König   2021-06-22  236  	if (dcb->active)
d40fce0f010662 Christian König   2021-06-22  237  		events = 0;
d40fce0f010662 Christian König   2021-06-22  238  	else
d40fce0f010662 Christian König   2021-06-22  239  		dcb->active = events;
d40fce0f010662 Christian König   2021-06-22  240  	spin_unlock_irq(&dmabuf->poll.lock);
d40fce0f010662 Christian König   2021-06-22  241  	if (!events)
d40fce0f010662 Christian König   2021-06-22  242  		return 0;
d40fce0f010662 Christian König   2021-06-22  243  
b016cd6ed4b772 Chris Wilson      2019-08-14  244  retry:
b016cd6ed4b772 Chris Wilson      2019-08-14  245  	seq = read_seqcount_begin(&resv->seq);
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  246  	rcu_read_lock();
b016cd6ed4b772 Chris Wilson      2019-08-14  247  
b016cd6ed4b772 Chris Wilson      2019-08-14  248  	fobj = rcu_dereference(resv->fence);
d40fce0f010662 Christian König   2021-06-22  249  	if (fobj && events & EPOLLOUT)
b016cd6ed4b772 Chris Wilson      2019-08-14  250  		shared_count = fobj->shared_count;
b016cd6ed4b772 Chris Wilson      2019-08-14  251  	else
b016cd6ed4b772 Chris Wilson      2019-08-14  252  		shared_count = 0;
d40fce0f010662 Christian König   2021-06-22  253  
d40fce0f010662 Christian König   2021-06-22  254  	for (i = 0; i < shared_count; ++i) {
d40fce0f010662 Christian König   2021-06-22  255  		fence = rcu_dereference(fobj->shared[i]);
d40fce0f010662 Christian König   2021-06-22  256  		fence = dma_fence_get_rcu(fence);
d40fce0f010662 Christian König   2021-06-22  257  		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
d40fce0f010662 Christian König   2021-06-22  258  			/* Concurrent modify detected, force re-check */
d40fce0f010662 Christian König   2021-06-22  259  			dma_fence_put(fence);
b016cd6ed4b772 Chris Wilson      2019-08-14  260  			rcu_read_unlock();
b016cd6ed4b772 Chris Wilson      2019-08-14  261  			goto retry;
b016cd6ed4b772 Chris Wilson      2019-08-14  262  		}
b016cd6ed4b772 Chris Wilson      2019-08-14  263  
d40fce0f010662 Christian König   2021-06-22  264  		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
d40fce0f010662 Christian König   2021-06-22  265  		if (!r) {
d40fce0f010662 Christian König   2021-06-22  266  			/* Callback queued */
d40fce0f010662 Christian König   2021-06-22  267  			events = 0;
d40fce0f010662 Christian König   2021-06-22  268  			goto out;
9b495a5887994a Maarten Lankhorst 2014-07-01  269  		}
d40fce0f010662 Christian König   2021-06-22  270  		dma_fence_put(fence);
04a5faa8cbe5a8 Maarten Lankhorst 2014-07-01  271  	}
9b495a5887994a Maarten Lankhorst 2014-07-01  272  
d40fce0f010662 Christian König   2021-06-22  273  	fence = dma_resv_get_excl(resv);
d40fce0f010662 Christian König   2021-06-22  274  	if (fence && shared_count == 0) {
d40fce0f010662 Christian König   2021-06-22  275  		fence = dma_fence_get_rcu(fence);
d40fce0f010662 Christian König   2021-06-22  276  		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
d40fce0f010662 Christian König   2021-06-22  277  			/* Concurrent modify detected, force re-check */
d40fce0f010662 Christian König   2021-06-22  278  			dma_fence_put(fence);
d40fce0f010662 Christian König   2021-06-22  279  			rcu_read_unlock();
d40fce0f010662 Christian König   2021-06-22  280  			goto retry;
9b495a5887994a Maarten Lankhorst 2014-07-01  281  
d40fce0f010662 Christian König   2021-06-22  282  		}
9b495a5887994a Maarten Lankhorst 2014-07-01  283  
d40fce0f010662 Christian König   2021-06-22  284  		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
d40fce0f010662 Christian König   2021-06-22  285  		if (!r) {
d40fce0f010662 Christian König   2021-06-22  286  			/* Callback queued */
d40fce0f010662 Christian König   2021-06-22  287  			events = 0;
9b495a5887994a Maarten Lankhorst 2014-07-01  288  			goto out;
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  289  		}
d40fce0f010662 Christian König   2021-06-22 @290  		dma_fence_put(fence_excl);
04a5faa8cbe5a8 Maarten Lankhorst 2014-07-01  291  	}
9b495a5887994a Maarten Lankhorst 2014-07-01  292  
9b495a5887994a Maarten Lankhorst 2014-07-01  293  	/* No callback queued, wake up any additional waiters. */
9b495a5887994a Maarten Lankhorst 2014-07-01  294  	dma_buf_poll_cb(NULL, &dcb->cb);
9b495a5887994a Maarten Lankhorst 2014-07-01  295  
9b495a5887994a Maarten Lankhorst 2014-07-01  296  out:
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  297  	rcu_read_unlock();
9b495a5887994a Maarten Lankhorst 2014-07-01  298  	return events;
9b495a5887994a Maarten Lankhorst 2014-07-01  299  }
9b495a5887994a Maarten Lankhorst 2014-07-01  300  

---
0-DAY CI Kernel Test Service, Intel Corporation
https://lists.01.org/hyperkitty/list/kbuild-all(a)lists.01.org

[-- Attachment #2: config.gz --]
[-- Type: application/gzip, Size: 47964 bytes --]

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
@ 2021-06-23  9:28 ` Dan Carpenter
  0 siblings, 0 replies; 10+ messages in thread
From: Dan Carpenter @ 2021-06-23  9:28 UTC (permalink / raw)
  To: kbuild, Christian König, daniel.vetter, dri-devel; +Cc: kbuild-all, lkp

Hi "Christian,

url:    https://github.com/0day-ci/linux/commits/Christian-K-nig/dma-buf-fix-and-rework-dma_buf_poll-v3/20210622-210643
base:   git://anongit.freedesktop.org/tegra/linux.git drm/tegra/for-next
config: i386-randconfig-m021-20210622 (attached as .config)
compiler: gcc-9 (Debian 9.3.0-22) 9.3.0

If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp@intel.com>
Reported-by: Dan Carpenter <dan.carpenter@oracle.com>

smatch warnings:
drivers/dma-buf/dma-buf.c:290 dma_buf_poll() error: uninitialized symbol 'fence_excl'.

vim +/fence_excl +290 drivers/dma-buf/dma-buf.c

afc9a42b7464f7 Al Viro           2017-07-03  208  static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
9b495a5887994a Maarten Lankhorst 2014-07-01  209  {
d40fce0f010662 Christian König   2021-06-22  210  	struct dma_buf_poll_cb_t *dcb;
9b495a5887994a Maarten Lankhorst 2014-07-01  211  	struct dma_buf *dmabuf;
52791eeec1d9f4 Christian König   2019-08-11  212  	struct dma_resv *resv;
52791eeec1d9f4 Christian König   2019-08-11  213  	struct dma_resv_list *fobj;
f54d1867005c33 Chris Wilson      2016-10-25  214  	struct dma_fence *fence_excl;
b016cd6ed4b772 Chris Wilson      2019-08-14  215  	unsigned shared_count, seq;
d40fce0f010662 Christian König   2021-06-22  216  	struct dma_fence *fence;
d40fce0f010662 Christian König   2021-06-22  217  	__poll_t events;
d40fce0f010662 Christian König   2021-06-22  218  	int r, i;
9b495a5887994a Maarten Lankhorst 2014-07-01  219  
9b495a5887994a Maarten Lankhorst 2014-07-01  220  	dmabuf = file->private_data;
9b495a5887994a Maarten Lankhorst 2014-07-01  221  	if (!dmabuf || !dmabuf->resv)
a9a08845e9acbd Linus Torvalds    2018-02-11  222  		return EPOLLERR;
9b495a5887994a Maarten Lankhorst 2014-07-01  223  
9b495a5887994a Maarten Lankhorst 2014-07-01  224  	resv = dmabuf->resv;
9b495a5887994a Maarten Lankhorst 2014-07-01  225  
9b495a5887994a Maarten Lankhorst 2014-07-01  226  	poll_wait(file, &dmabuf->poll, poll);
9b495a5887994a Maarten Lankhorst 2014-07-01  227  
a9a08845e9acbd Linus Torvalds    2018-02-11  228  	events = poll_requested_events(poll) & (EPOLLIN | EPOLLOUT);
9b495a5887994a Maarten Lankhorst 2014-07-01  229  	if (!events)
9b495a5887994a Maarten Lankhorst 2014-07-01  230  		return 0;
9b495a5887994a Maarten Lankhorst 2014-07-01  231  
d40fce0f010662 Christian König   2021-06-22  232  	dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
d40fce0f010662 Christian König   2021-06-22  233  
d40fce0f010662 Christian König   2021-06-22  234  	/* Only queue a new one if we are not still waiting for the old one */
d40fce0f010662 Christian König   2021-06-22  235  	spin_lock_irq(&dmabuf->poll.lock);
d40fce0f010662 Christian König   2021-06-22  236  	if (dcb->active)
d40fce0f010662 Christian König   2021-06-22  237  		events = 0;
d40fce0f010662 Christian König   2021-06-22  238  	else
d40fce0f010662 Christian König   2021-06-22  239  		dcb->active = events;
d40fce0f010662 Christian König   2021-06-22  240  	spin_unlock_irq(&dmabuf->poll.lock);
d40fce0f010662 Christian König   2021-06-22  241  	if (!events)
d40fce0f010662 Christian König   2021-06-22  242  		return 0;
d40fce0f010662 Christian König   2021-06-22  243  
b016cd6ed4b772 Chris Wilson      2019-08-14  244  retry:
b016cd6ed4b772 Chris Wilson      2019-08-14  245  	seq = read_seqcount_begin(&resv->seq);
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  246  	rcu_read_lock();
b016cd6ed4b772 Chris Wilson      2019-08-14  247  
b016cd6ed4b772 Chris Wilson      2019-08-14  248  	fobj = rcu_dereference(resv->fence);
d40fce0f010662 Christian König   2021-06-22  249  	if (fobj && events & EPOLLOUT)
b016cd6ed4b772 Chris Wilson      2019-08-14  250  		shared_count = fobj->shared_count;
b016cd6ed4b772 Chris Wilson      2019-08-14  251  	else
b016cd6ed4b772 Chris Wilson      2019-08-14  252  		shared_count = 0;
d40fce0f010662 Christian König   2021-06-22  253  
d40fce0f010662 Christian König   2021-06-22  254  	for (i = 0; i < shared_count; ++i) {
d40fce0f010662 Christian König   2021-06-22  255  		fence = rcu_dereference(fobj->shared[i]);
d40fce0f010662 Christian König   2021-06-22  256  		fence = dma_fence_get_rcu(fence);
d40fce0f010662 Christian König   2021-06-22  257  		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
d40fce0f010662 Christian König   2021-06-22  258  			/* Concurrent modify detected, force re-check */
d40fce0f010662 Christian König   2021-06-22  259  			dma_fence_put(fence);
b016cd6ed4b772 Chris Wilson      2019-08-14  260  			rcu_read_unlock();
b016cd6ed4b772 Chris Wilson      2019-08-14  261  			goto retry;
b016cd6ed4b772 Chris Wilson      2019-08-14  262  		}
b016cd6ed4b772 Chris Wilson      2019-08-14  263  
d40fce0f010662 Christian König   2021-06-22  264  		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
d40fce0f010662 Christian König   2021-06-22  265  		if (!r) {
d40fce0f010662 Christian König   2021-06-22  266  			/* Callback queued */
d40fce0f010662 Christian König   2021-06-22  267  			events = 0;
d40fce0f010662 Christian König   2021-06-22  268  			goto out;
9b495a5887994a Maarten Lankhorst 2014-07-01  269  		}
d40fce0f010662 Christian König   2021-06-22  270  		dma_fence_put(fence);
04a5faa8cbe5a8 Maarten Lankhorst 2014-07-01  271  	}
9b495a5887994a Maarten Lankhorst 2014-07-01  272  
d40fce0f010662 Christian König   2021-06-22  273  	fence = dma_resv_get_excl(resv);
d40fce0f010662 Christian König   2021-06-22  274  	if (fence && shared_count == 0) {
d40fce0f010662 Christian König   2021-06-22  275  		fence = dma_fence_get_rcu(fence);
d40fce0f010662 Christian König   2021-06-22  276  		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
d40fce0f010662 Christian König   2021-06-22  277  			/* Concurrent modify detected, force re-check */
d40fce0f010662 Christian König   2021-06-22  278  			dma_fence_put(fence);
d40fce0f010662 Christian König   2021-06-22  279  			rcu_read_unlock();
d40fce0f010662 Christian König   2021-06-22  280  			goto retry;
9b495a5887994a Maarten Lankhorst 2014-07-01  281  
d40fce0f010662 Christian König   2021-06-22  282  		}
9b495a5887994a Maarten Lankhorst 2014-07-01  283  
d40fce0f010662 Christian König   2021-06-22  284  		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
d40fce0f010662 Christian König   2021-06-22  285  		if (!r) {
d40fce0f010662 Christian König   2021-06-22  286  			/* Callback queued */
d40fce0f010662 Christian König   2021-06-22  287  			events = 0;
9b495a5887994a Maarten Lankhorst 2014-07-01  288  			goto out;
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  289  		}
d40fce0f010662 Christian König   2021-06-22 @290  		dma_fence_put(fence_excl);
                                                                              ^^^^^^^^^^
Never initialized.  Is part of the commit missing?

04a5faa8cbe5a8 Maarten Lankhorst 2014-07-01  291  	}
9b495a5887994a Maarten Lankhorst 2014-07-01  292  
9b495a5887994a Maarten Lankhorst 2014-07-01  293  	/* No callback queued, wake up any additional waiters. */
9b495a5887994a Maarten Lankhorst 2014-07-01  294  	dma_buf_poll_cb(NULL, &dcb->cb);
9b495a5887994a Maarten Lankhorst 2014-07-01  295  
9b495a5887994a Maarten Lankhorst 2014-07-01  296  out:
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  297  	rcu_read_unlock();
9b495a5887994a Maarten Lankhorst 2014-07-01  298  	return events;
9b495a5887994a Maarten Lankhorst 2014-07-01  299  }

---
0-DAY CI Kernel Test Service, Intel Corporation
https://lists.01.org/hyperkitty/list/kbuild-all@lists.01.org


^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
@ 2021-06-23  9:28 ` Dan Carpenter
  0 siblings, 0 replies; 10+ messages in thread
From: Dan Carpenter @ 2021-06-23  9:28 UTC (permalink / raw)
  To: kbuild-all

[-- Attachment #1: Type: text/plain, Size: 7708 bytes --]

Hi "Christian,

url:    https://github.com/0day-ci/linux/commits/Christian-K-nig/dma-buf-fix-and-rework-dma_buf_poll-v3/20210622-210643
base:   git://anongit.freedesktop.org/tegra/linux.git drm/tegra/for-next
config: i386-randconfig-m021-20210622 (attached as .config)
compiler: gcc-9 (Debian 9.3.0-22) 9.3.0

If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <lkp@intel.com>
Reported-by: Dan Carpenter <dan.carpenter@oracle.com>

smatch warnings:
drivers/dma-buf/dma-buf.c:290 dma_buf_poll() error: uninitialized symbol 'fence_excl'.

vim +/fence_excl +290 drivers/dma-buf/dma-buf.c

afc9a42b7464f7 Al Viro           2017-07-03  208  static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
9b495a5887994a Maarten Lankhorst 2014-07-01  209  {
d40fce0f010662 Christian König   2021-06-22  210  	struct dma_buf_poll_cb_t *dcb;
9b495a5887994a Maarten Lankhorst 2014-07-01  211  	struct dma_buf *dmabuf;
52791eeec1d9f4 Christian König   2019-08-11  212  	struct dma_resv *resv;
52791eeec1d9f4 Christian König   2019-08-11  213  	struct dma_resv_list *fobj;
f54d1867005c33 Chris Wilson      2016-10-25  214  	struct dma_fence *fence_excl;
b016cd6ed4b772 Chris Wilson      2019-08-14  215  	unsigned shared_count, seq;
d40fce0f010662 Christian König   2021-06-22  216  	struct dma_fence *fence;
d40fce0f010662 Christian König   2021-06-22  217  	__poll_t events;
d40fce0f010662 Christian König   2021-06-22  218  	int r, i;
9b495a5887994a Maarten Lankhorst 2014-07-01  219  
9b495a5887994a Maarten Lankhorst 2014-07-01  220  	dmabuf = file->private_data;
9b495a5887994a Maarten Lankhorst 2014-07-01  221  	if (!dmabuf || !dmabuf->resv)
a9a08845e9acbd Linus Torvalds    2018-02-11  222  		return EPOLLERR;
9b495a5887994a Maarten Lankhorst 2014-07-01  223  
9b495a5887994a Maarten Lankhorst 2014-07-01  224  	resv = dmabuf->resv;
9b495a5887994a Maarten Lankhorst 2014-07-01  225  
9b495a5887994a Maarten Lankhorst 2014-07-01  226  	poll_wait(file, &dmabuf->poll, poll);
9b495a5887994a Maarten Lankhorst 2014-07-01  227  
a9a08845e9acbd Linus Torvalds    2018-02-11  228  	events = poll_requested_events(poll) & (EPOLLIN | EPOLLOUT);
9b495a5887994a Maarten Lankhorst 2014-07-01  229  	if (!events)
9b495a5887994a Maarten Lankhorst 2014-07-01  230  		return 0;
9b495a5887994a Maarten Lankhorst 2014-07-01  231  
d40fce0f010662 Christian König   2021-06-22  232  	dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
d40fce0f010662 Christian König   2021-06-22  233  
d40fce0f010662 Christian König   2021-06-22  234  	/* Only queue a new one if we are not still waiting for the old one */
d40fce0f010662 Christian König   2021-06-22  235  	spin_lock_irq(&dmabuf->poll.lock);
d40fce0f010662 Christian König   2021-06-22  236  	if (dcb->active)
d40fce0f010662 Christian König   2021-06-22  237  		events = 0;
d40fce0f010662 Christian König   2021-06-22  238  	else
d40fce0f010662 Christian König   2021-06-22  239  		dcb->active = events;
d40fce0f010662 Christian König   2021-06-22  240  	spin_unlock_irq(&dmabuf->poll.lock);
d40fce0f010662 Christian König   2021-06-22  241  	if (!events)
d40fce0f010662 Christian König   2021-06-22  242  		return 0;
d40fce0f010662 Christian König   2021-06-22  243  
b016cd6ed4b772 Chris Wilson      2019-08-14  244  retry:
b016cd6ed4b772 Chris Wilson      2019-08-14  245  	seq = read_seqcount_begin(&resv->seq);
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  246  	rcu_read_lock();
b016cd6ed4b772 Chris Wilson      2019-08-14  247  
b016cd6ed4b772 Chris Wilson      2019-08-14  248  	fobj = rcu_dereference(resv->fence);
d40fce0f010662 Christian König   2021-06-22  249  	if (fobj && events & EPOLLOUT)
b016cd6ed4b772 Chris Wilson      2019-08-14  250  		shared_count = fobj->shared_count;
b016cd6ed4b772 Chris Wilson      2019-08-14  251  	else
b016cd6ed4b772 Chris Wilson      2019-08-14  252  		shared_count = 0;
d40fce0f010662 Christian König   2021-06-22  253  
d40fce0f010662 Christian König   2021-06-22  254  	for (i = 0; i < shared_count; ++i) {
d40fce0f010662 Christian König   2021-06-22  255  		fence = rcu_dereference(fobj->shared[i]);
d40fce0f010662 Christian König   2021-06-22  256  		fence = dma_fence_get_rcu(fence);
d40fce0f010662 Christian König   2021-06-22  257  		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
d40fce0f010662 Christian König   2021-06-22  258  			/* Concurrent modify detected, force re-check */
d40fce0f010662 Christian König   2021-06-22  259  			dma_fence_put(fence);
b016cd6ed4b772 Chris Wilson      2019-08-14  260  			rcu_read_unlock();
b016cd6ed4b772 Chris Wilson      2019-08-14  261  			goto retry;
b016cd6ed4b772 Chris Wilson      2019-08-14  262  		}
b016cd6ed4b772 Chris Wilson      2019-08-14  263  
d40fce0f010662 Christian König   2021-06-22  264  		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
d40fce0f010662 Christian König   2021-06-22  265  		if (!r) {
d40fce0f010662 Christian König   2021-06-22  266  			/* Callback queued */
d40fce0f010662 Christian König   2021-06-22  267  			events = 0;
d40fce0f010662 Christian König   2021-06-22  268  			goto out;
9b495a5887994a Maarten Lankhorst 2014-07-01  269  		}
d40fce0f010662 Christian König   2021-06-22  270  		dma_fence_put(fence);
04a5faa8cbe5a8 Maarten Lankhorst 2014-07-01  271  	}
9b495a5887994a Maarten Lankhorst 2014-07-01  272  
d40fce0f010662 Christian König   2021-06-22  273  	fence = dma_resv_get_excl(resv);
d40fce0f010662 Christian König   2021-06-22  274  	if (fence && shared_count == 0) {
d40fce0f010662 Christian König   2021-06-22  275  		fence = dma_fence_get_rcu(fence);
d40fce0f010662 Christian König   2021-06-22  276  		if (!fence || read_seqcount_retry(&resv->seq, seq)) {
d40fce0f010662 Christian König   2021-06-22  277  			/* Concurrent modify detected, force re-check */
d40fce0f010662 Christian König   2021-06-22  278  			dma_fence_put(fence);
d40fce0f010662 Christian König   2021-06-22  279  			rcu_read_unlock();
d40fce0f010662 Christian König   2021-06-22  280  			goto retry;
9b495a5887994a Maarten Lankhorst 2014-07-01  281  
d40fce0f010662 Christian König   2021-06-22  282  		}
9b495a5887994a Maarten Lankhorst 2014-07-01  283  
d40fce0f010662 Christian König   2021-06-22  284  		r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
d40fce0f010662 Christian König   2021-06-22  285  		if (!r) {
d40fce0f010662 Christian König   2021-06-22  286  			/* Callback queued */
d40fce0f010662 Christian König   2021-06-22  287  			events = 0;
9b495a5887994a Maarten Lankhorst 2014-07-01  288  			goto out;
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  289  		}
d40fce0f010662 Christian König   2021-06-22 @290  		dma_fence_put(fence_excl);
                                                                              ^^^^^^^^^^
Never initialized.  Is part of the commit missing?

04a5faa8cbe5a8 Maarten Lankhorst 2014-07-01  291  	}
9b495a5887994a Maarten Lankhorst 2014-07-01  292  
9b495a5887994a Maarten Lankhorst 2014-07-01  293  	/* No callback queued, wake up any additional waiters. */
9b495a5887994a Maarten Lankhorst 2014-07-01  294  	dma_buf_poll_cb(NULL, &dcb->cb);
9b495a5887994a Maarten Lankhorst 2014-07-01  295  
9b495a5887994a Maarten Lankhorst 2014-07-01  296  out:
3c3b177a9369b2 Maarten Lankhorst 2014-07-01  297  	rcu_read_unlock();
9b495a5887994a Maarten Lankhorst 2014-07-01  298  	return events;
9b495a5887994a Maarten Lankhorst 2014-07-01  299  }

---
0-DAY CI Kernel Test Service, Intel Corporation
https://lists.01.org/hyperkitty/list/kbuild-all(a)lists.01.org

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-22 17:02   ` Daniel Vetter
@ 2021-06-23 11:17     ` Christian König
  2021-06-23 11:30       ` Daniel Vetter
  0 siblings, 1 reply; 10+ messages in thread
From: Christian König @ 2021-06-23 11:17 UTC (permalink / raw)
  To: Daniel Vetter, Michel Dänzer, Pekka Paalanen, Simon Ser; +Cc: dri-devel

Am 22.06.21 um 19:02 schrieb Daniel Vetter:
> On Tue, Jun 22, 2021 at 3:07 PM Christian König
> <ckoenig.leichtzumerken@gmail.com> wrote:
>> Crap, hit enter to early before adding a cover letter.
>>
>> This is the same patch as before, but as requested I'm keeping the
>> exclusive fence handling as it is for now.
>>
>> Daniel can you double check this and/or make sure that it is tested?
>>
>> I only smoke tested it and the code is so complicated that I'm not sure
>> I catched all side effects.
> So I've thought about this some more, and we actually have docs for
> how this is supposed to work:
>
> https://dri.freedesktop.org/docs/drm/driver-api/dma-buf.html#implicit-fence-poll-support
>
> Docs are pretty clear that we want both read and write for EPOLLOUT or
> well both exclusive and shared fences. So I guess back to your actual
> thing, but maybe we should get some acks from userspace people for it
> (Michel, Pekka, Simon probably usual suspects).

Ok, sounds good to me. Going to send out a patch rebased to 
drm-misc-fixes today.

>
> The other thing I brought up and I haven't seen you reply to (maybe
> missed it) is whether we shouldn't just use dma_resv_get_fences(). We
> need to do the refcounting anyway, and this avoids us having to
> open-code this very nasty code. Finally, and imo most important, this
> is what's also used in drm_gem_fence_array_add_implicit(), which many
> drivers use to handle their implicit in-fences. So would be nice to
> have exactly matching code between that and what dma-buf poll does for
> "can I read" and "can I write".
>
> Thoughts?

Yeah, I've seen that. Just didn't had time to answer.

That goes into the same direction as my thinking that we need to 
centralize the RCU and synchronization handling in general.

What I don't like about the approach is that dma_resv_get_fences() needs 
to allocate memory. For a lot of use cases including dma_buf_poll that 
is rather overkill.

To unify the handling I think we should use the iterator I've create the 
prototype of. This way we only have an "for_write" parameter and the 
iterator in return gives you all the fences you need.

When you then extend that  approach we could say we have an enum 
describing the use case. Something like:
1. For explicit sync, just give me all the must sync fences from memory 
management.
2. For read, give me all the writers and memory management fences.
3. For write, give me all the readers and writers and memory management 
fences.
4. For memory management, give me everything including things like PTE 
updates/TLB flushes.

So instead of asking for some specific type of fence(s) the drivers 
tells the dma_resv object about their use case and in return get the 
fences they need to wait for.

This essentially means that we move the decision what to wait for from 
the drivers into the dma_resv object, which I think would be a massive 
improvement.

Functions like dma_resv_get_list(), dma_resv_get_excl(), 
dma_resv_get_excl_rcu() etc would then essentially be deprecated and 
instead you use dma_resv_get_fences(), dma_resv_for_each_fences(), 
dma_resv_wait_timeout(), dma_resv_test_signaled() with a proper use case.

What do you think?

Christian.

> -Daniel
>
>> Regards,
>> Christian.
>>
>> Am 22.06.21 um 15:04 schrieb Christian König:
>>> Daniel pointed me towards this function and there are multiple obvious problems
>>> in the implementation.
>>>
>>> First of all the retry loop is not working as intended. In general the retry
>>> makes only sense if you grab the reference first and then check the sequence
>>> values.
>>>
>>> It's also good practice to keep the reference around when installing callbacks
>>> to fences you don't own.
>>>
>>> And last the whole implementation was unnecessary complex and rather hard to
>>> understand which could lead to probably unexpected behavior of the IOCTL.
>>>
>>> Fix all this by reworking the implementation from scratch.
>>>
>>> Only mildly tested and needs a thoughtful review of the code.
>>>
>>> v2: fix the reference counting as well
>>> v3: keep the excl fence handling as is for stable
>>>
>>> Signed-off-by: Christian König <christian.koenig@amd.com>
>>> CC: stable@vger.kernel.org
>>> ---
>>>    drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
>>>    include/linux/dma-buf.h   |   2 +-
>>>    2 files changed, 55 insertions(+), 80 deletions(-)
>>>
>>> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
>>> index eadd1eaa2fb5..e97c3a9d98d5 100644
>>> --- a/drivers/dma-buf/dma-buf.c
>>> +++ b/drivers/dma-buf/dma-buf.c
>>> @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
>>>         * If you hit this BUG() it means someone dropped their ref to the
>>>         * dma-buf while still having pending operation to the buffer.
>>>         */
>>> -     BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
>>> +     BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
>>>
>>>        dmabuf->ops->release(dmabuf);
>>>
>>> @@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>>>        wake_up_locked_poll(dcb->poll, dcb->active);
>>>        dcb->active = 0;
>>>        spin_unlock_irqrestore(&dcb->poll->lock, flags);
>>> +     dma_fence_put(fence);
>>>    }
>>>
>>>    static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
>>>    {
>>> +     struct dma_buf_poll_cb_t *dcb;
>>>        struct dma_buf *dmabuf;
>>>        struct dma_resv *resv;
>>>        struct dma_resv_list *fobj;
>>>        struct dma_fence *fence_excl;
>>> -     __poll_t events;
>>>        unsigned shared_count, seq;
>>> +     struct dma_fence *fence;
>>> +     __poll_t events;
>>> +     int r, i;
>>>
>>>        dmabuf = file->private_data;
>>>        if (!dmabuf || !dmabuf->resv)
>>> @@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
>>>        if (!events)
>>>                return 0;
>>>
>>> +     dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
>>> +
>>> +     /* Only queue a new one if we are not still waiting for the old one */
>>> +     spin_lock_irq(&dmabuf->poll.lock);
>>> +     if (dcb->active)
>>> +             events = 0;
>>> +     else
>>> +             dcb->active = events;
>>> +     spin_unlock_irq(&dmabuf->poll.lock);
>>> +     if (!events)
>>> +             return 0;
>>> +
>>>    retry:
>>>        seq = read_seqcount_begin(&resv->seq);
>>>        rcu_read_lock();
>>>
>>>        fobj = rcu_dereference(resv->fence);
>>> -     if (fobj)
>>> +     if (fobj && events & EPOLLOUT)
>>>                shared_count = fobj->shared_count;
>>>        else
>>>                shared_count = 0;
>>> -     fence_excl = rcu_dereference(resv->fence_excl);
>>> -     if (read_seqcount_retry(&resv->seq, seq)) {
>>> -             rcu_read_unlock();
>>> -             goto retry;
>>> -     }
>>>
>>> -     if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
>>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
>>> -             __poll_t pevents = EPOLLIN;
>>> -
>>> -             if (shared_count == 0)
>>> -                     pevents |= EPOLLOUT;
>>> -
>>> -             spin_lock_irq(&dmabuf->poll.lock);
>>> -             if (dcb->active) {
>>> -                     dcb->active |= pevents;
>>> -                     events &= ~pevents;
>>> -             } else
>>> -                     dcb->active = pevents;
>>> -             spin_unlock_irq(&dmabuf->poll.lock);
>>> -
>>> -             if (events & pevents) {
>>> -                     if (!dma_fence_get_rcu(fence_excl)) {
>>> -                             /* force a recheck */
>>> -                             events &= ~pevents;
>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
>>> -                     } else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
>>> -                                                        dma_buf_poll_cb)) {
>>> -                             events &= ~pevents;
>>> -                             dma_fence_put(fence_excl);
>>> -                     } else {
>>> -                             /*
>>> -                              * No callback queued, wake up any additional
>>> -                              * waiters.
>>> -                              */
>>> -                             dma_fence_put(fence_excl);
>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
>>> -                     }
>>> +     for (i = 0; i < shared_count; ++i) {
>>> +             fence = rcu_dereference(fobj->shared[i]);
>>> +             fence = dma_fence_get_rcu(fence);
>>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
>>> +                     /* Concurrent modify detected, force re-check */
>>> +                     dma_fence_put(fence);
>>> +                     rcu_read_unlock();
>>> +                     goto retry;
>>>                }
>>> -     }
>>>
>>> -     if ((events & EPOLLOUT) && shared_count > 0) {
>>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
>>> -             int i;
>>> -
>>> -             /* Only queue a new callback if no event has fired yet */
>>> -             spin_lock_irq(&dmabuf->poll.lock);
>>> -             if (dcb->active)
>>> -                     events &= ~EPOLLOUT;
>>> -             else
>>> -                     dcb->active = EPOLLOUT;
>>> -             spin_unlock_irq(&dmabuf->poll.lock);
>>> -
>>> -             if (!(events & EPOLLOUT))
>>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
>>> +             if (!r) {
>>> +                     /* Callback queued */
>>> +                     events = 0;
>>>                        goto out;
>>> +             }
>>> +             dma_fence_put(fence);
>>> +     }
>>>
>>> -             for (i = 0; i < shared_count; ++i) {
>>> -                     struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
>>> -
>>> -                     if (!dma_fence_get_rcu(fence)) {
>>> -                             /*
>>> -                              * fence refcount dropped to zero, this means
>>> -                              * that fobj has been freed
>>> -                              *
>>> -                              * call dma_buf_poll_cb and force a recheck!
>>> -                              */
>>> -                             events &= ~EPOLLOUT;
>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
>>> -                             break;
>>> -                     }
>>> -                     if (!dma_fence_add_callback(fence, &dcb->cb,
>>> -                                                 dma_buf_poll_cb)) {
>>> -                             dma_fence_put(fence);
>>> -                             events &= ~EPOLLOUT;
>>> -                             break;
>>> -                     }
>>> +     fence = dma_resv_get_excl(resv);
>>> +     if (fence && shared_count == 0) {
>>> +             fence = dma_fence_get_rcu(fence);
>>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
>>> +                     /* Concurrent modify detected, force re-check */
>>>                        dma_fence_put(fence);
>>> +                     rcu_read_unlock();
>>> +                     goto retry;
>>> +
>>>                }
>>>
>>> -             /* No callback queued, wake up any additional waiters. */
>>> -             if (i == shared_count)
>>> -                     dma_buf_poll_cb(NULL, &dcb->cb);
>>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
>>> +             if (!r) {
>>> +                     /* Callback queued */
>>> +                     events = 0;
>>> +                     goto out;
>>> +             }
>>> +             dma_fence_put(fence_excl);
>>>        }
>>>
>>> +     /* No callback queued, wake up any additional waiters. */
>>> +     dma_buf_poll_cb(NULL, &dcb->cb);
>>> +
>>>    out:
>>>        rcu_read_unlock();
>>>        return events;
>>> @@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
>>>        dmabuf->owner = exp_info->owner;
>>>        spin_lock_init(&dmabuf->name_lock);
>>>        init_waitqueue_head(&dmabuf->poll);
>>> -     dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
>>> -     dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
>>> +     dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
>>> +     dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
>>>
>>>        if (!resv) {
>>>                resv = (struct dma_resv *)&dmabuf[1];
>>> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
>>> index efdc56b9d95f..7e747ad54c81 100644
>>> --- a/include/linux/dma-buf.h
>>> +++ b/include/linux/dma-buf.h
>>> @@ -329,7 +329,7 @@ struct dma_buf {
>>>                wait_queue_head_t *poll;
>>>
>>>                __poll_t active;
>>> -     } cb_excl, cb_shared;
>>> +     } cb_in, cb_out;
>>>    };
>>>
>>>    /**
>


^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-23 11:17     ` Christian König
@ 2021-06-23 11:30       ` Daniel Vetter
  2021-06-23 12:42         ` Christian König
  0 siblings, 1 reply; 10+ messages in thread
From: Daniel Vetter @ 2021-06-23 11:30 UTC (permalink / raw)
  To: Christian König; +Cc: Michel Dänzer, dri-devel

On Wed, Jun 23, 2021 at 1:17 PM Christian König
<ckoenig.leichtzumerken@gmail.com> wrote:
>
> Am 22.06.21 um 19:02 schrieb Daniel Vetter:
> > On Tue, Jun 22, 2021 at 3:07 PM Christian König
> > <ckoenig.leichtzumerken@gmail.com> wrote:
> >> Crap, hit enter to early before adding a cover letter.
> >>
> >> This is the same patch as before, but as requested I'm keeping the
> >> exclusive fence handling as it is for now.
> >>
> >> Daniel can you double check this and/or make sure that it is tested?
> >>
> >> I only smoke tested it and the code is so complicated that I'm not sure
> >> I catched all side effects.
> > So I've thought about this some more, and we actually have docs for
> > how this is supposed to work:
> >
> > https://dri.freedesktop.org/docs/drm/driver-api/dma-buf.html#implicit-fence-poll-support
> >
> > Docs are pretty clear that we want both read and write for EPOLLOUT or
> > well both exclusive and shared fences. So I guess back to your actual
> > thing, but maybe we should get some acks from userspace people for it
> > (Michel, Pekka, Simon probably usual suspects).
>
> Ok, sounds good to me. Going to send out a patch rebased to
> drm-misc-fixes today.
>
> >
> > The other thing I brought up and I haven't seen you reply to (maybe
> > missed it) is whether we shouldn't just use dma_resv_get_fences(). We
> > need to do the refcounting anyway, and this avoids us having to
> > open-code this very nasty code. Finally, and imo most important, this
> > is what's also used in drm_gem_fence_array_add_implicit(), which many
> > drivers use to handle their implicit in-fences. So would be nice to
> > have exactly matching code between that and what dma-buf poll does for
> > "can I read" and "can I write".
> >
> > Thoughts?
>
> Yeah, I've seen that. Just didn't had time to answer.
>
> That goes into the same direction as my thinking that we need to
> centralize the RCU and synchronization handling in general.
>
> What I don't like about the approach is that dma_resv_get_fences() needs
> to allocate memory. For a lot of use cases including dma_buf_poll that
> is rather overkill.
>
> To unify the handling I think we should use the iterator I've create the
> prototype of. This way we only have an "for_write" parameter and the
> iterator in return gives you all the fences you need.

Yeah I think in general I agree, especially in the CS code a bunch of
temporary allocations aren't great. But in dma_buf_poll I don't think
it's a place where anyone cares. That's why I think we can just use
that here and forget about all the trickiness. The gem helper otoh
wants an iterator (without retry even, since it's holding dma-resv
lock).

> When you then extend that  approach we could say we have an enum
> describing the use case. Something like:
> 1. For explicit sync, just give me all the must sync fences from memory
> management.
> 2. For read, give me all the writers and memory management fences.
> 3. For write, give me all the readers and writers and memory management
> fences.
> 4. For memory management, give me everything including things like PTE
> updates/TLB flushes.
>
> So instead of asking for some specific type of fence(s) the drivers
> tells the dma_resv object about their use case and in return get the
> fences they need to wait for.
>
> This essentially means that we move the decision what to wait for from
> the drivers into the dma_resv object, which I think would be a massive
> improvement.
>
> Functions like dma_resv_get_list(), dma_resv_get_excl(),
> dma_resv_get_excl_rcu() etc would then essentially be deprecated and
> instead you use dma_resv_get_fences(), dma_resv_for_each_fences(),
> dma_resv_wait_timeout(), dma_resv_test_signaled() with a proper use case.
>
> What do you think?

Yeah I think in general the direction makes sense, at least long term.
I think for now it's better to go with simplest solutions first until
we have everyone aligned on one set of rules, and have these rules
properly documented.
-Daniel

> Christian.
>
> > -Daniel
> >
> >> Regards,
> >> Christian.
> >>
> >> Am 22.06.21 um 15:04 schrieb Christian König:
> >>> Daniel pointed me towards this function and there are multiple obvious problems
> >>> in the implementation.
> >>>
> >>> First of all the retry loop is not working as intended. In general the retry
> >>> makes only sense if you grab the reference first and then check the sequence
> >>> values.
> >>>
> >>> It's also good practice to keep the reference around when installing callbacks
> >>> to fences you don't own.
> >>>
> >>> And last the whole implementation was unnecessary complex and rather hard to
> >>> understand which could lead to probably unexpected behavior of the IOCTL.
> >>>
> >>> Fix all this by reworking the implementation from scratch.
> >>>
> >>> Only mildly tested and needs a thoughtful review of the code.
> >>>
> >>> v2: fix the reference counting as well
> >>> v3: keep the excl fence handling as is for stable
> >>>
> >>> Signed-off-by: Christian König <christian.koenig@amd.com>
> >>> CC: stable@vger.kernel.org
> >>> ---
> >>>    drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
> >>>    include/linux/dma-buf.h   |   2 +-
> >>>    2 files changed, 55 insertions(+), 80 deletions(-)
> >>>
> >>> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> >>> index eadd1eaa2fb5..e97c3a9d98d5 100644
> >>> --- a/drivers/dma-buf/dma-buf.c
> >>> +++ b/drivers/dma-buf/dma-buf.c
> >>> @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
> >>>         * If you hit this BUG() it means someone dropped their ref to the
> >>>         * dma-buf while still having pending operation to the buffer.
> >>>         */
> >>> -     BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
> >>> +     BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
> >>>
> >>>        dmabuf->ops->release(dmabuf);
> >>>
> >>> @@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
> >>>        wake_up_locked_poll(dcb->poll, dcb->active);
> >>>        dcb->active = 0;
> >>>        spin_unlock_irqrestore(&dcb->poll->lock, flags);
> >>> +     dma_fence_put(fence);
> >>>    }
> >>>
> >>>    static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
> >>>    {
> >>> +     struct dma_buf_poll_cb_t *dcb;
> >>>        struct dma_buf *dmabuf;
> >>>        struct dma_resv *resv;
> >>>        struct dma_resv_list *fobj;
> >>>        struct dma_fence *fence_excl;
> >>> -     __poll_t events;
> >>>        unsigned shared_count, seq;
> >>> +     struct dma_fence *fence;
> >>> +     __poll_t events;
> >>> +     int r, i;
> >>>
> >>>        dmabuf = file->private_data;
> >>>        if (!dmabuf || !dmabuf->resv)
> >>> @@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
> >>>        if (!events)
> >>>                return 0;
> >>>
> >>> +     dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
> >>> +
> >>> +     /* Only queue a new one if we are not still waiting for the old one */
> >>> +     spin_lock_irq(&dmabuf->poll.lock);
> >>> +     if (dcb->active)
> >>> +             events = 0;
> >>> +     else
> >>> +             dcb->active = events;
> >>> +     spin_unlock_irq(&dmabuf->poll.lock);
> >>> +     if (!events)
> >>> +             return 0;
> >>> +
> >>>    retry:
> >>>        seq = read_seqcount_begin(&resv->seq);
> >>>        rcu_read_lock();
> >>>
> >>>        fobj = rcu_dereference(resv->fence);
> >>> -     if (fobj)
> >>> +     if (fobj && events & EPOLLOUT)
> >>>                shared_count = fobj->shared_count;
> >>>        else
> >>>                shared_count = 0;
> >>> -     fence_excl = rcu_dereference(resv->fence_excl);
> >>> -     if (read_seqcount_retry(&resv->seq, seq)) {
> >>> -             rcu_read_unlock();
> >>> -             goto retry;
> >>> -     }
> >>>
> >>> -     if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
> >>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
> >>> -             __poll_t pevents = EPOLLIN;
> >>> -
> >>> -             if (shared_count == 0)
> >>> -                     pevents |= EPOLLOUT;
> >>> -
> >>> -             spin_lock_irq(&dmabuf->poll.lock);
> >>> -             if (dcb->active) {
> >>> -                     dcb->active |= pevents;
> >>> -                     events &= ~pevents;
> >>> -             } else
> >>> -                     dcb->active = pevents;
> >>> -             spin_unlock_irq(&dmabuf->poll.lock);
> >>> -
> >>> -             if (events & pevents) {
> >>> -                     if (!dma_fence_get_rcu(fence_excl)) {
> >>> -                             /* force a recheck */
> >>> -                             events &= ~pevents;
> >>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
> >>> -                     } else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
> >>> -                                                        dma_buf_poll_cb)) {
> >>> -                             events &= ~pevents;
> >>> -                             dma_fence_put(fence_excl);
> >>> -                     } else {
> >>> -                             /*
> >>> -                              * No callback queued, wake up any additional
> >>> -                              * waiters.
> >>> -                              */
> >>> -                             dma_fence_put(fence_excl);
> >>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
> >>> -                     }
> >>> +     for (i = 0; i < shared_count; ++i) {
> >>> +             fence = rcu_dereference(fobj->shared[i]);
> >>> +             fence = dma_fence_get_rcu(fence);
> >>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> >>> +                     /* Concurrent modify detected, force re-check */
> >>> +                     dma_fence_put(fence);
> >>> +                     rcu_read_unlock();
> >>> +                     goto retry;
> >>>                }
> >>> -     }
> >>>
> >>> -     if ((events & EPOLLOUT) && shared_count > 0) {
> >>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
> >>> -             int i;
> >>> -
> >>> -             /* Only queue a new callback if no event has fired yet */
> >>> -             spin_lock_irq(&dmabuf->poll.lock);
> >>> -             if (dcb->active)
> >>> -                     events &= ~EPOLLOUT;
> >>> -             else
> >>> -                     dcb->active = EPOLLOUT;
> >>> -             spin_unlock_irq(&dmabuf->poll.lock);
> >>> -
> >>> -             if (!(events & EPOLLOUT))
> >>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> >>> +             if (!r) {
> >>> +                     /* Callback queued */
> >>> +                     events = 0;
> >>>                        goto out;
> >>> +             }
> >>> +             dma_fence_put(fence);
> >>> +     }
> >>>
> >>> -             for (i = 0; i < shared_count; ++i) {
> >>> -                     struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
> >>> -
> >>> -                     if (!dma_fence_get_rcu(fence)) {
> >>> -                             /*
> >>> -                              * fence refcount dropped to zero, this means
> >>> -                              * that fobj has been freed
> >>> -                              *
> >>> -                              * call dma_buf_poll_cb and force a recheck!
> >>> -                              */
> >>> -                             events &= ~EPOLLOUT;
> >>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
> >>> -                             break;
> >>> -                     }
> >>> -                     if (!dma_fence_add_callback(fence, &dcb->cb,
> >>> -                                                 dma_buf_poll_cb)) {
> >>> -                             dma_fence_put(fence);
> >>> -                             events &= ~EPOLLOUT;
> >>> -                             break;
> >>> -                     }
> >>> +     fence = dma_resv_get_excl(resv);
> >>> +     if (fence && shared_count == 0) {
> >>> +             fence = dma_fence_get_rcu(fence);
> >>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> >>> +                     /* Concurrent modify detected, force re-check */
> >>>                        dma_fence_put(fence);
> >>> +                     rcu_read_unlock();
> >>> +                     goto retry;
> >>> +
> >>>                }
> >>>
> >>> -             /* No callback queued, wake up any additional waiters. */
> >>> -             if (i == shared_count)
> >>> -                     dma_buf_poll_cb(NULL, &dcb->cb);
> >>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> >>> +             if (!r) {
> >>> +                     /* Callback queued */
> >>> +                     events = 0;
> >>> +                     goto out;
> >>> +             }
> >>> +             dma_fence_put(fence_excl);
> >>>        }
> >>>
> >>> +     /* No callback queued, wake up any additional waiters. */
> >>> +     dma_buf_poll_cb(NULL, &dcb->cb);
> >>> +
> >>>    out:
> >>>        rcu_read_unlock();
> >>>        return events;
> >>> @@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
> >>>        dmabuf->owner = exp_info->owner;
> >>>        spin_lock_init(&dmabuf->name_lock);
> >>>        init_waitqueue_head(&dmabuf->poll);
> >>> -     dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
> >>> -     dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
> >>> +     dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
> >>> +     dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
> >>>
> >>>        if (!resv) {
> >>>                resv = (struct dma_resv *)&dmabuf[1];
> >>> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
> >>> index efdc56b9d95f..7e747ad54c81 100644
> >>> --- a/include/linux/dma-buf.h
> >>> +++ b/include/linux/dma-buf.h
> >>> @@ -329,7 +329,7 @@ struct dma_buf {
> >>>                wait_queue_head_t *poll;
> >>>
> >>>                __poll_t active;
> >>> -     } cb_excl, cb_shared;
> >>> +     } cb_in, cb_out;
> >>>    };
> >>>
> >>>    /**
> >
>


-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch

^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-23 11:30       ` Daniel Vetter
@ 2021-06-23 12:42         ` Christian König
  2021-06-23 13:56           ` Daniel Vetter
  0 siblings, 1 reply; 10+ messages in thread
From: Christian König @ 2021-06-23 12:42 UTC (permalink / raw)
  To: Daniel Vetter; +Cc: Michel Dänzer, dri-devel

Am 23.06.21 um 13:30 schrieb Daniel Vetter:
> On Wed, Jun 23, 2021 at 1:17 PM Christian König
> <ckoenig.leichtzumerken@gmail.com> wrote:
>> Am 22.06.21 um 19:02 schrieb Daniel Vetter:
>>> On Tue, Jun 22, 2021 at 3:07 PM Christian König
>>> <ckoenig.leichtzumerken@gmail.com> wrote:
>>>> Crap, hit enter to early before adding a cover letter.
>>>>
>>>> This is the same patch as before, but as requested I'm keeping the
>>>> exclusive fence handling as it is for now.
>>>>
>>>> Daniel can you double check this and/or make sure that it is tested?
>>>>
>>>> I only smoke tested it and the code is so complicated that I'm not sure
>>>> I catched all side effects.
>>> So I've thought about this some more, and we actually have docs for
>>> how this is supposed to work:
>>>
>>> https://dri.freedesktop.org/docs/drm/driver-api/dma-buf.html#implicit-fence-poll-support
>>>
>>> Docs are pretty clear that we want both read and write for EPOLLOUT or
>>> well both exclusive and shared fences. So I guess back to your actual
>>> thing, but maybe we should get some acks from userspace people for it
>>> (Michel, Pekka, Simon probably usual suspects).
>> Ok, sounds good to me. Going to send out a patch rebased to
>> drm-misc-fixes today.
>>
>>> The other thing I brought up and I haven't seen you reply to (maybe
>>> missed it) is whether we shouldn't just use dma_resv_get_fences(). We
>>> need to do the refcounting anyway, and this avoids us having to
>>> open-code this very nasty code. Finally, and imo most important, this
>>> is what's also used in drm_gem_fence_array_add_implicit(), which many
>>> drivers use to handle their implicit in-fences. So would be nice to
>>> have exactly matching code between that and what dma-buf poll does for
>>> "can I read" and "can I write".
>>>
>>> Thoughts?
>> Yeah, I've seen that. Just didn't had time to answer.
>>
>> That goes into the same direction as my thinking that we need to
>> centralize the RCU and synchronization handling in general.
>>
>> What I don't like about the approach is that dma_resv_get_fences() needs
>> to allocate memory. For a lot of use cases including dma_buf_poll that
>> is rather overkill.
>>
>> To unify the handling I think we should use the iterator I've create the
>> prototype of. This way we only have an "for_write" parameter and the
>> iterator in return gives you all the fences you need.
> Yeah I think in general I agree, especially in the CS code a bunch of
> temporary allocations aren't great. But in dma_buf_poll I don't think
> it's a place where anyone cares. That's why I think we can just use
> that here and forget about all the trickiness. The gem helper otoh
> wants an iterator (without retry even, since it's holding dma-resv
> lock).

Well then I would rather say we make nails with heads and grab the 
reservation lock in dma_buf_poll.

That has at least less overhead than allocating memory, cause 
essentially what dma_buf_poll needs is a 
dma_resv_get_next_unsignaled_or_null_fence().

>> When you then extend that  approach we could say we have an enum
>> describing the use case. Something like:
>> 1. For explicit sync, just give me all the must sync fences from memory
>> management.
>> 2. For read, give me all the writers and memory management fences.
>> 3. For write, give me all the readers and writers and memory management
>> fences.
>> 4. For memory management, give me everything including things like PTE
>> updates/TLB flushes.
>>
>> So instead of asking for some specific type of fence(s) the drivers
>> tells the dma_resv object about their use case and in return get the
>> fences they need to wait for.
>>
>> This essentially means that we move the decision what to wait for from
>> the drivers into the dma_resv object, which I think would be a massive
>> improvement.
>>
>> Functions like dma_resv_get_list(), dma_resv_get_excl(),
>> dma_resv_get_excl_rcu() etc would then essentially be deprecated and
>> instead you use dma_resv_get_fences(), dma_resv_for_each_fences(),
>> dma_resv_wait_timeout(), dma_resv_test_signaled() with a proper use case.
>>
>> What do you think?
> Yeah I think in general the direction makes sense, at least long term.
> I think for now it's better to go with simplest solutions first until
> we have everyone aligned on one set of rules, and have these rules
> properly documented.

I think that looks rather good now, doesn't it?

Christian.

> -Daniel
>
>> Christian.
>>
>>> -Daniel
>>>
>>>> Regards,
>>>> Christian.
>>>>
>>>> Am 22.06.21 um 15:04 schrieb Christian König:
>>>>> Daniel pointed me towards this function and there are multiple obvious problems
>>>>> in the implementation.
>>>>>
>>>>> First of all the retry loop is not working as intended. In general the retry
>>>>> makes only sense if you grab the reference first and then check the sequence
>>>>> values.
>>>>>
>>>>> It's also good practice to keep the reference around when installing callbacks
>>>>> to fences you don't own.
>>>>>
>>>>> And last the whole implementation was unnecessary complex and rather hard to
>>>>> understand which could lead to probably unexpected behavior of the IOCTL.
>>>>>
>>>>> Fix all this by reworking the implementation from scratch.
>>>>>
>>>>> Only mildly tested and needs a thoughtful review of the code.
>>>>>
>>>>> v2: fix the reference counting as well
>>>>> v3: keep the excl fence handling as is for stable
>>>>>
>>>>> Signed-off-by: Christian König <christian.koenig@amd.com>
>>>>> CC: stable@vger.kernel.org
>>>>> ---
>>>>>     drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
>>>>>     include/linux/dma-buf.h   |   2 +-
>>>>>     2 files changed, 55 insertions(+), 80 deletions(-)
>>>>>
>>>>> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
>>>>> index eadd1eaa2fb5..e97c3a9d98d5 100644
>>>>> --- a/drivers/dma-buf/dma-buf.c
>>>>> +++ b/drivers/dma-buf/dma-buf.c
>>>>> @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
>>>>>          * If you hit this BUG() it means someone dropped their ref to the
>>>>>          * dma-buf while still having pending operation to the buffer.
>>>>>          */
>>>>> -     BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
>>>>> +     BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
>>>>>
>>>>>         dmabuf->ops->release(dmabuf);
>>>>>
>>>>> @@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
>>>>>         wake_up_locked_poll(dcb->poll, dcb->active);
>>>>>         dcb->active = 0;
>>>>>         spin_unlock_irqrestore(&dcb->poll->lock, flags);
>>>>> +     dma_fence_put(fence);
>>>>>     }
>>>>>
>>>>>     static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
>>>>>     {
>>>>> +     struct dma_buf_poll_cb_t *dcb;
>>>>>         struct dma_buf *dmabuf;
>>>>>         struct dma_resv *resv;
>>>>>         struct dma_resv_list *fobj;
>>>>>         struct dma_fence *fence_excl;
>>>>> -     __poll_t events;
>>>>>         unsigned shared_count, seq;
>>>>> +     struct dma_fence *fence;
>>>>> +     __poll_t events;
>>>>> +     int r, i;
>>>>>
>>>>>         dmabuf = file->private_data;
>>>>>         if (!dmabuf || !dmabuf->resv)
>>>>> @@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
>>>>>         if (!events)
>>>>>                 return 0;
>>>>>
>>>>> +     dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
>>>>> +
>>>>> +     /* Only queue a new one if we are not still waiting for the old one */
>>>>> +     spin_lock_irq(&dmabuf->poll.lock);
>>>>> +     if (dcb->active)
>>>>> +             events = 0;
>>>>> +     else
>>>>> +             dcb->active = events;
>>>>> +     spin_unlock_irq(&dmabuf->poll.lock);
>>>>> +     if (!events)
>>>>> +             return 0;
>>>>> +
>>>>>     retry:
>>>>>         seq = read_seqcount_begin(&resv->seq);
>>>>>         rcu_read_lock();
>>>>>
>>>>>         fobj = rcu_dereference(resv->fence);
>>>>> -     if (fobj)
>>>>> +     if (fobj && events & EPOLLOUT)
>>>>>                 shared_count = fobj->shared_count;
>>>>>         else
>>>>>                 shared_count = 0;
>>>>> -     fence_excl = rcu_dereference(resv->fence_excl);
>>>>> -     if (read_seqcount_retry(&resv->seq, seq)) {
>>>>> -             rcu_read_unlock();
>>>>> -             goto retry;
>>>>> -     }
>>>>>
>>>>> -     if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
>>>>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
>>>>> -             __poll_t pevents = EPOLLIN;
>>>>> -
>>>>> -             if (shared_count == 0)
>>>>> -                     pevents |= EPOLLOUT;
>>>>> -
>>>>> -             spin_lock_irq(&dmabuf->poll.lock);
>>>>> -             if (dcb->active) {
>>>>> -                     dcb->active |= pevents;
>>>>> -                     events &= ~pevents;
>>>>> -             } else
>>>>> -                     dcb->active = pevents;
>>>>> -             spin_unlock_irq(&dmabuf->poll.lock);
>>>>> -
>>>>> -             if (events & pevents) {
>>>>> -                     if (!dma_fence_get_rcu(fence_excl)) {
>>>>> -                             /* force a recheck */
>>>>> -                             events &= ~pevents;
>>>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
>>>>> -                     } else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
>>>>> -                                                        dma_buf_poll_cb)) {
>>>>> -                             events &= ~pevents;
>>>>> -                             dma_fence_put(fence_excl);
>>>>> -                     } else {
>>>>> -                             /*
>>>>> -                              * No callback queued, wake up any additional
>>>>> -                              * waiters.
>>>>> -                              */
>>>>> -                             dma_fence_put(fence_excl);
>>>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
>>>>> -                     }
>>>>> +     for (i = 0; i < shared_count; ++i) {
>>>>> +             fence = rcu_dereference(fobj->shared[i]);
>>>>> +             fence = dma_fence_get_rcu(fence);
>>>>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
>>>>> +                     /* Concurrent modify detected, force re-check */
>>>>> +                     dma_fence_put(fence);
>>>>> +                     rcu_read_unlock();
>>>>> +                     goto retry;
>>>>>                 }
>>>>> -     }
>>>>>
>>>>> -     if ((events & EPOLLOUT) && shared_count > 0) {
>>>>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
>>>>> -             int i;
>>>>> -
>>>>> -             /* Only queue a new callback if no event has fired yet */
>>>>> -             spin_lock_irq(&dmabuf->poll.lock);
>>>>> -             if (dcb->active)
>>>>> -                     events &= ~EPOLLOUT;
>>>>> -             else
>>>>> -                     dcb->active = EPOLLOUT;
>>>>> -             spin_unlock_irq(&dmabuf->poll.lock);
>>>>> -
>>>>> -             if (!(events & EPOLLOUT))
>>>>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
>>>>> +             if (!r) {
>>>>> +                     /* Callback queued */
>>>>> +                     events = 0;
>>>>>                         goto out;
>>>>> +             }
>>>>> +             dma_fence_put(fence);
>>>>> +     }
>>>>>
>>>>> -             for (i = 0; i < shared_count; ++i) {
>>>>> -                     struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
>>>>> -
>>>>> -                     if (!dma_fence_get_rcu(fence)) {
>>>>> -                             /*
>>>>> -                              * fence refcount dropped to zero, this means
>>>>> -                              * that fobj has been freed
>>>>> -                              *
>>>>> -                              * call dma_buf_poll_cb and force a recheck!
>>>>> -                              */
>>>>> -                             events &= ~EPOLLOUT;
>>>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
>>>>> -                             break;
>>>>> -                     }
>>>>> -                     if (!dma_fence_add_callback(fence, &dcb->cb,
>>>>> -                                                 dma_buf_poll_cb)) {
>>>>> -                             dma_fence_put(fence);
>>>>> -                             events &= ~EPOLLOUT;
>>>>> -                             break;
>>>>> -                     }
>>>>> +     fence = dma_resv_get_excl(resv);
>>>>> +     if (fence && shared_count == 0) {
>>>>> +             fence = dma_fence_get_rcu(fence);
>>>>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
>>>>> +                     /* Concurrent modify detected, force re-check */
>>>>>                         dma_fence_put(fence);
>>>>> +                     rcu_read_unlock();
>>>>> +                     goto retry;
>>>>> +
>>>>>                 }
>>>>>
>>>>> -             /* No callback queued, wake up any additional waiters. */
>>>>> -             if (i == shared_count)
>>>>> -                     dma_buf_poll_cb(NULL, &dcb->cb);
>>>>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
>>>>> +             if (!r) {
>>>>> +                     /* Callback queued */
>>>>> +                     events = 0;
>>>>> +                     goto out;
>>>>> +             }
>>>>> +             dma_fence_put(fence_excl);
>>>>>         }
>>>>>
>>>>> +     /* No callback queued, wake up any additional waiters. */
>>>>> +     dma_buf_poll_cb(NULL, &dcb->cb);
>>>>> +
>>>>>     out:
>>>>>         rcu_read_unlock();
>>>>>         return events;
>>>>> @@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
>>>>>         dmabuf->owner = exp_info->owner;
>>>>>         spin_lock_init(&dmabuf->name_lock);
>>>>>         init_waitqueue_head(&dmabuf->poll);
>>>>> -     dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
>>>>> -     dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
>>>>> +     dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
>>>>> +     dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
>>>>>
>>>>>         if (!resv) {
>>>>>                 resv = (struct dma_resv *)&dmabuf[1];
>>>>> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
>>>>> index efdc56b9d95f..7e747ad54c81 100644
>>>>> --- a/include/linux/dma-buf.h
>>>>> +++ b/include/linux/dma-buf.h
>>>>> @@ -329,7 +329,7 @@ struct dma_buf {
>>>>>                 wait_queue_head_t *poll;
>>>>>
>>>>>                 __poll_t active;
>>>>> -     } cb_excl, cb_shared;
>>>>> +     } cb_in, cb_out;
>>>>>     };
>>>>>
>>>>>     /**
>


^ permalink raw reply	[flat|nested] 10+ messages in thread

* Re: [PATCH] dma-buf: fix and rework dma_buf_poll v3
  2021-06-23 12:42         ` Christian König
@ 2021-06-23 13:56           ` Daniel Vetter
  0 siblings, 0 replies; 10+ messages in thread
From: Daniel Vetter @ 2021-06-23 13:56 UTC (permalink / raw)
  To: Christian König; +Cc: Michel Dänzer, dri-devel

On Wed, Jun 23, 2021 at 2:42 PM Christian König
<ckoenig.leichtzumerken@gmail.com> wrote:
>
> Am 23.06.21 um 13:30 schrieb Daniel Vetter:
> > On Wed, Jun 23, 2021 at 1:17 PM Christian König
> > <ckoenig.leichtzumerken@gmail.com> wrote:
> >> Am 22.06.21 um 19:02 schrieb Daniel Vetter:
> >>> On Tue, Jun 22, 2021 at 3:07 PM Christian König
> >>> <ckoenig.leichtzumerken@gmail.com> wrote:
> >>>> Crap, hit enter to early before adding a cover letter.
> >>>>
> >>>> This is the same patch as before, but as requested I'm keeping the
> >>>> exclusive fence handling as it is for now.
> >>>>
> >>>> Daniel can you double check this and/or make sure that it is tested?
> >>>>
> >>>> I only smoke tested it and the code is so complicated that I'm not sure
> >>>> I catched all side effects.
> >>> So I've thought about this some more, and we actually have docs for
> >>> how this is supposed to work:
> >>>
> >>> https://dri.freedesktop.org/docs/drm/driver-api/dma-buf.html#implicit-fence-poll-support
> >>>
> >>> Docs are pretty clear that we want both read and write for EPOLLOUT or
> >>> well both exclusive and shared fences. So I guess back to your actual
> >>> thing, but maybe we should get some acks from userspace people for it
> >>> (Michel, Pekka, Simon probably usual suspects).
> >> Ok, sounds good to me. Going to send out a patch rebased to
> >> drm-misc-fixes today.
> >>
> >>> The other thing I brought up and I haven't seen you reply to (maybe
> >>> missed it) is whether we shouldn't just use dma_resv_get_fences(). We
> >>> need to do the refcounting anyway, and this avoids us having to
> >>> open-code this very nasty code. Finally, and imo most important, this
> >>> is what's also used in drm_gem_fence_array_add_implicit(), which many
> >>> drivers use to handle their implicit in-fences. So would be nice to
> >>> have exactly matching code between that and what dma-buf poll does for
> >>> "can I read" and "can I write".
> >>>
> >>> Thoughts?
> >> Yeah, I've seen that. Just didn't had time to answer.
> >>
> >> That goes into the same direction as my thinking that we need to
> >> centralize the RCU and synchronization handling in general.
> >>
> >> What I don't like about the approach is that dma_resv_get_fences() needs
> >> to allocate memory. For a lot of use cases including dma_buf_poll that
> >> is rather overkill.
> >>
> >> To unify the handling I think we should use the iterator I've create the
> >> prototype of. This way we only have an "for_write" parameter and the
> >> iterator in return gives you all the fences you need.
> > Yeah I think in general I agree, especially in the CS code a bunch of
> > temporary allocations aren't great. But in dma_buf_poll I don't think
> > it's a place where anyone cares. That's why I think we can just use
> > that here and forget about all the trickiness. The gem helper otoh
> > wants an iterator (without retry even, since it's holding dma-resv
> > lock).
>
> Well then I would rather say we make nails with heads and grab the
> reservation lock in dma_buf_poll.
>
> That has at least less overhead than allocating memory, cause
> essentially what dma_buf_poll needs is a
> dma_resv_get_next_unsignaled_or_null_fence().

I'm all ok with that plan, avoids even more complexity.

> >> When you then extend that  approach we could say we have an enum
> >> describing the use case. Something like:
> >> 1. For explicit sync, just give me all the must sync fences from memory
> >> management.
> >> 2. For read, give me all the writers and memory management fences.
> >> 3. For write, give me all the readers and writers and memory management
> >> fences.
> >> 4. For memory management, give me everything including things like PTE
> >> updates/TLB flushes.
> >>
> >> So instead of asking for some specific type of fence(s) the drivers
> >> tells the dma_resv object about their use case and in return get the
> >> fences they need to wait for.
> >>
> >> This essentially means that we move the decision what to wait for from
> >> the drivers into the dma_resv object, which I think would be a massive
> >> improvement.
> >>
> >> Functions like dma_resv_get_list(), dma_resv_get_excl(),
> >> dma_resv_get_excl_rcu() etc would then essentially be deprecated and
> >> instead you use dma_resv_get_fences(), dma_resv_for_each_fences(),
> >> dma_resv_wait_timeout(), dma_resv_test_signaled() with a proper use case.
> >>
> >> What do you think?
> > Yeah I think in general the direction makes sense, at least long term.
> > I think for now it's better to go with simplest solutions first until
> > we have everyone aligned on one set of rules, and have these rules
> > properly documented.
>
> I think that looks rather good now, doesn't it?

Well we have 2 out of 3 pieces:
- ttm drivers need to wait in their pin: fixed&documented
- drivers need to follow the rules for setting dma_resv fences: amdgpu
fixed, patch set for other drivers + docs from me on the list
- drivers must not break the fence DAG in dma-resv: tbd, both
auditing/fixing drives and documenting it.

So getting there, but not yet fully arrived.
-Daniel

>
> Christian.
>
> > -Daniel
> >
> >> Christian.
> >>
> >>> -Daniel
> >>>
> >>>> Regards,
> >>>> Christian.
> >>>>
> >>>> Am 22.06.21 um 15:04 schrieb Christian König:
> >>>>> Daniel pointed me towards this function and there are multiple obvious problems
> >>>>> in the implementation.
> >>>>>
> >>>>> First of all the retry loop is not working as intended. In general the retry
> >>>>> makes only sense if you grab the reference first and then check the sequence
> >>>>> values.
> >>>>>
> >>>>> It's also good practice to keep the reference around when installing callbacks
> >>>>> to fences you don't own.
> >>>>>
> >>>>> And last the whole implementation was unnecessary complex and rather hard to
> >>>>> understand which could lead to probably unexpected behavior of the IOCTL.
> >>>>>
> >>>>> Fix all this by reworking the implementation from scratch.
> >>>>>
> >>>>> Only mildly tested and needs a thoughtful review of the code.
> >>>>>
> >>>>> v2: fix the reference counting as well
> >>>>> v3: keep the excl fence handling as is for stable
> >>>>>
> >>>>> Signed-off-by: Christian König <christian.koenig@amd.com>
> >>>>> CC: stable@vger.kernel.org
> >>>>> ---
> >>>>>     drivers/dma-buf/dma-buf.c | 133 ++++++++++++++++----------------------
> >>>>>     include/linux/dma-buf.h   |   2 +-
> >>>>>     2 files changed, 55 insertions(+), 80 deletions(-)
> >>>>>
> >>>>> diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c
> >>>>> index eadd1eaa2fb5..e97c3a9d98d5 100644
> >>>>> --- a/drivers/dma-buf/dma-buf.c
> >>>>> +++ b/drivers/dma-buf/dma-buf.c
> >>>>> @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry)
> >>>>>          * If you hit this BUG() it means someone dropped their ref to the
> >>>>>          * dma-buf while still having pending operation to the buffer.
> >>>>>          */
> >>>>> -     BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
> >>>>> +     BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
> >>>>>
> >>>>>         dmabuf->ops->release(dmabuf);
> >>>>>
> >>>>> @@ -202,16 +202,20 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
> >>>>>         wake_up_locked_poll(dcb->poll, dcb->active);
> >>>>>         dcb->active = 0;
> >>>>>         spin_unlock_irqrestore(&dcb->poll->lock, flags);
> >>>>> +     dma_fence_put(fence);
> >>>>>     }
> >>>>>
> >>>>>     static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
> >>>>>     {
> >>>>> +     struct dma_buf_poll_cb_t *dcb;
> >>>>>         struct dma_buf *dmabuf;
> >>>>>         struct dma_resv *resv;
> >>>>>         struct dma_resv_list *fobj;
> >>>>>         struct dma_fence *fence_excl;
> >>>>> -     __poll_t events;
> >>>>>         unsigned shared_count, seq;
> >>>>> +     struct dma_fence *fence;
> >>>>> +     __poll_t events;
> >>>>> +     int r, i;
> >>>>>
> >>>>>         dmabuf = file->private_data;
> >>>>>         if (!dmabuf || !dmabuf->resv)
> >>>>> @@ -225,99 +229,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll)
> >>>>>         if (!events)
> >>>>>                 return 0;
> >>>>>
> >>>>> +     dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
> >>>>> +
> >>>>> +     /* Only queue a new one if we are not still waiting for the old one */
> >>>>> +     spin_lock_irq(&dmabuf->poll.lock);
> >>>>> +     if (dcb->active)
> >>>>> +             events = 0;
> >>>>> +     else
> >>>>> +             dcb->active = events;
> >>>>> +     spin_unlock_irq(&dmabuf->poll.lock);
> >>>>> +     if (!events)
> >>>>> +             return 0;
> >>>>> +
> >>>>>     retry:
> >>>>>         seq = read_seqcount_begin(&resv->seq);
> >>>>>         rcu_read_lock();
> >>>>>
> >>>>>         fobj = rcu_dereference(resv->fence);
> >>>>> -     if (fobj)
> >>>>> +     if (fobj && events & EPOLLOUT)
> >>>>>                 shared_count = fobj->shared_count;
> >>>>>         else
> >>>>>                 shared_count = 0;
> >>>>> -     fence_excl = rcu_dereference(resv->fence_excl);
> >>>>> -     if (read_seqcount_retry(&resv->seq, seq)) {
> >>>>> -             rcu_read_unlock();
> >>>>> -             goto retry;
> >>>>> -     }
> >>>>>
> >>>>> -     if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
> >>>>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
> >>>>> -             __poll_t pevents = EPOLLIN;
> >>>>> -
> >>>>> -             if (shared_count == 0)
> >>>>> -                     pevents |= EPOLLOUT;
> >>>>> -
> >>>>> -             spin_lock_irq(&dmabuf->poll.lock);
> >>>>> -             if (dcb->active) {
> >>>>> -                     dcb->active |= pevents;
> >>>>> -                     events &= ~pevents;
> >>>>> -             } else
> >>>>> -                     dcb->active = pevents;
> >>>>> -             spin_unlock_irq(&dmabuf->poll.lock);
> >>>>> -
> >>>>> -             if (events & pevents) {
> >>>>> -                     if (!dma_fence_get_rcu(fence_excl)) {
> >>>>> -                             /* force a recheck */
> >>>>> -                             events &= ~pevents;
> >>>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
> >>>>> -                     } else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
> >>>>> -                                                        dma_buf_poll_cb)) {
> >>>>> -                             events &= ~pevents;
> >>>>> -                             dma_fence_put(fence_excl);
> >>>>> -                     } else {
> >>>>> -                             /*
> >>>>> -                              * No callback queued, wake up any additional
> >>>>> -                              * waiters.
> >>>>> -                              */
> >>>>> -                             dma_fence_put(fence_excl);
> >>>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
> >>>>> -                     }
> >>>>> +     for (i = 0; i < shared_count; ++i) {
> >>>>> +             fence = rcu_dereference(fobj->shared[i]);
> >>>>> +             fence = dma_fence_get_rcu(fence);
> >>>>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> >>>>> +                     /* Concurrent modify detected, force re-check */
> >>>>> +                     dma_fence_put(fence);
> >>>>> +                     rcu_read_unlock();
> >>>>> +                     goto retry;
> >>>>>                 }
> >>>>> -     }
> >>>>>
> >>>>> -     if ((events & EPOLLOUT) && shared_count > 0) {
> >>>>> -             struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
> >>>>> -             int i;
> >>>>> -
> >>>>> -             /* Only queue a new callback if no event has fired yet */
> >>>>> -             spin_lock_irq(&dmabuf->poll.lock);
> >>>>> -             if (dcb->active)
> >>>>> -                     events &= ~EPOLLOUT;
> >>>>> -             else
> >>>>> -                     dcb->active = EPOLLOUT;
> >>>>> -             spin_unlock_irq(&dmabuf->poll.lock);
> >>>>> -
> >>>>> -             if (!(events & EPOLLOUT))
> >>>>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> >>>>> +             if (!r) {
> >>>>> +                     /* Callback queued */
> >>>>> +                     events = 0;
> >>>>>                         goto out;
> >>>>> +             }
> >>>>> +             dma_fence_put(fence);
> >>>>> +     }
> >>>>>
> >>>>> -             for (i = 0; i < shared_count; ++i) {
> >>>>> -                     struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
> >>>>> -
> >>>>> -                     if (!dma_fence_get_rcu(fence)) {
> >>>>> -                             /*
> >>>>> -                              * fence refcount dropped to zero, this means
> >>>>> -                              * that fobj has been freed
> >>>>> -                              *
> >>>>> -                              * call dma_buf_poll_cb and force a recheck!
> >>>>> -                              */
> >>>>> -                             events &= ~EPOLLOUT;
> >>>>> -                             dma_buf_poll_cb(NULL, &dcb->cb);
> >>>>> -                             break;
> >>>>> -                     }
> >>>>> -                     if (!dma_fence_add_callback(fence, &dcb->cb,
> >>>>> -                                                 dma_buf_poll_cb)) {
> >>>>> -                             dma_fence_put(fence);
> >>>>> -                             events &= ~EPOLLOUT;
> >>>>> -                             break;
> >>>>> -                     }
> >>>>> +     fence = dma_resv_get_excl(resv);
> >>>>> +     if (fence && shared_count == 0) {
> >>>>> +             fence = dma_fence_get_rcu(fence);
> >>>>> +             if (!fence || read_seqcount_retry(&resv->seq, seq)) {
> >>>>> +                     /* Concurrent modify detected, force re-check */
> >>>>>                         dma_fence_put(fence);
> >>>>> +                     rcu_read_unlock();
> >>>>> +                     goto retry;
> >>>>> +
> >>>>>                 }
> >>>>>
> >>>>> -             /* No callback queued, wake up any additional waiters. */
> >>>>> -             if (i == shared_count)
> >>>>> -                     dma_buf_poll_cb(NULL, &dcb->cb);
> >>>>> +             r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
> >>>>> +             if (!r) {
> >>>>> +                     /* Callback queued */
> >>>>> +                     events = 0;
> >>>>> +                     goto out;
> >>>>> +             }
> >>>>> +             dma_fence_put(fence_excl);
> >>>>>         }
> >>>>>
> >>>>> +     /* No callback queued, wake up any additional waiters. */
> >>>>> +     dma_buf_poll_cb(NULL, &dcb->cb);
> >>>>> +
> >>>>>     out:
> >>>>>         rcu_read_unlock();
> >>>>>         return events;
> >>>>> @@ -562,8 +537,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info)
> >>>>>         dmabuf->owner = exp_info->owner;
> >>>>>         spin_lock_init(&dmabuf->name_lock);
> >>>>>         init_waitqueue_head(&dmabuf->poll);
> >>>>> -     dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
> >>>>> -     dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
> >>>>> +     dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
> >>>>> +     dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
> >>>>>
> >>>>>         if (!resv) {
> >>>>>                 resv = (struct dma_resv *)&dmabuf[1];
> >>>>> diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h
> >>>>> index efdc56b9d95f..7e747ad54c81 100644
> >>>>> --- a/include/linux/dma-buf.h
> >>>>> +++ b/include/linux/dma-buf.h
> >>>>> @@ -329,7 +329,7 @@ struct dma_buf {
> >>>>>                 wait_queue_head_t *poll;
> >>>>>
> >>>>>                 __poll_t active;
> >>>>> -     } cb_excl, cb_shared;
> >>>>> +     } cb_in, cb_out;
> >>>>>     };
> >>>>>
> >>>>>     /**
> >
>


-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch

^ permalink raw reply	[flat|nested] 10+ messages in thread

end of thread, other threads:[~2021-06-23 13:56 UTC | newest]

Thread overview: 10+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-06-22 13:04 [PATCH] dma-buf: fix and rework dma_buf_poll v3 Christian König
2021-06-22 13:07 ` Christian König
2021-06-22 17:02   ` Daniel Vetter
2021-06-23 11:17     ` Christian König
2021-06-23 11:30       ` Daniel Vetter
2021-06-23 12:42         ` Christian König
2021-06-23 13:56           ` Daniel Vetter
2021-06-23  1:23 kernel test robot
2021-06-23  9:28 ` Dan Carpenter
2021-06-23  9:28 ` Dan Carpenter

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.