From: "Christian König" <ckoenig.leichtzumerken@gmail.com>
To: linaro-mm-sig@lists.linaro.org, dri-devel@lists.freedesktop.org,
linux-media@vger.kernel.org, intel-gfx@lists.freedesktop.org
Cc: daniel@ffwll.ch, tvrtko.ursulin@linux.intel.com
Subject: [PATCH 05/27] dma-buf: use new iterator in dma_resv_get_fences v3
Date: Fri, 24 Sep 2021 17:30:51 +0200 [thread overview]
Message-ID: <20210924153113.2159-5-christian.koenig@amd.com> (raw)
In-Reply-To: <20210924153113.2159-1-christian.koenig@amd.com>
This makes the function much simpler since the complex
retry logic is now handled elsewhere.
v2: use sizeof(void*) instead
v3: fix rebase bug
Signed-off-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Daniel Vetter <daniel.vetter@ffwll.ch>
---
drivers/dma-buf/dma-resv.c | 108 ++++++++++++-------------------------
1 file changed, 35 insertions(+), 73 deletions(-)
diff --git a/drivers/dma-buf/dma-resv.c b/drivers/dma-buf/dma-resv.c
index 17ef7472830b..39a3eac28834 100644
--- a/drivers/dma-buf/dma-resv.c
+++ b/drivers/dma-buf/dma-resv.c
@@ -533,99 +533,61 @@ EXPORT_SYMBOL(dma_resv_copy_fences);
* dma_resv_get_fences - Get an object's shared and exclusive
* fences without update side lock held
* @obj: the reservation object
- * @pfence_excl: the returned exclusive fence (or NULL)
- * @pshared_count: the number of shared fences returned
- * @pshared: the array of shared fence ptrs returned (array is krealloc'd to
+ * @fence_excl: the returned exclusive fence (or NULL)
+ * @shared_count: the number of shared fences returned
+ * @shared: the array of shared fence ptrs returned (array is krealloc'd to
* the required size, and must be freed by caller)
*
* Retrieve all fences from the reservation object. If the pointer for the
* exclusive fence is not specified the fence is put into the array of the
* shared fences as well. Returns either zero or -ENOMEM.
*/
-int dma_resv_get_fences(struct dma_resv *obj, struct dma_fence **pfence_excl,
- unsigned int *pshared_count,
- struct dma_fence ***pshared)
+int dma_resv_get_fences(struct dma_resv *obj, struct dma_fence **fence_excl,
+ unsigned int *shared_count, struct dma_fence ***shared)
{
- struct dma_fence **shared = NULL;
- struct dma_fence *fence_excl;
- unsigned int shared_count;
- int ret = 1;
-
- do {
- struct dma_resv_list *fobj;
- unsigned int i, seq;
- size_t sz = 0;
-
- shared_count = i = 0;
-
- rcu_read_lock();
- seq = read_seqcount_begin(&obj->seq);
+ struct dma_resv_iter cursor;
+ struct dma_fence *fence;
- fence_excl = dma_resv_excl_fence(obj);
- if (fence_excl && !dma_fence_get_rcu(fence_excl))
- goto unlock;
+ *shared_count = 0;
+ *shared = NULL;
- fobj = dma_resv_shared_list(obj);
- if (fobj)
- sz += sizeof(*shared) * fobj->shared_max;
+ if (fence_excl)
+ *fence_excl = NULL;
- if (!pfence_excl && fence_excl)
- sz += sizeof(*shared);
+ dma_resv_iter_begin(&cursor, obj, true);
+ dma_resv_for_each_fence_unlocked(&cursor, fence) {
- if (sz) {
- struct dma_fence **nshared;
+ if (dma_resv_iter_is_restarted(&cursor)) {
+ unsigned int count;
- nshared = krealloc(shared, sz,
- GFP_NOWAIT | __GFP_NOWARN);
- if (!nshared) {
- rcu_read_unlock();
+ while (*shared_count)
+ dma_fence_put((*shared)[--(*shared_count)]);
- dma_fence_put(fence_excl);
- fence_excl = NULL;
+ if (fence_excl)
+ dma_fence_put(*fence_excl);
- nshared = krealloc(shared, sz, GFP_KERNEL);
- if (nshared) {
- shared = nshared;
- continue;
- }
+ count = cursor.fences ? cursor.fences->shared_count : 0;
+ count += fence_excl ? 0 : 1;
- ret = -ENOMEM;
- break;
+ /* Eventually re-allocate the array */
+ *shared = krealloc_array(*shared, count,
+ sizeof(void *),
+ GFP_KERNEL);
+ if (count && !*shared) {
+ dma_resv_iter_end(&cursor);
+ return -ENOMEM;
}
- shared = nshared;
- shared_count = fobj ? fobj->shared_count : 0;
- for (i = 0; i < shared_count; ++i) {
- shared[i] = rcu_dereference(fobj->shared[i]);
- if (!dma_fence_get_rcu(shared[i]))
- break;
- }
- }
-
- if (i != shared_count || read_seqcount_retry(&obj->seq, seq)) {
- while (i--)
- dma_fence_put(shared[i]);
- dma_fence_put(fence_excl);
- goto unlock;
}
- ret = 0;
-unlock:
- rcu_read_unlock();
- } while (ret);
-
- if (pfence_excl)
- *pfence_excl = fence_excl;
- else if (fence_excl)
- shared[shared_count++] = fence_excl;
-
- if (!shared_count) {
- kfree(shared);
- shared = NULL;
+ dma_fence_get(fence);
+ if (dma_resv_iter_is_exclusive(&cursor) && fence_excl)
+ *fence_excl = fence;
+ else
+ (*shared)[(*shared_count)++] = fence;
}
+ dma_resv_iter_end(&cursor);
- *pshared_count = shared_count;
- *pshared = shared;
- return ret;
+ return 0;
}
EXPORT_SYMBOL_GPL(dma_resv_get_fences);
--
2.25.1
next prev parent reply other threads:[~2021-09-24 15:31 UTC|newest]
Thread overview: 27+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-09-24 15:30 [PATCH 01/27] dma-buf: add dma_resv_for_each_fence_unlocked v6 Christian König
2021-09-24 15:30 ` [PATCH 02/27] dma-buf: add dma_resv_for_each_fence Christian König
2021-09-24 15:30 ` [PATCH 03/27] dma-buf: add dma_resv selftest Christian König
2021-09-24 15:30 ` [PATCH 04/27] dma-buf: use new iterator in dma_resv_copy_fences Christian König
2021-09-24 15:30 ` Christian König [this message]
2021-09-24 15:30 ` [PATCH 06/27] dma-buf: use new iterator in dma_resv_wait_timeout Christian König
2021-09-24 15:30 ` [PATCH 07/27] dma-buf: use new iterator in dma_resv_test_signaled Christian König
2021-09-24 15:30 ` [PATCH 08/27] dma-buf: use the new iterator in dma_buf_debug_show Christian König
2021-09-24 15:30 ` [PATCH 09/27] drm/ttm: use the new iterator in ttm_bo_flush_all_fences Christian König
2021-09-24 15:30 ` [PATCH 10/27] drm/amdgpu: use the new iterator in amdgpu_sync_resv Christian König
2021-09-24 15:30 ` [PATCH 11/27] drm/amdgpu: use new iterator in amdgpu_ttm_bo_eviction_valuable Christian König
2021-09-24 15:30 ` [PATCH 12/27] drm/amdgpu: use new iterator in amdgpu_vm_prt_fini Christian König
2021-09-24 15:30 ` [PATCH 13/27] drm/msm: use new iterator in msm_gem_describe Christian König
2021-09-24 15:31 ` [PATCH 14/27] drm/radeon: use new iterator in radeon_sync_resv Christian König
2021-09-24 15:31 ` [PATCH 15/27] drm/scheduler: use new iterator in drm_sched_job_add_implicit_dependencies v2 Christian König
2021-09-24 15:31 ` [PATCH 16/27] drm/i915: use the new iterator in i915_gem_busy_ioctl v2 Christian König
2021-09-24 15:31 ` [PATCH 17/27] drm/i915: use the new iterator in i915_sw_fence_await_reservation v3 Christian König
2021-09-24 15:31 ` [PATCH 18/27] drm/i915: use the new iterator in i915_request_await_object v2 Christian König
2021-09-24 15:31 ` [PATCH 19/27] drm/i915: use new iterator in i915_gem_object_wait_reservation Christian König
2021-09-24 15:31 ` [PATCH 20/27] drm/i915: use new iterator in i915_gem_object_wait_priority Christian König
2021-09-24 15:31 ` [PATCH 21/27] drm/i915: use new cursor in intel_prepare_plane_fb Christian König
2021-09-24 15:31 ` [PATCH 22/27] drm: use new iterator in drm_gem_fence_array_add_implicit v3 Christian König
2021-09-24 15:31 ` [PATCH 23/27] drm: use new iterator in drm_gem_plane_helper_prepare_fb Christian König
2021-09-24 15:31 ` [PATCH 24/27] drm/nouveau: use the new iterator in nouveau_fence_sync Christian König
2021-09-24 15:31 ` [PATCH 25/27] drm/nouveau: use the new interator in nv50_wndw_prepare_fb Christian König
2021-09-24 15:31 ` [PATCH 26/27] drm/etnaviv: use new iterator in etnaviv_gem_describe Christian König
2021-09-24 15:31 ` [PATCH 27/27] drm/etnaviv: replace dma_resv_get_excl_unlocked Christian König
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210924153113.2159-5-christian.koenig@amd.com \
--to=ckoenig.leichtzumerken@gmail.com \
--cc=daniel@ffwll.ch \
--cc=dri-devel@lists.freedesktop.org \
--cc=intel-gfx@lists.freedesktop.org \
--cc=linaro-mm-sig@lists.linaro.org \
--cc=linux-media@vger.kernel.org \
--cc=tvrtko.ursulin@linux.intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).