All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
@ 2021-04-27 10:54 Christian König
  2021-04-27 10:54 ` [PATCH 2/5] drm/amdgpu: use cursor functions in amdgpu_bo_in_cpu_visible_vram Christian König
                   ` (5 more replies)
  0 siblings, 6 replies; 12+ messages in thread
From: Christian König @ 2021-04-27 10:54 UTC (permalink / raw)
  To: amd-gfx

Now that we found the underlying problem we can re-apply this patch.

This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.

v2: rebase on KFD changes

Signed-off-by: Christian König <christian.koenig@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55 +++++++++-----------------
 drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
 drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
 3 files changed, 20 insertions(+), 40 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
index dae51992c607..fa43d332a979 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
@@ -37,6 +37,7 @@
 #include "amdgpu_gmc.h"
 #include "amdgpu_xgmi.h"
 #include "amdgpu_dma_buf.h"
+#include "amdgpu_res_cursor.h"
 #include "kfd_svm.h"
 
 /**
@@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct amdgpu_vm_update_params *params,
  * @last: last mapped entry
  * @flags: flags for the entries
  * @offset: offset into nodes and pages_addr
- * @nodes: array of drm_mm_nodes with the MC addresses
+ * @res: ttm_resource to map
  * @pages_addr: DMA addresses to use for mapping
  * @fence: optional resulting fence
  *
@@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 				bool unlocked, struct dma_resv *resv,
 				uint64_t start, uint64_t last,
 				uint64_t flags, uint64_t offset,
-				struct drm_mm_node *nodes,
+				struct ttm_resource *res,
 				dma_addr_t *pages_addr,
 				struct dma_fence **fence)
 {
 	struct amdgpu_vm_update_params params;
+	struct amdgpu_res_cursor cursor;
 	enum amdgpu_sync_mode sync_mode;
-	uint64_t pfn;
 	int r;
 
 	memset(&params, 0, sizeof(params));
@@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 	else
 		sync_mode = AMDGPU_SYNC_EXPLICIT;
 
-	pfn = offset >> PAGE_SHIFT;
-	if (nodes) {
-		while (pfn >= nodes->size) {
-			pfn -= nodes->size;
-			++nodes;
-		}
-	}
-
 	amdgpu_vm_eviction_lock(vm);
 	if (vm->evicting) {
 		r = -EBUSY;
@@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 	if (r)
 		goto error_unlock;
 
-	do {
+	amdgpu_res_first(res, offset, (last - start + 1) * AMDGPU_GPU_PAGE_SIZE,
+			 &cursor);
+	while (cursor.remaining) {
 		uint64_t tmp, num_entries, addr;
 
-
-		num_entries = last - start + 1;
-		if (nodes) {
-			addr = nodes->start << PAGE_SHIFT;
-			num_entries = min((nodes->size - pfn) *
-				AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
-		} else {
-			addr = 0;
-		}
-
+		num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
 		if (pages_addr) {
 			bool contiguous = true;
 
 			if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
+				uint64_t pfn = cursor.start >> PAGE_SHIFT;
 				uint64_t count;
 
 				contiguous = pages_addr[pfn + 1] ==
@@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 			}
 
 			if (!contiguous) {
-				addr = pfn << PAGE_SHIFT;
+				addr = cursor.start;
 				params.pages_addr = pages_addr;
 			} else {
-				addr = pages_addr[pfn];
+				addr = pages_addr[cursor.start >> PAGE_SHIFT];
 				params.pages_addr = NULL;
 			}
 
 		} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
-			addr += bo_adev->vm_manager.vram_base_offset;
-			addr += pfn << PAGE_SHIFT;
+			addr = bo_adev->vm_manager.vram_base_offset +
+				cursor.start;
+		} else {
+			addr = 0;
 		}
 
 		tmp = start + num_entries;
@@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 		if (r)
 			goto error_unlock;
 
-		pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
-		if (nodes && nodes->size == pfn) {
-			pfn = 0;
-			++nodes;
-		}
+		amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
 		start = tmp;
-
-	} while (unlikely(start != last + 1));
+	};
 
 	r = vm->update_funcs->commit(&params, fence);
 
@@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
 	struct amdgpu_bo_va_mapping *mapping;
 	dma_addr_t *pages_addr = NULL;
 	struct ttm_resource *mem;
-	struct drm_mm_node *nodes;
 	struct dma_fence **last_update;
 	struct dma_resv *resv;
 	uint64_t flags;
@@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
 
 	if (clear || !bo) {
 		mem = NULL;
-		nodes = NULL;
 		resv = vm->root.base.bo->tbo.base.resv;
 	} else {
 		struct drm_gem_object *obj = &bo->tbo.base;
@@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
 				bo = gem_to_amdgpu_bo(gobj);
 		}
 		mem = &bo->tbo.mem;
-		nodes = mem->mm_node;
 		if (mem->mem_type == TTM_PL_TT)
 			pages_addr = bo->tbo.ttm->dma_address;
 	}
@@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
 		r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false, false,
 						resv, mapping->start,
 						mapping->last, update_flags,
-						mapping->offset, nodes,
+						mapping->offset, mem,
 						pages_addr, last_update);
 		if (r)
 			return r;
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
index e5a3f18be2b7..1ae5ea8db497 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
@@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
 				bool unlocked, struct dma_resv *resv,
 				uint64_t start, uint64_t last,
 				uint64_t flags, uint64_t offset,
-				struct drm_mm_node *nodes,
+				struct ttm_resource *res,
 				dma_addr_t *pages_addr,
 				struct dma_fence **fence);
 int amdgpu_vm_bo_update(struct amdgpu_device *adev,
diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
index e4ce97ab6e26..0b0e76e16ddc 100644
--- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
+++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
@@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device *adev, struct amdgpu_vm *vm,
 					prange->mapping.start,
 					prange->mapping.last, pte_flags,
 					prange->mapping.offset,
-					prange->ttm_res ?
-						prange->ttm_res->mm_node : NULL,
+					prange->ttm_res,
 					dma_addr, &vm->last_update);
 	if (r) {
 		pr_debug("failed %d to map to gpu 0x%lx\n", r, prange->start);
-- 
2.25.1

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply related	[flat|nested] 12+ messages in thread

* [PATCH 2/5] drm/amdgpu: use cursor functions in amdgpu_bo_in_cpu_visible_vram
  2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
@ 2021-04-27 10:54 ` Christian König
  2021-04-27 10:54 ` [PATCH 3/5] drm/amdgpu: set the contiguous flag if possible Christian König
                   ` (4 subsequent siblings)
  5 siblings, 0 replies; 12+ messages in thread
From: Christian König @ 2021-04-27 10:54 UTC (permalink / raw)
  To: amd-gfx

One of the last remaining uses of drm_mm_node.

Signed-off-by: Christian König <christian.koenig@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_object.h | 15 +++++++++------
 1 file changed, 9 insertions(+), 6 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
index e0ec48d6a3fd..1cebe24cca71 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_object.h
@@ -30,6 +30,8 @@
 
 #include <drm/amdgpu_drm.h>
 #include "amdgpu.h"
+#include "amdgpu_res_cursor.h"
+
 #ifdef CONFIG_MMU_NOTIFIER
 #include <linux/mmu_notifier.h>
 #endif
@@ -215,18 +217,19 @@ static inline u64 amdgpu_bo_mmap_offset(struct amdgpu_bo *bo)
 static inline bool amdgpu_bo_in_cpu_visible_vram(struct amdgpu_bo *bo)
 {
 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);
-	unsigned fpfn = adev->gmc.visible_vram_size >> PAGE_SHIFT;
-	struct drm_mm_node *node = bo->tbo.mem.mm_node;
-	unsigned long pages_left;
+	struct amdgpu_res_cursor cursor;
 
 	if (bo->tbo.mem.mem_type != TTM_PL_VRAM)
 		return false;
 
-	for (pages_left = bo->tbo.mem.num_pages; pages_left;
-	     pages_left -= node->size, node++)
-		if (node->start < fpfn)
+	amdgpu_res_first(&bo->tbo.mem, 0, amdgpu_bo_size(bo), &cursor);
+	while (cursor.remaining) {
+		if (cursor.start < adev->gmc.visible_vram_size)
 			return true;
 
+		amdgpu_res_next(&cursor, cursor.size);
+	}
+
 	return false;
 }
 
-- 
2.25.1

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply related	[flat|nested] 12+ messages in thread

* [PATCH 3/5] drm/amdgpu: set the contiguous flag if possible
  2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
  2021-04-27 10:54 ` [PATCH 2/5] drm/amdgpu: use cursor functions in amdgpu_bo_in_cpu_visible_vram Christian König
@ 2021-04-27 10:54 ` Christian König
  2021-04-27 10:54 ` [PATCH 4/5] drm/amdgpu: check contiguous flags instead of mm_node Christian König
                   ` (3 subsequent siblings)
  5 siblings, 0 replies; 12+ messages in thread
From: Christian König @ 2021-04-27 10:54 UTC (permalink / raw)
  To: amd-gfx

This avoids reallocating scanout BOs on first pin in a lot of cases.

Signed-off-by: Christian König <christian.koenig@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c | 5 +++--
 1 file changed, 3 insertions(+), 2 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
index e2cbe19404c0..87c40c4cdd50 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
@@ -446,10 +446,11 @@ static int amdgpu_vram_mgr_new(struct ttm_resource_manager *man,
 	}
 	spin_unlock(&mgr->lock);
 
-	atomic64_add(vis_usage, &mgr->vis_usage);
+	if (i == 1)
+		mem->placement |= TTM_PL_FLAG_CONTIGUOUS;
 
+	atomic64_add(vis_usage, &mgr->vis_usage);
 	mem->mm_node = nodes;
-
 	return 0;
 
 error:
-- 
2.25.1

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply related	[flat|nested] 12+ messages in thread

* [PATCH 4/5] drm/amdgpu: check contiguous flags instead of mm_node
  2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
  2021-04-27 10:54 ` [PATCH 2/5] drm/amdgpu: use cursor functions in amdgpu_bo_in_cpu_visible_vram Christian König
  2021-04-27 10:54 ` [PATCH 3/5] drm/amdgpu: set the contiguous flag if possible Christian König
@ 2021-04-27 10:54 ` Christian König
  2021-04-27 10:54 ` [PATCH 5/5] drm/amdgpu: move struct amdgpu_vram_reservation into vram mgr Christian König
                   ` (2 subsequent siblings)
  5 siblings, 0 replies; 12+ messages in thread
From: Christian König @ 2021-04-27 10:54 UTC (permalink / raw)
  To: amd-gfx

Drop the last user of drm_mm_node.

Signed-off-by: Christian König <christian.koenig@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 11 ++++-------
 1 file changed, 4 insertions(+), 7 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
index 7e7d8330d64b..b068253575eb 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
@@ -576,10 +576,10 @@ static int amdgpu_bo_move(struct ttm_buffer_object *bo, bool evict,
  *
  * Called by ttm_mem_io_reserve() ultimately via ttm_bo_vm_fault()
  */
-static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resource *mem)
+static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev,
+				     struct ttm_resource *mem)
 {
 	struct amdgpu_device *adev = amdgpu_ttm_adev(bdev);
-	struct drm_mm_node *mm_node = mem->mm_node;
 	size_t bus_size = (size_t)mem->num_pages << PAGE_SHIFT;
 
 	switch (mem->mem_type) {
@@ -593,12 +593,9 @@ static int amdgpu_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resourc
 		/* check if it's visible */
 		if ((mem->bus.offset + bus_size) > adev->gmc.visible_vram_size)
 			return -EINVAL;
-		/* Only physically contiguous buffers apply. In a contiguous
-		 * buffer, size of the first mm_node would match the number of
-		 * pages in ttm_resource.
-		 */
+
 		if (adev->mman.aper_base_kaddr &&
-		    (mm_node->size == mem->num_pages))
+		    mem->placement & TTM_PL_FLAG_CONTIGUOUS)
 			mem->bus.addr = (u8 *)adev->mman.aper_base_kaddr +
 					mem->bus.offset;
 
-- 
2.25.1

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply related	[flat|nested] 12+ messages in thread

* [PATCH 5/5] drm/amdgpu: move struct amdgpu_vram_reservation into vram mgr
  2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
                   ` (2 preceding siblings ...)
  2021-04-27 10:54 ` [PATCH 4/5] drm/amdgpu: check contiguous flags instead of mm_node Christian König
@ 2021-04-27 10:54 ` Christian König
  2021-04-30  8:32 ` [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Felix Kuehling
  2021-04-30 15:19 ` Nirmoy
  5 siblings, 0 replies; 12+ messages in thread
From: Christian König @ 2021-04-27 10:54 UTC (permalink / raw)
  To: amd-gfx

Not used outside of that file.

Signed-off-by: Christian König <christian.koenig@amd.com>
---
 drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h      | 5 -----
 drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c | 5 +++++
 2 files changed, 5 insertions(+), 5 deletions(-)

diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
index 9e38475e0f8d..eb84a69c4b74 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.h
@@ -37,11 +37,6 @@
 
 #define AMDGPU_POISON	0xd0bed0be
 
-struct amdgpu_vram_reservation {
-	struct list_head node;
-	struct drm_mm_node mm_node;
-};
-
 struct amdgpu_vram_mgr {
 	struct ttm_resource_manager manager;
 	struct drm_mm mm;
diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
index 87c40c4cdd50..a52e17ed3df6 100644
--- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
+++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
@@ -29,6 +29,11 @@
 #include "amdgpu_atomfirmware.h"
 #include "atom.h"
 
+struct amdgpu_vram_reservation {
+	struct list_head node;
+	struct drm_mm_node mm_node;
+};
+
 static inline struct amdgpu_vram_mgr *
 to_vram_mgr(struct ttm_resource_manager *man)
 {
-- 
2.25.1

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply related	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
                   ` (3 preceding siblings ...)
  2021-04-27 10:54 ` [PATCH 5/5] drm/amdgpu: move struct amdgpu_vram_reservation into vram mgr Christian König
@ 2021-04-30  8:32 ` Felix Kuehling
  2021-04-30  8:54   ` Christian König
  2021-04-30 15:19 ` Nirmoy
  5 siblings, 1 reply; 12+ messages in thread
From: Felix Kuehling @ 2021-04-30  8:32 UTC (permalink / raw)
  To: Christian König, amd-gfx

Am 2021-04-27 um 6:54 a.m. schrieb Christian König:
> Now that we found the underlying problem we can re-apply this patch.
>
> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>
> v2: rebase on KFD changes
>
> Signed-off-by: Christian König <christian.koenig@amd.com>

Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>


> ---
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55 +++++++++-----------------
>  drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>  drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>  3 files changed, 20 insertions(+), 40 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index dae51992c607..fa43d332a979 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -37,6 +37,7 @@
>  #include "amdgpu_gmc.h"
>  #include "amdgpu_xgmi.h"
>  #include "amdgpu_dma_buf.h"
> +#include "amdgpu_res_cursor.h"
>  #include "kfd_svm.h"
>  
>  /**
> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct amdgpu_vm_update_params *params,
>   * @last: last mapped entry
>   * @flags: flags for the entries
>   * @offset: offset into nodes and pages_addr
> - * @nodes: array of drm_mm_nodes with the MC addresses
> + * @res: ttm_resource to map
>   * @pages_addr: DMA addresses to use for mapping
>   * @fence: optional resulting fence
>   *
> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  				bool unlocked, struct dma_resv *resv,
>  				uint64_t start, uint64_t last,
>  				uint64_t flags, uint64_t offset,
> -				struct drm_mm_node *nodes,
> +				struct ttm_resource *res,
>  				dma_addr_t *pages_addr,
>  				struct dma_fence **fence)
>  {
>  	struct amdgpu_vm_update_params params;
> +	struct amdgpu_res_cursor cursor;
>  	enum amdgpu_sync_mode sync_mode;
> -	uint64_t pfn;
>  	int r;
>  
>  	memset(&params, 0, sizeof(params));
> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  	else
>  		sync_mode = AMDGPU_SYNC_EXPLICIT;
>  
> -	pfn = offset >> PAGE_SHIFT;
> -	if (nodes) {
> -		while (pfn >= nodes->size) {
> -			pfn -= nodes->size;
> -			++nodes;
> -		}
> -	}
> -
>  	amdgpu_vm_eviction_lock(vm);
>  	if (vm->evicting) {
>  		r = -EBUSY;
> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  	if (r)
>  		goto error_unlock;
>  
> -	do {
> +	amdgpu_res_first(res, offset, (last - start + 1) * AMDGPU_GPU_PAGE_SIZE,
> +			 &cursor);
> +	while (cursor.remaining) {
>  		uint64_t tmp, num_entries, addr;
>  
> -
> -		num_entries = last - start + 1;
> -		if (nodes) {
> -			addr = nodes->start << PAGE_SHIFT;
> -			num_entries = min((nodes->size - pfn) *
> -				AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
> -		} else {
> -			addr = 0;
> -		}
> -
> +		num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>  		if (pages_addr) {
>  			bool contiguous = true;
>  
>  			if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
> +				uint64_t pfn = cursor.start >> PAGE_SHIFT;
>  				uint64_t count;
>  
>  				contiguous = pages_addr[pfn + 1] ==
> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  			}
>  
>  			if (!contiguous) {
> -				addr = pfn << PAGE_SHIFT;
> +				addr = cursor.start;
>  				params.pages_addr = pages_addr;
>  			} else {
> -				addr = pages_addr[pfn];
> +				addr = pages_addr[cursor.start >> PAGE_SHIFT];
>  				params.pages_addr = NULL;
>  			}
>  
>  		} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
> -			addr += bo_adev->vm_manager.vram_base_offset;
> -			addr += pfn << PAGE_SHIFT;
> +			addr = bo_adev->vm_manager.vram_base_offset +
> +				cursor.start;
> +		} else {
> +			addr = 0;
>  		}
>  
>  		tmp = start + num_entries;
> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  		if (r)
>  			goto error_unlock;
>  
> -		pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
> -		if (nodes && nodes->size == pfn) {
> -			pfn = 0;
> -			++nodes;
> -		}
> +		amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
>  		start = tmp;
> -
> -	} while (unlikely(start != last + 1));
> +	};
>  
>  	r = vm->update_funcs->commit(&params, fence);
>  
> @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>  	struct amdgpu_bo_va_mapping *mapping;
>  	dma_addr_t *pages_addr = NULL;
>  	struct ttm_resource *mem;
> -	struct drm_mm_node *nodes;
>  	struct dma_fence **last_update;
>  	struct dma_resv *resv;
>  	uint64_t flags;
> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>  
>  	if (clear || !bo) {
>  		mem = NULL;
> -		nodes = NULL;
>  		resv = vm->root.base.bo->tbo.base.resv;
>  	} else {
>  		struct drm_gem_object *obj = &bo->tbo.base;
> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>  				bo = gem_to_amdgpu_bo(gobj);
>  		}
>  		mem = &bo->tbo.mem;
> -		nodes = mem->mm_node;
>  		if (mem->mem_type == TTM_PL_TT)
>  			pages_addr = bo->tbo.ttm->dma_address;
>  	}
> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>  		r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false, false,
>  						resv, mapping->start,
>  						mapping->last, update_flags,
> -						mapping->offset, nodes,
> +						mapping->offset, mem,
>  						pages_addr, last_update);
>  		if (r)
>  			return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> index e5a3f18be2b7..1ae5ea8db497 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>  				bool unlocked, struct dma_resv *resv,
>  				uint64_t start, uint64_t last,
>  				uint64_t flags, uint64_t offset,
> -				struct drm_mm_node *nodes,
> +				struct ttm_resource *res,
>  				dma_addr_t *pages_addr,
>  				struct dma_fence **fence);
>  int amdgpu_vm_bo_update(struct amdgpu_device *adev,
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> index e4ce97ab6e26..0b0e76e16ddc 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device *adev, struct amdgpu_vm *vm,
>  					prange->mapping.start,
>  					prange->mapping.last, pte_flags,
>  					prange->mapping.offset,
> -					prange->ttm_res ?
> -						prange->ttm_res->mm_node : NULL,
> +					prange->ttm_res,
>  					dma_addr, &vm->last_update);
>  	if (r) {
>  		pr_debug("failed %d to map to gpu 0x%lx\n", r, prange->start);
_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-30  8:32 ` [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Felix Kuehling
@ 2021-04-30  8:54   ` Christian König
  2021-04-30 13:45     ` Felix Kuehling
  0 siblings, 1 reply; 12+ messages in thread
From: Christian König @ 2021-04-30  8:54 UTC (permalink / raw)
  To: Felix Kuehling, amd-gfx



Am 30.04.21 um 10:32 schrieb Felix Kuehling:
> Am 2021-04-27 um 6:54 a.m. schrieb Christian König:
>> Now that we found the underlying problem we can re-apply this patch.
>>
>> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>>
>> v2: rebase on KFD changes
>>
>> Signed-off-by: Christian König <christian.koenig@amd.com>
> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>

Thanks, just for this patch or the series?

Christian.

>
>
>> ---
>>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55 +++++++++-----------------
>>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>>   drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>>   3 files changed, 20 insertions(+), 40 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>> index dae51992c607..fa43d332a979 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>> @@ -37,6 +37,7 @@
>>   #include "amdgpu_gmc.h"
>>   #include "amdgpu_xgmi.h"
>>   #include "amdgpu_dma_buf.h"
>> +#include "amdgpu_res_cursor.h"
>>   #include "kfd_svm.h"
>>   
>>   /**
>> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct amdgpu_vm_update_params *params,
>>    * @last: last mapped entry
>>    * @flags: flags for the entries
>>    * @offset: offset into nodes and pages_addr
>> - * @nodes: array of drm_mm_nodes with the MC addresses
>> + * @res: ttm_resource to map
>>    * @pages_addr: DMA addresses to use for mapping
>>    * @fence: optional resulting fence
>>    *
>> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>>   				bool unlocked, struct dma_resv *resv,
>>   				uint64_t start, uint64_t last,
>>   				uint64_t flags, uint64_t offset,
>> -				struct drm_mm_node *nodes,
>> +				struct ttm_resource *res,
>>   				dma_addr_t *pages_addr,
>>   				struct dma_fence **fence)
>>   {
>>   	struct amdgpu_vm_update_params params;
>> +	struct amdgpu_res_cursor cursor;
>>   	enum amdgpu_sync_mode sync_mode;
>> -	uint64_t pfn;
>>   	int r;
>>   
>>   	memset(&params, 0, sizeof(params));
>> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>>   	else
>>   		sync_mode = AMDGPU_SYNC_EXPLICIT;
>>   
>> -	pfn = offset >> PAGE_SHIFT;
>> -	if (nodes) {
>> -		while (pfn >= nodes->size) {
>> -			pfn -= nodes->size;
>> -			++nodes;
>> -		}
>> -	}
>> -
>>   	amdgpu_vm_eviction_lock(vm);
>>   	if (vm->evicting) {
>>   		r = -EBUSY;
>> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>>   	if (r)
>>   		goto error_unlock;
>>   
>> -	do {
>> +	amdgpu_res_first(res, offset, (last - start + 1) * AMDGPU_GPU_PAGE_SIZE,
>> +			 &cursor);
>> +	while (cursor.remaining) {
>>   		uint64_t tmp, num_entries, addr;
>>   
>> -
>> -		num_entries = last - start + 1;
>> -		if (nodes) {
>> -			addr = nodes->start << PAGE_SHIFT;
>> -			num_entries = min((nodes->size - pfn) *
>> -				AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
>> -		} else {
>> -			addr = 0;
>> -		}
>> -
>> +		num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>>   		if (pages_addr) {
>>   			bool contiguous = true;
>>   
>>   			if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
>> +				uint64_t pfn = cursor.start >> PAGE_SHIFT;
>>   				uint64_t count;
>>   
>>   				contiguous = pages_addr[pfn + 1] ==
>> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>>   			}
>>   
>>   			if (!contiguous) {
>> -				addr = pfn << PAGE_SHIFT;
>> +				addr = cursor.start;
>>   				params.pages_addr = pages_addr;
>>   			} else {
>> -				addr = pages_addr[pfn];
>> +				addr = pages_addr[cursor.start >> PAGE_SHIFT];
>>   				params.pages_addr = NULL;
>>   			}
>>   
>>   		} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
>> -			addr += bo_adev->vm_manager.vram_base_offset;
>> -			addr += pfn << PAGE_SHIFT;
>> +			addr = bo_adev->vm_manager.vram_base_offset +
>> +				cursor.start;
>> +		} else {
>> +			addr = 0;
>>   		}
>>   
>>   		tmp = start + num_entries;
>> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>>   		if (r)
>>   			goto error_unlock;
>>   
>> -		pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
>> -		if (nodes && nodes->size == pfn) {
>> -			pfn = 0;
>> -			++nodes;
>> -		}
>> +		amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
>>   		start = tmp;
>> -
>> -	} while (unlikely(start != last + 1));
>> +	};
>>   
>>   	r = vm->update_funcs->commit(&params, fence);
>>   
>> @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>>   	struct amdgpu_bo_va_mapping *mapping;
>>   	dma_addr_t *pages_addr = NULL;
>>   	struct ttm_resource *mem;
>> -	struct drm_mm_node *nodes;
>>   	struct dma_fence **last_update;
>>   	struct dma_resv *resv;
>>   	uint64_t flags;
>> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>>   
>>   	if (clear || !bo) {
>>   		mem = NULL;
>> -		nodes = NULL;
>>   		resv = vm->root.base.bo->tbo.base.resv;
>>   	} else {
>>   		struct drm_gem_object *obj = &bo->tbo.base;
>> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>>   				bo = gem_to_amdgpu_bo(gobj);
>>   		}
>>   		mem = &bo->tbo.mem;
>> -		nodes = mem->mm_node;
>>   		if (mem->mem_type == TTM_PL_TT)
>>   			pages_addr = bo->tbo.ttm->dma_address;
>>   	}
>> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>>   		r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false, false,
>>   						resv, mapping->start,
>>   						mapping->last, update_flags,
>> -						mapping->offset, nodes,
>> +						mapping->offset, mem,
>>   						pages_addr, last_update);
>>   		if (r)
>>   			return r;
>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>> index e5a3f18be2b7..1ae5ea8db497 100644
>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>>   				bool unlocked, struct dma_resv *resv,
>>   				uint64_t start, uint64_t last,
>>   				uint64_t flags, uint64_t offset,
>> -				struct drm_mm_node *nodes,
>> +				struct ttm_resource *res,
>>   				dma_addr_t *pages_addr,
>>   				struct dma_fence **fence);
>>   int amdgpu_vm_bo_update(struct amdgpu_device *adev,
>> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>> index e4ce97ab6e26..0b0e76e16ddc 100644
>> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device *adev, struct amdgpu_vm *vm,
>>   					prange->mapping.start,
>>   					prange->mapping.last, pte_flags,
>>   					prange->mapping.offset,
>> -					prange->ttm_res ?
>> -						prange->ttm_res->mm_node : NULL,
>> +					prange->ttm_res,
>>   					dma_addr, &vm->last_update);
>>   	if (r) {
>>   		pr_debug("failed %d to map to gpu 0x%lx\n", r, prange->start);

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-30  8:54   ` Christian König
@ 2021-04-30 13:45     ` Felix Kuehling
  2021-04-30 13:50       ` Christian König
  0 siblings, 1 reply; 12+ messages in thread
From: Felix Kuehling @ 2021-04-30 13:45 UTC (permalink / raw)
  To: Christian König, amd-gfx

Am 2021-04-30 um 4:54 a.m. schrieb Christian König:
>
>
> Am 30.04.21 um 10:32 schrieb Felix Kuehling:
>> Am 2021-04-27 um 6:54 a.m. schrieb Christian König:
>>> Now that we found the underlying problem we can re-apply this patch.
>>>
>>> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>>>
>>> v2: rebase on KFD changes
>>>
>>> Signed-off-by: Christian König <christian.koenig@amd.com>
>> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>
>
> Thanks, just for this patch or the series?

I meant it for the patch. I could probably extend it to the series. But
I'm a bit confused how the patches relate to each other. There seems to
be an implication that everything leads up to enabling patch 5, i.e.
removing users of mm_nodes so that you can in patch 5 make the data
structure containing mm_nodes private.

But it's not the same data structure. I don't think patch 5 really
depends on any of the previous patches, because all the other patches
deal with ttm_resource.mm_nodes and patch 5 deals with
amdgpu_vram_reservation.mm_nodes.

Regards,
  Felix


>
> Christian.
>
>>
>>
>>> ---
>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55
>>> +++++++++-----------------
>>>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>>>   drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>>>   3 files changed, 20 insertions(+), 40 deletions(-)
>>>
>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>> index dae51992c607..fa43d332a979 100644
>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>> @@ -37,6 +37,7 @@
>>>   #include "amdgpu_gmc.h"
>>>   #include "amdgpu_xgmi.h"
>>>   #include "amdgpu_dma_buf.h"
>>> +#include "amdgpu_res_cursor.h"
>>>   #include "kfd_svm.h"
>>>     /**
>>> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct
>>> amdgpu_vm_update_params *params,
>>>    * @last: last mapped entry
>>>    * @flags: flags for the entries
>>>    * @offset: offset into nodes and pages_addr
>>> - * @nodes: array of drm_mm_nodes with the MC addresses
>>> + * @res: ttm_resource to map
>>>    * @pages_addr: DMA addresses to use for mapping
>>>    * @fence: optional resulting fence
>>>    *
>>> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct
>>> amdgpu_device *adev,
>>>                   bool unlocked, struct dma_resv *resv,
>>>                   uint64_t start, uint64_t last,
>>>                   uint64_t flags, uint64_t offset,
>>> -                struct drm_mm_node *nodes,
>>> +                struct ttm_resource *res,
>>>                   dma_addr_t *pages_addr,
>>>                   struct dma_fence **fence)
>>>   {
>>>       struct amdgpu_vm_update_params params;
>>> +    struct amdgpu_res_cursor cursor;
>>>       enum amdgpu_sync_mode sync_mode;
>>> -    uint64_t pfn;
>>>       int r;
>>>         memset(&params, 0, sizeof(params));
>>> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct
>>> amdgpu_device *adev,
>>>       else
>>>           sync_mode = AMDGPU_SYNC_EXPLICIT;
>>>   -    pfn = offset >> PAGE_SHIFT;
>>> -    if (nodes) {
>>> -        while (pfn >= nodes->size) {
>>> -            pfn -= nodes->size;
>>> -            ++nodes;
>>> -        }
>>> -    }
>>> -
>>>       amdgpu_vm_eviction_lock(vm);
>>>       if (vm->evicting) {
>>>           r = -EBUSY;
>>> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct
>>> amdgpu_device *adev,
>>>       if (r)
>>>           goto error_unlock;
>>>   -    do {
>>> +    amdgpu_res_first(res, offset, (last - start + 1) *
>>> AMDGPU_GPU_PAGE_SIZE,
>>> +             &cursor);
>>> +    while (cursor.remaining) {
>>>           uint64_t tmp, num_entries, addr;
>>>   -
>>> -        num_entries = last - start + 1;
>>> -        if (nodes) {
>>> -            addr = nodes->start << PAGE_SHIFT;
>>> -            num_entries = min((nodes->size - pfn) *
>>> -                AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
>>> -        } else {
>>> -            addr = 0;
>>> -        }
>>> -
>>> +        num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>>>           if (pages_addr) {
>>>               bool contiguous = true;
>>>                 if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
>>> +                uint64_t pfn = cursor.start >> PAGE_SHIFT;
>>>                   uint64_t count;
>>>                     contiguous = pages_addr[pfn + 1] ==
>>> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct
>>> amdgpu_device *adev,
>>>               }
>>>                 if (!contiguous) {
>>> -                addr = pfn << PAGE_SHIFT;
>>> +                addr = cursor.start;
>>>                   params.pages_addr = pages_addr;
>>>               } else {
>>> -                addr = pages_addr[pfn];
>>> +                addr = pages_addr[cursor.start >> PAGE_SHIFT];
>>>                   params.pages_addr = NULL;
>>>               }
>>>             } else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
>>> -            addr += bo_adev->vm_manager.vram_base_offset;
>>> -            addr += pfn << PAGE_SHIFT;
>>> +            addr = bo_adev->vm_manager.vram_base_offset +
>>> +                cursor.start;
>>> +        } else {
>>> +            addr = 0;
>>>           }
>>>             tmp = start + num_entries;
>>> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct
>>> amdgpu_device *adev,
>>>           if (r)
>>>               goto error_unlock;
>>>   -        pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
>>> -        if (nodes && nodes->size == pfn) {
>>> -            pfn = 0;
>>> -            ++nodes;
>>> -        }
>>> +        amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
>>>           start = tmp;
>>> -
>>> -    } while (unlikely(start != last + 1));
>>> +    };
>>>         r = vm->update_funcs->commit(&params, fence);
>>>   @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>> *adev, struct amdgpu_bo_va *bo_va,
>>>       struct amdgpu_bo_va_mapping *mapping;
>>>       dma_addr_t *pages_addr = NULL;
>>>       struct ttm_resource *mem;
>>> -    struct drm_mm_node *nodes;
>>>       struct dma_fence **last_update;
>>>       struct dma_resv *resv;
>>>       uint64_t flags;
>>> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>> *adev, struct amdgpu_bo_va *bo_va,
>>>         if (clear || !bo) {
>>>           mem = NULL;
>>> -        nodes = NULL;
>>>           resv = vm->root.base.bo->tbo.base.resv;
>>>       } else {
>>>           struct drm_gem_object *obj = &bo->tbo.base;
>>> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>> *adev, struct amdgpu_bo_va *bo_va,
>>>                   bo = gem_to_amdgpu_bo(gobj);
>>>           }
>>>           mem = &bo->tbo.mem;
>>> -        nodes = mem->mm_node;
>>>           if (mem->mem_type == TTM_PL_TT)
>>>               pages_addr = bo->tbo.ttm->dma_address;
>>>       }
>>> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>> *adev, struct amdgpu_bo_va *bo_va,
>>>           r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false,
>>> false,
>>>                           resv, mapping->start,
>>>                           mapping->last, update_flags,
>>> -                        mapping->offset, nodes,
>>> +                        mapping->offset, mem,
>>>                           pages_addr, last_update);
>>>           if (r)
>>>               return r;
>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>> index e5a3f18be2b7..1ae5ea8db497 100644
>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct
>>> amdgpu_device *adev,
>>>                   bool unlocked, struct dma_resv *resv,
>>>                   uint64_t start, uint64_t last,
>>>                   uint64_t flags, uint64_t offset,
>>> -                struct drm_mm_node *nodes,
>>> +                struct ttm_resource *res,
>>>                   dma_addr_t *pages_addr,
>>>                   struct dma_fence **fence);
>>>   int amdgpu_vm_bo_update(struct amdgpu_device *adev,
>>> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>> b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>> index e4ce97ab6e26..0b0e76e16ddc 100644
>>> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device
>>> *adev, struct amdgpu_vm *vm,
>>>                       prange->mapping.start,
>>>                       prange->mapping.last, pte_flags,
>>>                       prange->mapping.offset,
>>> -                    prange->ttm_res ?
>>> -                        prange->ttm_res->mm_node : NULL,
>>> +                    prange->ttm_res,
>>>                       dma_addr, &vm->last_update);
>>>       if (r) {
>>>           pr_debug("failed %d to map to gpu 0x%lx\n", r,
>>> prange->start);
>
_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-30 13:45     ` Felix Kuehling
@ 2021-04-30 13:50       ` Christian König
  2021-04-30 14:57         ` Felix Kuehling
  0 siblings, 1 reply; 12+ messages in thread
From: Christian König @ 2021-04-30 13:50 UTC (permalink / raw)
  To: Felix Kuehling, amd-gfx



Am 30.04.21 um 15:45 schrieb Felix Kuehling:
> Am 2021-04-30 um 4:54 a.m. schrieb Christian König:
>>
>> Am 30.04.21 um 10:32 schrieb Felix Kuehling:
>>> Am 2021-04-27 um 6:54 a.m. schrieb Christian König:
>>>> Now that we found the underlying problem we can re-apply this patch.
>>>>
>>>> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>>>>
>>>> v2: rebase on KFD changes
>>>>
>>>> Signed-off-by: Christian König <christian.koenig@amd.com>
>>> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>
>> Thanks, just for this patch or the series?
> I meant it for the patch. I could probably extend it to the series. But
> I'm a bit confused how the patches relate to each other. There seems to
> be an implication that everything leads up to enabling patch 5, i.e.
> removing users of mm_nodes so that you can in patch 5 make the data
> structure containing mm_nodes private.
>
> But it's not the same data structure. I don't think patch 5 really
> depends on any of the previous patches, because all the other patches
> deal with ttm_resource.mm_nodes and patch 5 deals with
> amdgpu_vram_reservation.mm_nodes.

No, you are close but still misunderstood the intention here a bit.

I want to stop using drm_mm_node directly in the amdgpu code so that we 
are able to switch the VRAM management to a buddy allocator sooner or later.

mm_nodes and amdgpu_vram_reservation.mm_nodes are both of that type.

Regards,
Christian.

>
> Regards,
>    Felix
>
>
>> Christian.
>>
>>>
>>>> ---
>>>>    drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55
>>>> +++++++++-----------------
>>>>    drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>>>>    drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>>>>    3 files changed, 20 insertions(+), 40 deletions(-)
>>>>
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> index dae51992c607..fa43d332a979 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>> @@ -37,6 +37,7 @@
>>>>    #include "amdgpu_gmc.h"
>>>>    #include "amdgpu_xgmi.h"
>>>>    #include "amdgpu_dma_buf.h"
>>>> +#include "amdgpu_res_cursor.h"
>>>>    #include "kfd_svm.h"
>>>>      /**
>>>> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct
>>>> amdgpu_vm_update_params *params,
>>>>     * @last: last mapped entry
>>>>     * @flags: flags for the entries
>>>>     * @offset: offset into nodes and pages_addr
>>>> - * @nodes: array of drm_mm_nodes with the MC addresses
>>>> + * @res: ttm_resource to map
>>>>     * @pages_addr: DMA addresses to use for mapping
>>>>     * @fence: optional resulting fence
>>>>     *
>>>> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct
>>>> amdgpu_device *adev,
>>>>                    bool unlocked, struct dma_resv *resv,
>>>>                    uint64_t start, uint64_t last,
>>>>                    uint64_t flags, uint64_t offset,
>>>> -                struct drm_mm_node *nodes,
>>>> +                struct ttm_resource *res,
>>>>                    dma_addr_t *pages_addr,
>>>>                    struct dma_fence **fence)
>>>>    {
>>>>        struct amdgpu_vm_update_params params;
>>>> +    struct amdgpu_res_cursor cursor;
>>>>        enum amdgpu_sync_mode sync_mode;
>>>> -    uint64_t pfn;
>>>>        int r;
>>>>          memset(&params, 0, sizeof(params));
>>>> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct
>>>> amdgpu_device *adev,
>>>>        else
>>>>            sync_mode = AMDGPU_SYNC_EXPLICIT;
>>>>    -    pfn = offset >> PAGE_SHIFT;
>>>> -    if (nodes) {
>>>> -        while (pfn >= nodes->size) {
>>>> -            pfn -= nodes->size;
>>>> -            ++nodes;
>>>> -        }
>>>> -    }
>>>> -
>>>>        amdgpu_vm_eviction_lock(vm);
>>>>        if (vm->evicting) {
>>>>            r = -EBUSY;
>>>> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct
>>>> amdgpu_device *adev,
>>>>        if (r)
>>>>            goto error_unlock;
>>>>    -    do {
>>>> +    amdgpu_res_first(res, offset, (last - start + 1) *
>>>> AMDGPU_GPU_PAGE_SIZE,
>>>> +             &cursor);
>>>> +    while (cursor.remaining) {
>>>>            uint64_t tmp, num_entries, addr;
>>>>    -
>>>> -        num_entries = last - start + 1;
>>>> -        if (nodes) {
>>>> -            addr = nodes->start << PAGE_SHIFT;
>>>> -            num_entries = min((nodes->size - pfn) *
>>>> -                AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
>>>> -        } else {
>>>> -            addr = 0;
>>>> -        }
>>>> -
>>>> +        num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>>>>            if (pages_addr) {
>>>>                bool contiguous = true;
>>>>                  if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
>>>> +                uint64_t pfn = cursor.start >> PAGE_SHIFT;
>>>>                    uint64_t count;
>>>>                      contiguous = pages_addr[pfn + 1] ==
>>>> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct
>>>> amdgpu_device *adev,
>>>>                }
>>>>                  if (!contiguous) {
>>>> -                addr = pfn << PAGE_SHIFT;
>>>> +                addr = cursor.start;
>>>>                    params.pages_addr = pages_addr;
>>>>                } else {
>>>> -                addr = pages_addr[pfn];
>>>> +                addr = pages_addr[cursor.start >> PAGE_SHIFT];
>>>>                    params.pages_addr = NULL;
>>>>                }
>>>>              } else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
>>>> -            addr += bo_adev->vm_manager.vram_base_offset;
>>>> -            addr += pfn << PAGE_SHIFT;
>>>> +            addr = bo_adev->vm_manager.vram_base_offset +
>>>> +                cursor.start;
>>>> +        } else {
>>>> +            addr = 0;
>>>>            }
>>>>              tmp = start + num_entries;
>>>> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct
>>>> amdgpu_device *adev,
>>>>            if (r)
>>>>                goto error_unlock;
>>>>    -        pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
>>>> -        if (nodes && nodes->size == pfn) {
>>>> -            pfn = 0;
>>>> -            ++nodes;
>>>> -        }
>>>> +        amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
>>>>            start = tmp;
>>>> -
>>>> -    } while (unlikely(start != last + 1));
>>>> +    };
>>>>          r = vm->update_funcs->commit(&params, fence);
>>>>    @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>        struct amdgpu_bo_va_mapping *mapping;
>>>>        dma_addr_t *pages_addr = NULL;
>>>>        struct ttm_resource *mem;
>>>> -    struct drm_mm_node *nodes;
>>>>        struct dma_fence **last_update;
>>>>        struct dma_resv *resv;
>>>>        uint64_t flags;
>>>> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>          if (clear || !bo) {
>>>>            mem = NULL;
>>>> -        nodes = NULL;
>>>>            resv = vm->root.base.bo->tbo.base.resv;
>>>>        } else {
>>>>            struct drm_gem_object *obj = &bo->tbo.base;
>>>> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>                    bo = gem_to_amdgpu_bo(gobj);
>>>>            }
>>>>            mem = &bo->tbo.mem;
>>>> -        nodes = mem->mm_node;
>>>>            if (mem->mem_type == TTM_PL_TT)
>>>>                pages_addr = bo->tbo.ttm->dma_address;
>>>>        }
>>>> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>            r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false,
>>>> false,
>>>>                            resv, mapping->start,
>>>>                            mapping->last, update_flags,
>>>> -                        mapping->offset, nodes,
>>>> +                        mapping->offset, mem,
>>>>                            pages_addr, last_update);
>>>>            if (r)
>>>>                return r;
>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>> index e5a3f18be2b7..1ae5ea8db497 100644
>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct
>>>> amdgpu_device *adev,
>>>>                    bool unlocked, struct dma_resv *resv,
>>>>                    uint64_t start, uint64_t last,
>>>>                    uint64_t flags, uint64_t offset,
>>>> -                struct drm_mm_node *nodes,
>>>> +                struct ttm_resource *res,
>>>>                    dma_addr_t *pages_addr,
>>>>                    struct dma_fence **fence);
>>>>    int amdgpu_vm_bo_update(struct amdgpu_device *adev,
>>>> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>> b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>> index e4ce97ab6e26..0b0e76e16ddc 100644
>>>> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device
>>>> *adev, struct amdgpu_vm *vm,
>>>>                        prange->mapping.start,
>>>>                        prange->mapping.last, pte_flags,
>>>>                        prange->mapping.offset,
>>>> -                    prange->ttm_res ?
>>>> -                        prange->ttm_res->mm_node : NULL,
>>>> +                    prange->ttm_res,
>>>>                        dma_addr, &vm->last_update);
>>>>        if (r) {
>>>>            pr_debug("failed %d to map to gpu 0x%lx\n", r,
>>>> prange->start);

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-30 13:50       ` Christian König
@ 2021-04-30 14:57         ` Felix Kuehling
  2021-04-30 14:58           ` Christian König
  0 siblings, 1 reply; 12+ messages in thread
From: Felix Kuehling @ 2021-04-30 14:57 UTC (permalink / raw)
  To: Christian König, amd-gfx

Am 2021-04-30 um 9:50 a.m. schrieb Christian König:
>
>
> Am 30.04.21 um 15:45 schrieb Felix Kuehling:
>> Am 2021-04-30 um 4:54 a.m. schrieb Christian König:
>>>
>>> Am 30.04.21 um 10:32 schrieb Felix Kuehling:
>>>> Am 2021-04-27 um 6:54 a.m. schrieb Christian König:
>>>>> Now that we found the underlying problem we can re-apply this patch.
>>>>>
>>>>> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>>>>>
>>>>> v2: rebase on KFD changes
>>>>>
>>>>> Signed-off-by: Christian König <christian.koenig@amd.com>
>>>> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>
>>> Thanks, just for this patch or the series?
>> I meant it for the patch. I could probably extend it to the series. But
>> I'm a bit confused how the patches relate to each other. There seems to
>> be an implication that everything leads up to enabling patch 5, i.e.
>> removing users of mm_nodes so that you can in patch 5 make the data
>> structure containing mm_nodes private.
>>
>> But it's not the same data structure. I don't think patch 5 really
>> depends on any of the previous patches, because all the other patches
>> deal with ttm_resource.mm_nodes and patch 5 deals with
>> amdgpu_vram_reservation.mm_nodes.
>
> No, you are close but still misunderstood the intention here a bit.
>
> I want to stop using drm_mm_node directly in the amdgpu code so that
> we are able to switch the VRAM management to a buddy allocator sooner
> or later.
>
> mm_nodes and amdgpu_vram_reservation.mm_nodes are both of that type.

Yes, I got that. So all of these patches are preparation for something
bigger. That's the part that wasn't clear. The series is

Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>

We still have a few uses of mm_node in kfd_migrate.c. That code is about
to change as Alejandro is beginning to rework our VRAM management for
HMM migration to allow partial migration failures and mixed VRAM/System
mappings. But we'll still need a way to find VRAM addresses within a BO.

Regards,
  Felix


>
> Regards,
> Christian.
>
>>
>> Regards,
>>    Felix
>>
>>
>>> Christian.
>>>
>>>>
>>>>> ---
>>>>>    drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55
>>>>> +++++++++-----------------
>>>>>    drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>>>>>    drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>>>>>    3 files changed, 20 insertions(+), 40 deletions(-)
>>>>>
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>> index dae51992c607..fa43d332a979 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>> @@ -37,6 +37,7 @@
>>>>>    #include "amdgpu_gmc.h"
>>>>>    #include "amdgpu_xgmi.h"
>>>>>    #include "amdgpu_dma_buf.h"
>>>>> +#include "amdgpu_res_cursor.h"
>>>>>    #include "kfd_svm.h"
>>>>>      /**
>>>>> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct
>>>>> amdgpu_vm_update_params *params,
>>>>>     * @last: last mapped entry
>>>>>     * @flags: flags for the entries
>>>>>     * @offset: offset into nodes and pages_addr
>>>>> - * @nodes: array of drm_mm_nodes with the MC addresses
>>>>> + * @res: ttm_resource to map
>>>>>     * @pages_addr: DMA addresses to use for mapping
>>>>>     * @fence: optional resulting fence
>>>>>     *
>>>>> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>> amdgpu_device *adev,
>>>>>                    bool unlocked, struct dma_resv *resv,
>>>>>                    uint64_t start, uint64_t last,
>>>>>                    uint64_t flags, uint64_t offset,
>>>>> -                struct drm_mm_node *nodes,
>>>>> +                struct ttm_resource *res,
>>>>>                    dma_addr_t *pages_addr,
>>>>>                    struct dma_fence **fence)
>>>>>    {
>>>>>        struct amdgpu_vm_update_params params;
>>>>> +    struct amdgpu_res_cursor cursor;
>>>>>        enum amdgpu_sync_mode sync_mode;
>>>>> -    uint64_t pfn;
>>>>>        int r;
>>>>>          memset(&params, 0, sizeof(params));
>>>>> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>> amdgpu_device *adev,
>>>>>        else
>>>>>            sync_mode = AMDGPU_SYNC_EXPLICIT;
>>>>>    -    pfn = offset >> PAGE_SHIFT;
>>>>> -    if (nodes) {
>>>>> -        while (pfn >= nodes->size) {
>>>>> -            pfn -= nodes->size;
>>>>> -            ++nodes;
>>>>> -        }
>>>>> -    }
>>>>> -
>>>>>        amdgpu_vm_eviction_lock(vm);
>>>>>        if (vm->evicting) {
>>>>>            r = -EBUSY;
>>>>> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>> amdgpu_device *adev,
>>>>>        if (r)
>>>>>            goto error_unlock;
>>>>>    -    do {
>>>>> +    amdgpu_res_first(res, offset, (last - start + 1) *
>>>>> AMDGPU_GPU_PAGE_SIZE,
>>>>> +             &cursor);
>>>>> +    while (cursor.remaining) {
>>>>>            uint64_t tmp, num_entries, addr;
>>>>>    -
>>>>> -        num_entries = last - start + 1;
>>>>> -        if (nodes) {
>>>>> -            addr = nodes->start << PAGE_SHIFT;
>>>>> -            num_entries = min((nodes->size - pfn) *
>>>>> -                AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
>>>>> -        } else {
>>>>> -            addr = 0;
>>>>> -        }
>>>>> -
>>>>> +        num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>>>>>            if (pages_addr) {
>>>>>                bool contiguous = true;
>>>>>                  if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
>>>>> +                uint64_t pfn = cursor.start >> PAGE_SHIFT;
>>>>>                    uint64_t count;
>>>>>                      contiguous = pages_addr[pfn + 1] ==
>>>>> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>> amdgpu_device *adev,
>>>>>                }
>>>>>                  if (!contiguous) {
>>>>> -                addr = pfn << PAGE_SHIFT;
>>>>> +                addr = cursor.start;
>>>>>                    params.pages_addr = pages_addr;
>>>>>                } else {
>>>>> -                addr = pages_addr[pfn];
>>>>> +                addr = pages_addr[cursor.start >> PAGE_SHIFT];
>>>>>                    params.pages_addr = NULL;
>>>>>                }
>>>>>              } else if (flags & (AMDGPU_PTE_VALID |
>>>>> AMDGPU_PTE_PRT)) {
>>>>> -            addr += bo_adev->vm_manager.vram_base_offset;
>>>>> -            addr += pfn << PAGE_SHIFT;
>>>>> +            addr = bo_adev->vm_manager.vram_base_offset +
>>>>> +                cursor.start;
>>>>> +        } else {
>>>>> +            addr = 0;
>>>>>            }
>>>>>              tmp = start + num_entries;
>>>>> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>> amdgpu_device *adev,
>>>>>            if (r)
>>>>>                goto error_unlock;
>>>>>    -        pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
>>>>> -        if (nodes && nodes->size == pfn) {
>>>>> -            pfn = 0;
>>>>> -            ++nodes;
>>>>> -        }
>>>>> +        amdgpu_res_next(&cursor, num_entries *
>>>>> AMDGPU_GPU_PAGE_SIZE);
>>>>>            start = tmp;
>>>>> -
>>>>> -    } while (unlikely(start != last + 1));
>>>>> +    };
>>>>>          r = vm->update_funcs->commit(&params, fence);
>>>>>    @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>        struct amdgpu_bo_va_mapping *mapping;
>>>>>        dma_addr_t *pages_addr = NULL;
>>>>>        struct ttm_resource *mem;
>>>>> -    struct drm_mm_node *nodes;
>>>>>        struct dma_fence **last_update;
>>>>>        struct dma_resv *resv;
>>>>>        uint64_t flags;
>>>>> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>          if (clear || !bo) {
>>>>>            mem = NULL;
>>>>> -        nodes = NULL;
>>>>>            resv = vm->root.base.bo->tbo.base.resv;
>>>>>        } else {
>>>>>            struct drm_gem_object *obj = &bo->tbo.base;
>>>>> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>                    bo = gem_to_amdgpu_bo(gobj);
>>>>>            }
>>>>>            mem = &bo->tbo.mem;
>>>>> -        nodes = mem->mm_node;
>>>>>            if (mem->mem_type == TTM_PL_TT)
>>>>>                pages_addr = bo->tbo.ttm->dma_address;
>>>>>        }
>>>>> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>            r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false,
>>>>> false,
>>>>>                            resv, mapping->start,
>>>>>                            mapping->last, update_flags,
>>>>> -                        mapping->offset, nodes,
>>>>> +                        mapping->offset, mem,
>>>>>                            pages_addr, last_update);
>>>>>            if (r)
>>>>>                return r;
>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>> index e5a3f18be2b7..1ae5ea8db497 100644
>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>> amdgpu_device *adev,
>>>>>                    bool unlocked, struct dma_resv *resv,
>>>>>                    uint64_t start, uint64_t last,
>>>>>                    uint64_t flags, uint64_t offset,
>>>>> -                struct drm_mm_node *nodes,
>>>>> +                struct ttm_resource *res,
>>>>>                    dma_addr_t *pages_addr,
>>>>>                    struct dma_fence **fence);
>>>>>    int amdgpu_vm_bo_update(struct amdgpu_device *adev,
>>>>> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>> b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>> index e4ce97ab6e26..0b0e76e16ddc 100644
>>>>> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device
>>>>> *adev, struct amdgpu_vm *vm,
>>>>>                        prange->mapping.start,
>>>>>                        prange->mapping.last, pte_flags,
>>>>>                        prange->mapping.offset,
>>>>> -                    prange->ttm_res ?
>>>>> -                        prange->ttm_res->mm_node : NULL,
>>>>> +                    prange->ttm_res,
>>>>>                        dma_addr, &vm->last_update);
>>>>>        if (r) {
>>>>>            pr_debug("failed %d to map to gpu 0x%lx\n", r,
>>>>> prange->start);
>
_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-30 14:57         ` Felix Kuehling
@ 2021-04-30 14:58           ` Christian König
  0 siblings, 0 replies; 12+ messages in thread
From: Christian König @ 2021-04-30 14:58 UTC (permalink / raw)
  To: Felix Kuehling, amd-gfx

Am 30.04.21 um 16:57 schrieb Felix Kuehling:
> Am 2021-04-30 um 9:50 a.m. schrieb Christian König:
>>
>> Am 30.04.21 um 15:45 schrieb Felix Kuehling:
>>> Am 2021-04-30 um 4:54 a.m. schrieb Christian König:
>>>> Am 30.04.21 um 10:32 schrieb Felix Kuehling:
>>>>> Am 2021-04-27 um 6:54 a.m. schrieb Christian König:
>>>>>> Now that we found the underlying problem we can re-apply this patch.
>>>>>>
>>>>>> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>>>>>>
>>>>>> v2: rebase on KFD changes
>>>>>>
>>>>>> Signed-off-by: Christian König <christian.koenig@amd.com>
>>>>> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>
>>>> Thanks, just for this patch or the series?
>>> I meant it for the patch. I could probably extend it to the series. But
>>> I'm a bit confused how the patches relate to each other. There seems to
>>> be an implication that everything leads up to enabling patch 5, i.e.
>>> removing users of mm_nodes so that you can in patch 5 make the data
>>> structure containing mm_nodes private.
>>>
>>> But it's not the same data structure. I don't think patch 5 really
>>> depends on any of the previous patches, because all the other patches
>>> deal with ttm_resource.mm_nodes and patch 5 deals with
>>> amdgpu_vram_reservation.mm_nodes.
>> No, you are close but still misunderstood the intention here a bit.
>>
>> I want to stop using drm_mm_node directly in the amdgpu code so that
>> we are able to switch the VRAM management to a buddy allocator sooner
>> or later.
>>
>> mm_nodes and amdgpu_vram_reservation.mm_nodes are both of that type.
> Yes, I got that. So all of these patches are preparation for something
> bigger. That's the part that wasn't clear. The series is
>
> Acked-by: Felix Kuehling <Felix.Kuehling@amd.com>
>
> We still have a few uses of mm_node in kfd_migrate.c. That code is about
> to change as Alejandro is beginning to rework our VRAM management for
> HMM migration to allow partial migration failures and mixed VRAM/System
> mappings. But we'll still need a way to find VRAM addresses within a BO.

Using amdgpu_res_coursor.h for that should work.

Regards,
Christian.

>
> Regards,
>    Felix
>
>
>> Regards,
>> Christian.
>>
>>> Regards,
>>>     Felix
>>>
>>>
>>>> Christian.
>>>>
>>>>>> ---
>>>>>>     drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55
>>>>>> +++++++++-----------------
>>>>>>     drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>>>>>>     drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>>>>>>     3 files changed, 20 insertions(+), 40 deletions(-)
>>>>>>
>>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>>> index dae51992c607..fa43d332a979 100644
>>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
>>>>>> @@ -37,6 +37,7 @@
>>>>>>     #include "amdgpu_gmc.h"
>>>>>>     #include "amdgpu_xgmi.h"
>>>>>>     #include "amdgpu_dma_buf.h"
>>>>>> +#include "amdgpu_res_cursor.h"
>>>>>>     #include "kfd_svm.h"
>>>>>>       /**
>>>>>> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct
>>>>>> amdgpu_vm_update_params *params,
>>>>>>      * @last: last mapped entry
>>>>>>      * @flags: flags for the entries
>>>>>>      * @offset: offset into nodes and pages_addr
>>>>>> - * @nodes: array of drm_mm_nodes with the MC addresses
>>>>>> + * @res: ttm_resource to map
>>>>>>      * @pages_addr: DMA addresses to use for mapping
>>>>>>      * @fence: optional resulting fence
>>>>>>      *
>>>>>> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>>> amdgpu_device *adev,
>>>>>>                     bool unlocked, struct dma_resv *resv,
>>>>>>                     uint64_t start, uint64_t last,
>>>>>>                     uint64_t flags, uint64_t offset,
>>>>>> -                struct drm_mm_node *nodes,
>>>>>> +                struct ttm_resource *res,
>>>>>>                     dma_addr_t *pages_addr,
>>>>>>                     struct dma_fence **fence)
>>>>>>     {
>>>>>>         struct amdgpu_vm_update_params params;
>>>>>> +    struct amdgpu_res_cursor cursor;
>>>>>>         enum amdgpu_sync_mode sync_mode;
>>>>>> -    uint64_t pfn;
>>>>>>         int r;
>>>>>>           memset(&params, 0, sizeof(params));
>>>>>> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>>> amdgpu_device *adev,
>>>>>>         else
>>>>>>             sync_mode = AMDGPU_SYNC_EXPLICIT;
>>>>>>     -    pfn = offset >> PAGE_SHIFT;
>>>>>> -    if (nodes) {
>>>>>> -        while (pfn >= nodes->size) {
>>>>>> -            pfn -= nodes->size;
>>>>>> -            ++nodes;
>>>>>> -        }
>>>>>> -    }
>>>>>> -
>>>>>>         amdgpu_vm_eviction_lock(vm);
>>>>>>         if (vm->evicting) {
>>>>>>             r = -EBUSY;
>>>>>> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>>> amdgpu_device *adev,
>>>>>>         if (r)
>>>>>>             goto error_unlock;
>>>>>>     -    do {
>>>>>> +    amdgpu_res_first(res, offset, (last - start + 1) *
>>>>>> AMDGPU_GPU_PAGE_SIZE,
>>>>>> +             &cursor);
>>>>>> +    while (cursor.remaining) {
>>>>>>             uint64_t tmp, num_entries, addr;
>>>>>>     -
>>>>>> -        num_entries = last - start + 1;
>>>>>> -        if (nodes) {
>>>>>> -            addr = nodes->start << PAGE_SHIFT;
>>>>>> -            num_entries = min((nodes->size - pfn) *
>>>>>> -                AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
>>>>>> -        } else {
>>>>>> -            addr = 0;
>>>>>> -        }
>>>>>> -
>>>>>> +        num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>>>>>>             if (pages_addr) {
>>>>>>                 bool contiguous = true;
>>>>>>                   if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
>>>>>> +                uint64_t pfn = cursor.start >> PAGE_SHIFT;
>>>>>>                     uint64_t count;
>>>>>>                       contiguous = pages_addr[pfn + 1] ==
>>>>>> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>>> amdgpu_device *adev,
>>>>>>                 }
>>>>>>                   if (!contiguous) {
>>>>>> -                addr = pfn << PAGE_SHIFT;
>>>>>> +                addr = cursor.start;
>>>>>>                     params.pages_addr = pages_addr;
>>>>>>                 } else {
>>>>>> -                addr = pages_addr[pfn];
>>>>>> +                addr = pages_addr[cursor.start >> PAGE_SHIFT];
>>>>>>                     params.pages_addr = NULL;
>>>>>>                 }
>>>>>>               } else if (flags & (AMDGPU_PTE_VALID |
>>>>>> AMDGPU_PTE_PRT)) {
>>>>>> -            addr += bo_adev->vm_manager.vram_base_offset;
>>>>>> -            addr += pfn << PAGE_SHIFT;
>>>>>> +            addr = bo_adev->vm_manager.vram_base_offset +
>>>>>> +                cursor.start;
>>>>>> +        } else {
>>>>>> +            addr = 0;
>>>>>>             }
>>>>>>               tmp = start + num_entries;
>>>>>> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>>> amdgpu_device *adev,
>>>>>>             if (r)
>>>>>>                 goto error_unlock;
>>>>>>     -        pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
>>>>>> -        if (nodes && nodes->size == pfn) {
>>>>>> -            pfn = 0;
>>>>>> -            ++nodes;
>>>>>> -        }
>>>>>> +        amdgpu_res_next(&cursor, num_entries *
>>>>>> AMDGPU_GPU_PAGE_SIZE);
>>>>>>             start = tmp;
>>>>>> -
>>>>>> -    } while (unlikely(start != last + 1));
>>>>>> +    };
>>>>>>           r = vm->update_funcs->commit(&params, fence);
>>>>>>     @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>>         struct amdgpu_bo_va_mapping *mapping;
>>>>>>         dma_addr_t *pages_addr = NULL;
>>>>>>         struct ttm_resource *mem;
>>>>>> -    struct drm_mm_node *nodes;
>>>>>>         struct dma_fence **last_update;
>>>>>>         struct dma_resv *resv;
>>>>>>         uint64_t flags;
>>>>>> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>>           if (clear || !bo) {
>>>>>>             mem = NULL;
>>>>>> -        nodes = NULL;
>>>>>>             resv = vm->root.base.bo->tbo.base.resv;
>>>>>>         } else {
>>>>>>             struct drm_gem_object *obj = &bo->tbo.base;
>>>>>> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>>                     bo = gem_to_amdgpu_bo(gobj);
>>>>>>             }
>>>>>>             mem = &bo->tbo.mem;
>>>>>> -        nodes = mem->mm_node;
>>>>>>             if (mem->mem_type == TTM_PL_TT)
>>>>>>                 pages_addr = bo->tbo.ttm->dma_address;
>>>>>>         }
>>>>>> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device
>>>>>> *adev, struct amdgpu_bo_va *bo_va,
>>>>>>             r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false,
>>>>>> false,
>>>>>>                             resv, mapping->start,
>>>>>>                             mapping->last, update_flags,
>>>>>> -                        mapping->offset, nodes,
>>>>>> +                        mapping->offset, mem,
>>>>>>                             pages_addr, last_update);
>>>>>>             if (r)
>>>>>>                 return r;
>>>>>> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>>> b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>>> index e5a3f18be2b7..1ae5ea8db497 100644
>>>>>> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>>> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
>>>>>> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct
>>>>>> amdgpu_device *adev,
>>>>>>                     bool unlocked, struct dma_resv *resv,
>>>>>>                     uint64_t start, uint64_t last,
>>>>>>                     uint64_t flags, uint64_t offset,
>>>>>> -                struct drm_mm_node *nodes,
>>>>>> +                struct ttm_resource *res,
>>>>>>                     dma_addr_t *pages_addr,
>>>>>>                     struct dma_fence **fence);
>>>>>>     int amdgpu_vm_bo_update(struct amdgpu_device *adev,
>>>>>> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>>> b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>>> index e4ce97ab6e26..0b0e76e16ddc 100644
>>>>>> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>>> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
>>>>>> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device
>>>>>> *adev, struct amdgpu_vm *vm,
>>>>>>                         prange->mapping.start,
>>>>>>                         prange->mapping.last, pte_flags,
>>>>>>                         prange->mapping.offset,
>>>>>> -                    prange->ttm_res ?
>>>>>> -                        prange->ttm_res->mm_node : NULL,
>>>>>> +                    prange->ttm_res,
>>>>>>                         dma_addr, &vm->last_update);
>>>>>>         if (r) {
>>>>>>             pr_debug("failed %d to map to gpu 0x%lx\n", r,
>>>>>> prange->start);

_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

* Re: [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2
  2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
                   ` (4 preceding siblings ...)
  2021-04-30  8:32 ` [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Felix Kuehling
@ 2021-04-30 15:19 ` Nirmoy
  5 siblings, 0 replies; 12+ messages in thread
From: Nirmoy @ 2021-04-30 15:19 UTC (permalink / raw)
  To: amd-gfx

Series is Tested-by: Nirmoy Das <nirmoy.das@amd.com>

On 4/27/21 12:54 PM, Christian König wrote:
> Now that we found the underlying problem we can re-apply this patch.
>
> This reverts commit 867fee7f8821ff42e7308088cf0c3450ac49c17c.
>
> v2: rebase on KFD changes
>
> Signed-off-by: Christian König <christian.koenig@amd.com>
> ---
>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 55 +++++++++-----------------
>   drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h |  2 +-
>   drivers/gpu/drm/amd/amdkfd/kfd_svm.c   |  3 +-
>   3 files changed, 20 insertions(+), 40 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index dae51992c607..fa43d332a979 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -37,6 +37,7 @@
>   #include "amdgpu_gmc.h"
>   #include "amdgpu_xgmi.h"
>   #include "amdgpu_dma_buf.h"
> +#include "amdgpu_res_cursor.h"
>   #include "kfd_svm.h"
>   
>   /**
> @@ -1606,7 +1607,7 @@ static int amdgpu_vm_update_ptes(struct amdgpu_vm_update_params *params,
>    * @last: last mapped entry
>    * @flags: flags for the entries
>    * @offset: offset into nodes and pages_addr
> - * @nodes: array of drm_mm_nodes with the MC addresses
> + * @res: ttm_resource to map
>    * @pages_addr: DMA addresses to use for mapping
>    * @fence: optional resulting fence
>    *
> @@ -1621,13 +1622,13 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   				bool unlocked, struct dma_resv *resv,
>   				uint64_t start, uint64_t last,
>   				uint64_t flags, uint64_t offset,
> -				struct drm_mm_node *nodes,
> +				struct ttm_resource *res,
>   				dma_addr_t *pages_addr,
>   				struct dma_fence **fence)
>   {
>   	struct amdgpu_vm_update_params params;
> +	struct amdgpu_res_cursor cursor;
>   	enum amdgpu_sync_mode sync_mode;
> -	uint64_t pfn;
>   	int r;
>   
>   	memset(&params, 0, sizeof(params));
> @@ -1645,14 +1646,6 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   	else
>   		sync_mode = AMDGPU_SYNC_EXPLICIT;
>   
> -	pfn = offset >> PAGE_SHIFT;
> -	if (nodes) {
> -		while (pfn >= nodes->size) {
> -			pfn -= nodes->size;
> -			++nodes;
> -		}
> -	}
> -
>   	amdgpu_vm_eviction_lock(vm);
>   	if (vm->evicting) {
>   		r = -EBUSY;
> @@ -1671,23 +1664,17 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   	if (r)
>   		goto error_unlock;
>   
> -	do {
> +	amdgpu_res_first(res, offset, (last - start + 1) * AMDGPU_GPU_PAGE_SIZE,
> +			 &cursor);
> +	while (cursor.remaining) {
>   		uint64_t tmp, num_entries, addr;
>   
> -
> -		num_entries = last - start + 1;
> -		if (nodes) {
> -			addr = nodes->start << PAGE_SHIFT;
> -			num_entries = min((nodes->size - pfn) *
> -				AMDGPU_GPU_PAGES_IN_CPU_PAGE, num_entries);
> -		} else {
> -			addr = 0;
> -		}
> -
> +		num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
>   		if (pages_addr) {
>   			bool contiguous = true;
>   
>   			if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
> +				uint64_t pfn = cursor.start >> PAGE_SHIFT;
>   				uint64_t count;
>   
>   				contiguous = pages_addr[pfn + 1] ==
> @@ -1707,16 +1694,18 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   			}
>   
>   			if (!contiguous) {
> -				addr = pfn << PAGE_SHIFT;
> +				addr = cursor.start;
>   				params.pages_addr = pages_addr;
>   			} else {
> -				addr = pages_addr[pfn];
> +				addr = pages_addr[cursor.start >> PAGE_SHIFT];
>   				params.pages_addr = NULL;
>   			}
>   
>   		} else if (flags & (AMDGPU_PTE_VALID | AMDGPU_PTE_PRT)) {
> -			addr += bo_adev->vm_manager.vram_base_offset;
> -			addr += pfn << PAGE_SHIFT;
> +			addr = bo_adev->vm_manager.vram_base_offset +
> +				cursor.start;
> +		} else {
> +			addr = 0;
>   		}
>   
>   		tmp = start + num_entries;
> @@ -1724,14 +1713,9 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   		if (r)
>   			goto error_unlock;
>   
> -		pfn += num_entries / AMDGPU_GPU_PAGES_IN_CPU_PAGE;
> -		if (nodes && nodes->size == pfn) {
> -			pfn = 0;
> -			++nodes;
> -		}
> +		amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
>   		start = tmp;
> -
> -	} while (unlikely(start != last + 1));
> +	};
>   
>   	r = vm->update_funcs->commit(&params, fence);
>   
> @@ -1760,7 +1744,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>   	struct amdgpu_bo_va_mapping *mapping;
>   	dma_addr_t *pages_addr = NULL;
>   	struct ttm_resource *mem;
> -	struct drm_mm_node *nodes;
>   	struct dma_fence **last_update;
>   	struct dma_resv *resv;
>   	uint64_t flags;
> @@ -1769,7 +1752,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>   
>   	if (clear || !bo) {
>   		mem = NULL;
> -		nodes = NULL;
>   		resv = vm->root.base.bo->tbo.base.resv;
>   	} else {
>   		struct drm_gem_object *obj = &bo->tbo.base;
> @@ -1784,7 +1766,6 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>   				bo = gem_to_amdgpu_bo(gobj);
>   		}
>   		mem = &bo->tbo.mem;
> -		nodes = mem->mm_node;
>   		if (mem->mem_type == TTM_PL_TT)
>   			pages_addr = bo->tbo.ttm->dma_address;
>   	}
> @@ -1833,7 +1814,7 @@ int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va,
>   		r = amdgpu_vm_bo_update_mapping(adev, bo_adev, vm, false, false,
>   						resv, mapping->start,
>   						mapping->last, update_flags,
> -						mapping->offset, nodes,
> +						mapping->offset, mem,
>   						pages_addr, last_update);
>   		if (r)
>   			return r;
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> index e5a3f18be2b7..1ae5ea8db497 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> @@ -402,7 +402,7 @@ int amdgpu_vm_bo_update_mapping(struct amdgpu_device *adev,
>   				bool unlocked, struct dma_resv *resv,
>   				uint64_t start, uint64_t last,
>   				uint64_t flags, uint64_t offset,
> -				struct drm_mm_node *nodes,
> +				struct ttm_resource *res,
>   				dma_addr_t *pages_addr,
>   				struct dma_fence **fence);
>   int amdgpu_vm_bo_update(struct amdgpu_device *adev,
> diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> index e4ce97ab6e26..0b0e76e16ddc 100644
> --- a/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> +++ b/drivers/gpu/drm/amd/amdkfd/kfd_svm.c
> @@ -1151,8 +1151,7 @@ svm_range_map_to_gpu(struct amdgpu_device *adev, struct amdgpu_vm *vm,
>   					prange->mapping.start,
>   					prange->mapping.last, pte_flags,
>   					prange->mapping.offset,
> -					prange->ttm_res ?
> -						prange->ttm_res->mm_node : NULL,
> +					prange->ttm_res,
>   					dma_addr, &vm->last_update);
>   	if (r) {
>   		pr_debug("failed %d to map to gpu 0x%lx\n", r, prange->start);
_______________________________________________
amd-gfx mailing list
amd-gfx@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/amd-gfx

^ permalink raw reply	[flat|nested] 12+ messages in thread

end of thread, other threads:[~2021-04-30 15:20 UTC | newest]

Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-04-27 10:54 [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Christian König
2021-04-27 10:54 ` [PATCH 2/5] drm/amdgpu: use cursor functions in amdgpu_bo_in_cpu_visible_vram Christian König
2021-04-27 10:54 ` [PATCH 3/5] drm/amdgpu: set the contiguous flag if possible Christian König
2021-04-27 10:54 ` [PATCH 4/5] drm/amdgpu: check contiguous flags instead of mm_node Christian König
2021-04-27 10:54 ` [PATCH 5/5] drm/amdgpu: move struct amdgpu_vram_reservation into vram mgr Christian König
2021-04-30  8:32 ` [PATCH 1/5] drm/amdgpu: re-apply "use the new cursor in the VM code" v2 Felix Kuehling
2021-04-30  8:54   ` Christian König
2021-04-30 13:45     ` Felix Kuehling
2021-04-30 13:50       ` Christian König
2021-04-30 14:57         ` Felix Kuehling
2021-04-30 14:58           ` Christian König
2021-04-30 15:19 ` Nirmoy

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.