* [PATCH 0/3] Fix non-contiguous VRAM BO access in Xe
@ 2024-10-18 21:11 Matthew Brost
2024-10-18 21:11 ` [PATCH 1/3] drm/ttm: Add ttm_bo_access Matthew Brost
` (2 more replies)
0 siblings, 3 replies; 4+ messages in thread
From: Matthew Brost @ 2024-10-18 21:11 UTC (permalink / raw)
To: intel-xe, dri-devel; +Cc: matthew.auld, thomas.hellstrom
Mapping a non-contiguous VRAM BO doesn't work, start to fix this.
A follow up series should cleanup any remaining mapping of
non-contiguous VRAM BOs, add non-visible access support to
xe_ttm_access_memory, and warn / error on mapping a BO which cannot be
mapped.
Matthew Brost (3):
drm/ttm: Add ttm_bo_access
drm/xe: Add xe_ttm_access_memory
drm/xe: Use ttm_bo_access in xe_vm_snapshot_capture_delayed
drivers/gpu/drm/ttm/ttm_bo_util.c | 85 +++++++++++++++++++++++++++++++
drivers/gpu/drm/ttm/ttm_bo_vm.c | 65 +----------------------
drivers/gpu/drm/xe/xe_bo.c | 57 +++++++++++++++++++--
drivers/gpu/drm/xe/xe_vm.c | 17 +++----
include/drm/ttm/ttm_bo.h | 2 +
5 files changed, 148 insertions(+), 78 deletions(-)
--
2.34.1
^ permalink raw reply [flat|nested] 4+ messages in thread
* [PATCH 1/3] drm/ttm: Add ttm_bo_access
2024-10-18 21:11 [PATCH 0/3] Fix non-contiguous VRAM BO access in Xe Matthew Brost
@ 2024-10-18 21:11 ` Matthew Brost
2024-10-18 21:11 ` [PATCH 2/3] drm/xe: Add xe_ttm_access_memory Matthew Brost
2024-10-18 21:11 ` [PATCH 3/3] drm/xe: Use ttm_bo_access in xe_vm_snapshot_capture_delayed Matthew Brost
2 siblings, 0 replies; 4+ messages in thread
From: Matthew Brost @ 2024-10-18 21:11 UTC (permalink / raw)
To: intel-xe, dri-devel; +Cc: matthew.auld, thomas.hellstrom
Non-contiguous VRAM cannot easily be mapped in TTM nor can non-visible
VRAM easily be accessed. Add ttm_bo_access, which is similar to
ttm_bo_vm_access, to access such memory.
Reported-by: Christoph Manszewski <christoph.manszewski@intel.com>
Suggested-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
---
drivers/gpu/drm/ttm/ttm_bo_util.c | 85 +++++++++++++++++++++++++++++++
drivers/gpu/drm/ttm/ttm_bo_vm.c | 65 +----------------------
include/drm/ttm/ttm_bo.h | 2 +
3 files changed, 88 insertions(+), 64 deletions(-)
diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
index d939925efa81..9e427c8342ab 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_util.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
@@ -919,3 +919,88 @@ s64 ttm_lru_walk_for_evict(struct ttm_lru_walk *walk, struct ttm_device *bdev,
return progress;
}
+
+static int ttm_bo_access_kmap(struct ttm_buffer_object *bo,
+ unsigned long offset,
+ uint8_t *buf, int len, int write)
+{
+ unsigned long page = offset >> PAGE_SHIFT;
+ unsigned long bytes_left = len;
+ int ret;
+
+ /* Copy a page at a time, that way no extra virtual address
+ * mapping is needed
+ */
+ offset -= page << PAGE_SHIFT;
+ do {
+ unsigned long bytes = min(bytes_left, PAGE_SIZE - offset);
+ struct ttm_bo_kmap_obj map;
+ void *ptr;
+ bool is_iomem;
+
+ ret = ttm_bo_kmap(bo, page, 1, &map);
+ if (ret)
+ return ret;
+
+ ptr = (uint8_t *)ttm_kmap_obj_virtual(&map, &is_iomem) + offset;
+ WARN_ON_ONCE(is_iomem);
+ if (write)
+ memcpy(ptr, buf, bytes);
+ else
+ memcpy(buf, ptr, bytes);
+ ttm_bo_kunmap(&map);
+
+ page++;
+ buf += bytes;
+ bytes_left -= bytes;
+ offset = 0;
+ } while (bytes_left);
+
+ return len;
+}
+/**
+ * ttm_bo_access - Helper to access a buffer object
+ *
+ * @bo: ttm buffer object
+ * @offset: access offset into buffer object
+ * @buf: pointer to caller memory to read into or write from
+ * @len: length of access
+ * @write: write access
+ *
+ * Utility function to access a buffer object. Useful when buffer object cannot
+ * be easily mapped (non-contiguous, non-visible, etc...).
+ *
+ * Returns:
+ * Number of bytes accessed or errno
+ */
+int ttm_bo_access(struct ttm_buffer_object *bo, unsigned long offset,
+ void *buf, int len, int write)
+{
+ int ret;
+
+ if (len < 1 || (offset + len) > bo->base.size)
+ return -EIO;
+
+ ret = ttm_bo_reserve(bo, true, false, NULL);
+ if (ret)
+ return ret;
+
+ switch (bo->resource->mem_type) {
+ case TTM_PL_SYSTEM:
+ fallthrough;
+ case TTM_PL_TT:
+ ret = ttm_bo_access_kmap(bo, offset, buf, len, write);
+ break;
+ default:
+ if (bo->bdev->funcs->access_memory)
+ ret = bo->bdev->funcs->access_memory(
+ bo, offset, buf, len, write);
+ else
+ ret = -EIO;
+ }
+
+ ttm_bo_unreserve(bo);
+
+ return ret;
+}
+EXPORT_SYMBOL(ttm_bo_access);
diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
index 2c699ed1963a..20b1e5f78684 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
@@ -366,45 +366,6 @@ void ttm_bo_vm_close(struct vm_area_struct *vma)
}
EXPORT_SYMBOL(ttm_bo_vm_close);
-static int ttm_bo_vm_access_kmap(struct ttm_buffer_object *bo,
- unsigned long offset,
- uint8_t *buf, int len, int write)
-{
- unsigned long page = offset >> PAGE_SHIFT;
- unsigned long bytes_left = len;
- int ret;
-
- /* Copy a page at a time, that way no extra virtual address
- * mapping is needed
- */
- offset -= page << PAGE_SHIFT;
- do {
- unsigned long bytes = min(bytes_left, PAGE_SIZE - offset);
- struct ttm_bo_kmap_obj map;
- void *ptr;
- bool is_iomem;
-
- ret = ttm_bo_kmap(bo, page, 1, &map);
- if (ret)
- return ret;
-
- ptr = (uint8_t *)ttm_kmap_obj_virtual(&map, &is_iomem) + offset;
- WARN_ON_ONCE(is_iomem);
- if (write)
- memcpy(ptr, buf, bytes);
- else
- memcpy(buf, ptr, bytes);
- ttm_bo_kunmap(&map);
-
- page++;
- buf += bytes;
- bytes_left -= bytes;
- offset = 0;
- } while (bytes_left);
-
- return len;
-}
-
int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr,
void *buf, int len, int write)
{
@@ -412,32 +373,8 @@ int ttm_bo_vm_access(struct vm_area_struct *vma, unsigned long addr,
unsigned long offset = (addr) - vma->vm_start +
((vma->vm_pgoff - drm_vma_node_start(&bo->base.vma_node))
<< PAGE_SHIFT);
- int ret;
-
- if (len < 1 || (offset + len) > bo->base.size)
- return -EIO;
- ret = ttm_bo_reserve(bo, true, false, NULL);
- if (ret)
- return ret;
-
- switch (bo->resource->mem_type) {
- case TTM_PL_SYSTEM:
- fallthrough;
- case TTM_PL_TT:
- ret = ttm_bo_vm_access_kmap(bo, offset, buf, len, write);
- break;
- default:
- if (bo->bdev->funcs->access_memory)
- ret = bo->bdev->funcs->access_memory(
- bo, offset, buf, len, write);
- else
- ret = -EIO;
- }
-
- ttm_bo_unreserve(bo);
-
- return ret;
+ return ttm_bo_access(bo, offset, buf, len, write);
}
EXPORT_SYMBOL(ttm_bo_vm_access);
diff --git a/include/drm/ttm/ttm_bo.h b/include/drm/ttm/ttm_bo.h
index 5804408815be..8ea11cd8df39 100644
--- a/include/drm/ttm/ttm_bo.h
+++ b/include/drm/ttm/ttm_bo.h
@@ -421,6 +421,8 @@ void ttm_bo_unpin(struct ttm_buffer_object *bo);
int ttm_bo_evict_first(struct ttm_device *bdev,
struct ttm_resource_manager *man,
struct ttm_operation_ctx *ctx);
+int ttm_bo_access(struct ttm_buffer_object *bo, unsigned long offset,
+ void *buf, int len, int write);
vm_fault_t ttm_bo_vm_reserve(struct ttm_buffer_object *bo,
struct vm_fault *vmf);
vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf,
--
2.34.1
^ permalink raw reply related [flat|nested] 4+ messages in thread
* [PATCH 2/3] drm/xe: Add xe_ttm_access_memory
2024-10-18 21:11 [PATCH 0/3] Fix non-contiguous VRAM BO access in Xe Matthew Brost
2024-10-18 21:11 ` [PATCH 1/3] drm/ttm: Add ttm_bo_access Matthew Brost
@ 2024-10-18 21:11 ` Matthew Brost
2024-10-18 21:11 ` [PATCH 3/3] drm/xe: Use ttm_bo_access in xe_vm_snapshot_capture_delayed Matthew Brost
2 siblings, 0 replies; 4+ messages in thread
From: Matthew Brost @ 2024-10-18 21:11 UTC (permalink / raw)
To: intel-xe, dri-devel; +Cc: matthew.auld, thomas.hellstrom
Non-contiguous VRAM cannot easily be mapped in TTM nor can non-visible
VRAM easily be accessed. Add xe_ttm_access_memory which hooks into
ttm_bo_access to access such memory.
Reported-by: Christoph Manszewski <christoph.manszewski@intel.com>
Suggested-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
---
drivers/gpu/drm/xe/xe_bo.c | 57 ++++++++++++++++++++++++++++++++++++--
1 file changed, 54 insertions(+), 3 deletions(-)
diff --git a/drivers/gpu/drm/xe/xe_bo.c b/drivers/gpu/drm/xe/xe_bo.c
index 5b232f2951b1..9a5c1ed7ae97 100644
--- a/drivers/gpu/drm/xe/xe_bo.c
+++ b/drivers/gpu/drm/xe/xe_bo.c
@@ -442,6 +442,14 @@ static void xe_ttm_tt_destroy(struct ttm_device *ttm_dev, struct ttm_tt *tt)
kfree(tt);
}
+static bool xe_ttm_resource_visible(struct ttm_resource *mem)
+{
+ struct xe_ttm_vram_mgr_resource *vres =
+ to_xe_ttm_vram_mgr_resource(mem);
+
+ return vres->used_visible_size == mem->size;
+}
+
static int xe_ttm_io_mem_reserve(struct ttm_device *bdev,
struct ttm_resource *mem)
{
@@ -453,11 +461,9 @@ static int xe_ttm_io_mem_reserve(struct ttm_device *bdev,
return 0;
case XE_PL_VRAM0:
case XE_PL_VRAM1: {
- struct xe_ttm_vram_mgr_resource *vres =
- to_xe_ttm_vram_mgr_resource(mem);
struct xe_mem_region *vram = res_to_mem_region(mem);
- if (vres->used_visible_size < mem->size)
+ if (!xe_ttm_resource_visible(mem))
return -EINVAL;
mem->bus.offset = mem->start << PAGE_SHIFT;
@@ -1111,6 +1117,50 @@ static void xe_ttm_bo_swap_notify(struct ttm_buffer_object *ttm_bo)
}
}
+static int xe_ttm_access_memory(struct ttm_buffer_object *ttm_bo,
+ unsigned long offset, void *buf, int len,
+ int write)
+{
+ struct xe_bo *bo = ttm_to_xe_bo(ttm_bo);
+ struct xe_device *xe = ttm_to_xe_device(ttm_bo->bdev);
+ struct iosys_map vmap;
+ struct xe_res_cursor cursor;
+ struct xe_mem_region *vram;
+ int bytes_left = len;
+
+ xe_bo_assert_held(bo);
+
+ if (!mem_type_is_vram(ttm_bo->resource->mem_type))
+ return -EIO;
+
+ /* FIXME: Use GPU for non-visible VRAM */
+ if (!xe_ttm_resource_visible(ttm_bo->resource))
+ return -EIO;
+
+ vram = res_to_mem_region(ttm_bo->resource);
+ xe_res_first(ttm_bo->resource, offset & PAGE_MASK, bo->size, &cursor);
+
+ do {
+ unsigned long page_offset = (offset & ~PAGE_MASK);
+ int byte_count = min((int)(PAGE_SIZE - page_offset), bytes_left);
+
+ iosys_map_set_vaddr_iomem(&vmap, (u8 __iomem *)vram->mapping +
+ cursor.start);
+ if (write)
+ xe_map_memcpy_to(xe, &vmap, page_offset, buf, byte_count);
+ else
+ xe_map_memcpy_from(xe, buf, &vmap, page_offset, byte_count);
+
+ offset += byte_count;
+ buf += byte_count;
+ bytes_left -= byte_count;
+ if (bytes_left)
+ xe_res_next(&cursor, PAGE_SIZE);
+ } while (bytes_left);
+
+ return len;
+}
+
const struct ttm_device_funcs xe_ttm_funcs = {
.ttm_tt_create = xe_ttm_tt_create,
.ttm_tt_populate = xe_ttm_tt_populate,
@@ -1120,6 +1170,7 @@ const struct ttm_device_funcs xe_ttm_funcs = {
.move = xe_bo_move,
.io_mem_reserve = xe_ttm_io_mem_reserve,
.io_mem_pfn = xe_ttm_io_mem_pfn,
+ .access_memory = xe_ttm_access_memory,
.release_notify = xe_ttm_bo_release_notify,
.eviction_valuable = ttm_bo_eviction_valuable,
.delete_mem_notify = xe_ttm_bo_delete_mem_notify,
--
2.34.1
^ permalink raw reply related [flat|nested] 4+ messages in thread
* [PATCH 3/3] drm/xe: Use ttm_bo_access in xe_vm_snapshot_capture_delayed
2024-10-18 21:11 [PATCH 0/3] Fix non-contiguous VRAM BO access in Xe Matthew Brost
2024-10-18 21:11 ` [PATCH 1/3] drm/ttm: Add ttm_bo_access Matthew Brost
2024-10-18 21:11 ` [PATCH 2/3] drm/xe: Add xe_ttm_access_memory Matthew Brost
@ 2024-10-18 21:11 ` Matthew Brost
2 siblings, 0 replies; 4+ messages in thread
From: Matthew Brost @ 2024-10-18 21:11 UTC (permalink / raw)
To: intel-xe, dri-devel; +Cc: matthew.auld, thomas.hellstrom
Non-contiguous mapping of BO in VRAM doesn't work, use ttm_bo_access
instead.
Fixes: 0eb2a18a8fad ("drm/xe: Implement VM snapshot support for BO's and userptr")
Suggested-by: Matthew Auld <matthew.auld@intel.com>
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
---
drivers/gpu/drm/xe/xe_vm.c | 17 ++++++-----------
1 file changed, 6 insertions(+), 11 deletions(-)
diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
index c99380271de6..2eae9ce41c06 100644
--- a/drivers/gpu/drm/xe/xe_vm.c
+++ b/drivers/gpu/drm/xe/xe_vm.c
@@ -3303,7 +3303,6 @@ void xe_vm_snapshot_capture_delayed(struct xe_vm_snapshot *snap)
for (int i = 0; i < snap->num_snaps; i++) {
struct xe_bo *bo = snap->snap[i].bo;
- struct iosys_map src;
int err;
if (IS_ERR(snap->snap[i].data))
@@ -3316,16 +3315,12 @@ void xe_vm_snapshot_capture_delayed(struct xe_vm_snapshot *snap)
}
if (bo) {
- xe_bo_lock(bo, false);
- err = ttm_bo_vmap(&bo->ttm, &src);
- if (!err) {
- xe_map_memcpy_from(xe_bo_device(bo),
- snap->snap[i].data,
- &src, snap->snap[i].bo_ofs,
- snap->snap[i].len);
- ttm_bo_vunmap(&bo->ttm, &src);
- }
- xe_bo_unlock(bo);
+ err = ttm_bo_access(&bo->ttm, snap->snap[i].bo_ofs,
+ snap->snap[i].data, snap->snap[i].len, 0);
+ if (!(err < 0) && err != snap->snap[i].len)
+ err = -EIO;
+ else
+ err = 0;
} else {
void __user *userptr = (void __user *)(size_t)snap->snap[i].bo_ofs;
--
2.34.1
^ permalink raw reply related [flat|nested] 4+ messages in thread
end of thread, other threads:[~2024-10-18 21:10 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-10-18 21:11 [PATCH 0/3] Fix non-contiguous VRAM BO access in Xe Matthew Brost
2024-10-18 21:11 ` [PATCH 1/3] drm/ttm: Add ttm_bo_access Matthew Brost
2024-10-18 21:11 ` [PATCH 2/3] drm/xe: Add xe_ttm_access_memory Matthew Brost
2024-10-18 21:11 ` [PATCH 3/3] drm/xe: Use ttm_bo_access in xe_vm_snapshot_capture_delayed Matthew Brost
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox