From: Oak Zeng <oak.zeng@intel.com>
To: intel-xe@lists.freedesktop.org
Subject: [CI 41/44] drm/xe/svm: Migration from sram to vram for system allocator
Date: Fri, 14 Jun 2024 17:58:14 -0400 [thread overview]
Message-ID: <20240614215817.1097633-41-oak.zeng@intel.com> (raw)
In-Reply-To: <20240614215817.1097633-1-oak.zeng@intel.com>
If applicable, migrate a range of hmmptr from sram to vram for
system allocator. Traditional userptr is not migrated. Only userptr
created during fault (aka userptr split from system allocator vma,
aka fault userptr in codes) can be migrated.
Instead of the whole userptr, only a sub-range of userptr is populated
and bind to GPU page table. Right now the range granularity is 2MiB,
which will be overwritten by memory attribute API.
FIXME: The migration should be conditional on user memory attributes
setting. Add this logic when memory attributes are supported
v1: Use none NULL owner when call drm_svm_hmmptr_populate (Himal)
Cc: Thomas Hellström <thomas.hellstrom@linux.intel.com>
Cc: Matthew Brost <matthew.brost@intel.com>
Cc: Brian Welty <brian.welty@intel.com>
Signed-off-by: Himal Prasad Ghimiray <himal.prasad.ghimiray@intel.com>
Signed-off-by: Oak Zeng <oak.zeng@intel.com>
---
drivers/gpu/drm/xe/xe_gt_pagefault.c | 42 +++++++++++++++++++++++++---
drivers/gpu/drm/xe/xe_vm.c | 18 +++++++-----
drivers/gpu/drm/xe/xe_vm.h | 2 +-
3 files changed, 50 insertions(+), 12 deletions(-)
diff --git a/drivers/gpu/drm/xe/xe_gt_pagefault.c b/drivers/gpu/drm/xe/xe_gt_pagefault.c
index 93efee125fc0..f2a4abcef3b1 100644
--- a/drivers/gpu/drm/xe/xe_gt_pagefault.c
+++ b/drivers/gpu/drm/xe/xe_gt_pagefault.c
@@ -128,6 +128,12 @@ static int handle_vma_pagefault(struct xe_tile *tile, struct pagefault *pf,
ktime_t end = 0;
int err;
bool atomic;
+ /**FIXME: use migration granularity from memory attributes*/
+ u64 migrate_granularity = SZ_2M;
+ u64 fault_addr = pf->page_addr;
+ u64 fault_offset, fault_cpu_addr;
+ u64 aligned_cpu_fault_start, aligned_cpu_fault_end;
+ u64 cpu_va_start, cpu_va_end, gpu_va_start, gpu_va_end;
trace_xe_vma_pagefault(vma);
atomic = access_is_atomic(pf->access_type);
@@ -144,13 +150,41 @@ static int handle_vma_pagefault(struct xe_tile *tile, struct pagefault *pf,
}
}
+ fault_offset = fault_addr - xe_vma_start(vma);
+ fault_cpu_addr = xe_vma_userptr(vma) + fault_offset;
+ aligned_cpu_fault_start = ALIGN_DOWN(fault_cpu_addr, migrate_granularity);
+ aligned_cpu_fault_end = aligned_cpu_fault_start + migrate_granularity;
+
+ if (xe_vma_is_userptr(vma)) {
+ cpu_va_start = max_t(u64, xe_vma_userptr(vma), aligned_cpu_fault_start);
+ cpu_va_end = min_t(u64, xe_vma_userptr_end(vma), aligned_cpu_fault_end);
+ gpu_va_start = xe_vma_start(vma) + (cpu_va_start - xe_vma_userptr(vma));
+ gpu_va_end = xe_vma_end(vma) - (xe_vma_userptr_end(vma) - cpu_va_end);
+ } else {
+ gpu_va_start = xe_vma_start(vma);
+ gpu_va_end = xe_vma_end(vma);
+ }
+
retry_userptr:
xe_vm_userptr_garbage_collector(vm);
if (xe_vma_is_userptr(vma) &&
xe_vma_userptr_check_repin(to_userptr_vma(vma))) {
struct xe_userptr_vma *uvma = to_userptr_vma(vma);
-
- err = xe_vma_userptr_pin_pages(uvma);
+ struct xe_userptr *userptr = &uvma->userptr;
+ struct drm_hmmptr *hmmptr = &userptr->hmmptr;
+
+ mmap_read_lock(hmmptr->notifier.mm);
+ if (xe_vma_is_fault_userptr(vma)) {
+ /**FIXME: add migration policy here*/
+ err = drm_svm_migrate_hmmptr_to_vram(&vm->gpuvm, &tile->mem.vram.drm_mr,
+ hmmptr, cpu_va_start, cpu_va_end);
+ if (err) {
+ mmap_read_unlock(hmmptr->notifier.mm);
+ return err;
+ }
+ }
+ err = xe_vma_userptr_pin_pages(uvma, cpu_va_start, cpu_va_end, true);
+ mmap_read_unlock(hmmptr->notifier.mm);
if (err)
return err;
}
@@ -167,8 +201,8 @@ static int handle_vma_pagefault(struct xe_tile *tile, struct pagefault *pf,
/* Bind VMA only to the GT that has faulted */
trace_xe_vma_pf_bind(vma);
- fence = xe_vma_rebind(vm, vma, xe_vma_start(vma),
- xe_vma_end(vma), BIT(tile->id));
+ fence = xe_vma_rebind(vm, vma, gpu_va_start,
+ gpu_va_end, BIT(tile->id));
if (IS_ERR(fence)) {
err = PTR_ERR(fence);
if (xe_vm_validate_should_retry(&exec, err, &end))
diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
index 418420918bd3..29eb561de326 100644
--- a/drivers/gpu/drm/xe/xe_vm.c
+++ b/drivers/gpu/drm/xe/xe_vm.c
@@ -76,6 +76,7 @@ static void xe_vma_userptr_dma_map_pages(struct xe_userptr_vma *uvma,
xe_assert(xe, xe_vma_is_userptr(vma));
xe_assert(xe, start >= xe_vma_userptr(vma));
xe_assert(xe, end <= xe_vma_userptr_end(vma));
+
drm_svm_hmmptr_map_dma_pages(hmmptr, page_idx, npages);
}
@@ -96,7 +97,7 @@ static void xe_vma_userptr_dma_unmap_pages(struct xe_userptr_vma *uvma,
drm_svm_hmmptr_unmap_dma_pages(hmmptr, page_idx, npages);
}
-int xe_vma_userptr_pin_pages(struct xe_userptr_vma *uvma)
+int xe_vma_userptr_pin_pages(struct xe_userptr_vma *uvma, u64 start, u64 end, bool mmap_locked)
{
struct drm_hmmptr *hmmptr = &uvma->userptr.hmmptr;
struct xe_vma *vma = &uvma->vma;
@@ -106,14 +107,15 @@ int xe_vma_userptr_pin_pages(struct xe_userptr_vma *uvma)
lockdep_assert_held(&vm->lock);
xe_assert(xe, xe_vma_is_userptr(vma));
+ xe_assert(xe, start >= xe_vma_userptr(vma));
+ xe_assert(xe, end <= xe_vma_userptr_end(vma));
- ret = drm_svm_hmmptr_populate(hmmptr, NULL, xe_vma_userptr(vma),
- xe_vma_userptr(vma) + xe_vma_size(vma),
- !xe_vma_read_only(vma), false);
+ ret = drm_svm_hmmptr_populate(hmmptr, xe, start, end,
+ !xe_vma_read_only(vma), mmap_locked);
if (ret)
return ret;
- xe_vma_userptr_dma_map_pages(uvma, xe_vma_userptr(vma), xe_vma_userptr_end(vma));
+ xe_vma_userptr_dma_map_pages(uvma, start, end);
return 0;
}
@@ -736,7 +738,8 @@ int xe_vm_userptr_pin(struct xe_vm *vm)
/* Pin and move to temporary list */
list_for_each_entry_safe(uvma, next, &vm->userptr.repin_list,
userptr.repin_link) {
- err = xe_vma_userptr_pin_pages(uvma);
+ err = xe_vma_userptr_pin_pages(uvma, xe_vma_userptr(&uvma->vma),
+ xe_vma_userptr_end(&uvma->vma), false);
if (err == -EFAULT) {
list_del_init(&uvma->userptr.repin_link);
@@ -2116,7 +2119,8 @@ static struct xe_vma *new_vma(struct xe_vm *vm, struct drm_gpuva_op_map *op,
drm_exec_fini(&exec);
if (xe_vma_is_userptr(vma) && !xe_vma_is_fault_userptr(vma)) {
- err = xe_vma_userptr_pin_pages(to_userptr_vma(vma));
+ err = xe_vma_userptr_pin_pages(to_userptr_vma(vma), xe_vma_userptr(vma),
+ xe_vma_userptr_end(vma), false);
if (err) {
prep_vma_destroy(vm, vma, false);
xe_vma_destroy_unlocked(vma);
diff --git a/drivers/gpu/drm/xe/xe_vm.h b/drivers/gpu/drm/xe/xe_vm.h
index b3b6ceec39ba..f24891cb1fcb 100644
--- a/drivers/gpu/drm/xe/xe_vm.h
+++ b/drivers/gpu/drm/xe/xe_vm.h
@@ -253,7 +253,7 @@ static inline void xe_vm_reactivate_rebind(struct xe_vm *vm)
}
}
-int xe_vma_userptr_pin_pages(struct xe_userptr_vma *uvma);
+int xe_vma_userptr_pin_pages(struct xe_userptr_vma *uvma, u64 start, u64 end, bool mmap_locked);
int xe_vma_userptr_check_repin(struct xe_userptr_vma *uvma);
--
2.26.3
next prev parent reply other threads:[~2024-06-14 21:49 UTC|newest]
Thread overview: 60+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-06-14 21:57 [CI 01/44] mm/hmm: let users to tag specific PFNs Oak Zeng
2024-06-14 21:57 ` [CI 02/44] dma-mapping: provide an interface to allocate IOVA Oak Zeng
2024-06-14 21:57 ` [CI 03/44] dma-mapping: provide callbacks to link/unlink pages to specific IOVA Oak Zeng
2024-06-14 21:57 ` [CI 04/44] iommu/dma: Provide an interface to allow preallocate IOVA Oak Zeng
2024-06-14 21:57 ` [CI 05/44] iommu/dma: Prepare map/unmap page functions to receive IOVA Oak Zeng
2024-06-14 21:57 ` [CI 06/44] iommu/dma: Implement link/unlink page callbacks Oak Zeng
2024-06-14 21:57 ` [CI 07/44] mm: Add an empty implementation of zone_device_page_init Oak Zeng
2024-06-14 21:57 ` [CI 08/44] drm: Move GPUVA_START/LAST to drm_gpuvm.h Oak Zeng
2024-06-14 21:57 ` [CI 09/44] drm/svm: Mark drm_gpuvm to participate SVM Oak Zeng
2024-06-14 21:57 ` [CI 10/44] drm/svm: introduce drm_mem_region concept Oak Zeng
2024-06-14 21:57 ` [CI 11/44] drm/svm: introduce hmmptr and helper functions Oak Zeng
2024-06-14 21:57 ` [CI 12/44] drm/svm: Introduce helper to remap drm memory region Oak Zeng
2024-06-14 21:57 ` [CI 13/44] drm/svm: handle CPU page fault Oak Zeng
2024-06-14 21:57 ` [CI 14/44] drm/svm: Migrate a range of hmmptr to vram Oak Zeng
2024-06-14 21:57 ` [CI 15/44] drm/svm: Add DRM SVM documentation Oak Zeng
2024-06-14 21:57 ` [CI 16/44] drm/svm: Introduce DRM_SVM kernel config Oak Zeng
2024-06-14 21:57 ` [CI 17/44] drm/xe: s/xe_tile_migrate_engine/xe_tile_migrate_exec_queue Oak Zeng
2024-06-14 21:57 ` [CI 18/44] drm/xe: Add xe_vm_pgtable_update_op to xe_vma_ops Oak Zeng
2024-06-14 21:57 ` [CI 19/44] drm/xe: Convert multiple bind ops into single job Oak Zeng
2024-06-14 21:57 ` [CI 20/44] drm/xe: Update VM trace events Oak Zeng
2024-06-14 21:57 ` [CI 21/44] drm/xe: Update PT layer with better error handling Oak Zeng
2024-06-14 21:57 ` [CI 22/44] drm/xe: Retry BO allocation Oak Zeng
2024-06-14 21:57 ` [CI 23/44] drm/xe/uapi: Add DRM_XE_VM_BIND_FLAG_SYSTEM_ALLOCATOR flag Oak Zeng
2024-06-14 21:57 ` [CI 24/44] drm/xe: Add a helper to calculate userptr end address Oak Zeng
2024-06-14 21:57 ` [CI 25/44] drm/xe: Add dma_addr res cursor Oak Zeng
2024-06-14 21:57 ` [CI 26/44] drm/xe: Use drm_mem_region for xe Oak Zeng
2024-06-14 21:58 ` [CI 27/44] drm/xe: use drm_hmmptr in xe Oak Zeng
2024-06-14 21:58 ` [CI 28/44] drm/xe: Moving to range based vma invalidation Oak Zeng
2024-06-14 21:58 ` [CI 29/44] drm/xe: Support range based page table update Oak Zeng
2024-06-14 21:58 ` [CI 30/44] drm/xe/uapi: Add DRM_XE_VM_CREATE_FLAG_PARTICIPATE_SVM flag Oak Zeng
2024-06-14 21:58 ` [CI 31/44] drm/xe/svm: Create userptr if page fault occurs on system_allocator VMA Oak Zeng
2024-06-14 21:58 ` [CI 32/44] drm/xe/svm: Add faulted userptr VMA garbage collector Oak Zeng
2024-06-14 21:58 ` [CI 33/44] drm/xe: Introduce helper to get tile from memory region Oak Zeng
2024-06-14 21:58 ` [CI 34/44] drm/xe/svm: implement functions to allocate and free device memory Oak Zeng
2024-06-14 21:58 ` [CI 35/44] drm/xe/svm: Get drm device from drm memory region Oak Zeng
2024-06-14 21:58 ` [CI 36/44] drm/xe/svm: Get page map owner of a " Oak Zeng
2024-06-14 21:58 ` [CI 37/44] drm/xe/svm: Add migrate layer functions for SVM support Oak Zeng
2024-06-14 21:58 ` [CI 38/44] drm/xe/svm: introduce svm migration function Oak Zeng
2024-06-14 21:58 ` [CI 39/44] drm/xe/svm: Register xe memory region to drm layer Oak Zeng
2024-06-14 21:58 ` [CI 40/44] drm/xe/svm: Introduce DRM_XE_SVM kernel config Oak Zeng
2024-06-14 21:58 ` Oak Zeng [this message]
2024-06-14 21:58 ` [CI 42/44] drm/xe/svm: Determine a vma is backed by device memory Oak Zeng
2024-06-14 21:58 ` [CI 43/44] drm/xe/svm: Introduce hmm_pfn array based resource cursor Oak Zeng
2024-06-14 21:58 ` [CI 44/44] drm/xe: Enable system allocator uAPI Oak Zeng
2024-06-14 22:41 ` ✓ CI.Patch_applied: success for series starting with [CI,01/44] mm/hmm: let users to tag specific PFNs Patchwork
2024-06-14 22:42 ` ✗ CI.checkpatch: warning " Patchwork
2024-06-14 22:43 ` ✓ CI.KUnit: success " Patchwork
2024-06-14 22:55 ` ✓ CI.Build: " Patchwork
2024-06-14 22:57 ` ✗ CI.Hooks: failure " Patchwork
2024-06-14 22:58 ` ✗ CI.checksparse: warning " Patchwork
2024-06-14 23:21 ` ✓ CI.BAT: success " Patchwork
2024-06-15 2:56 ` ✓ CI.Patch_applied: success for series starting with [CI,01/44] mm/hmm: let users to tag specific PFNs (rev2) Patchwork
2024-06-15 2:57 ` ✗ CI.checkpatch: warning " Patchwork
2024-06-15 2:58 ` ✓ CI.KUnit: success " Patchwork
2024-06-15 3:10 ` ✓ CI.Build: " Patchwork
2024-06-15 3:12 ` ✗ CI.Hooks: failure " Patchwork
2024-06-15 3:14 ` ✗ CI.checksparse: warning " Patchwork
2024-06-15 3:36 ` ✓ CI.BAT: success " Patchwork
2024-06-15 13:59 ` ✗ CI.FULL: failure for series starting with [CI,01/44] mm/hmm: let users to tag specific PFNs Patchwork
2024-06-18 8:53 ` ✓ CI.FULL: success for series starting with [CI,01/44] mm/hmm: let users to tag specific PFNs (rev2) Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20240614215817.1097633-41-oak.zeng@intel.com \
--to=oak.zeng@intel.com \
--cc=intel-xe@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox