From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Date: Tue, 19 Sep 2023 15:31:34 +0000 From: Matthew Brost To: Rodrigo Vivi Message-ID: References: <20230919142000.91363-1-rodrigo.vivi@intel.com> <20230919142000.91363-4-rodrigo.vivi@intel.com> Content-Type: text/plain; charset="us-ascii" Content-Disposition: inline In-Reply-To: <20230919142000.91363-4-rodrigo.vivi@intel.com> MIME-Version: 1.0 Subject: Re: [igt-dev] [PATCH i-g-t 03/16] drm-uapi/xe_drm: Separate VM_BIND's operation and flag, align with latest uapi List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: igt-dev@lists.freedesktop.org, intel-xe@lists.freedesktop.org Errors-To: igt-dev-bounces@lists.freedesktop.org Sender: "igt-dev" List-ID: On Tue, Sep 19, 2023 at 10:19:46AM -0400, Rodrigo Vivi wrote: > From: Francois Dugast > > Align with commit ("drm/xe/uapi: Separate VM_BIND's operation and flag") > > Signed-off-by: Francois Dugast > Signed-off-by: Rodrigo Vivi Reviewed-by: Matthew Brost > --- > include/drm-uapi/xe_drm.h | 14 ++++++++------ > lib/intel_batchbuffer.c | 11 +++++++---- > lib/xe/xe_ioctl.c | 31 ++++++++++++++++--------------- > lib/xe/xe_ioctl.h | 6 +++--- > lib/xe/xe_util.c | 9 ++++++--- > tests/intel/xe_exec_basic.c | 2 +- > tests/intel/xe_exec_threads.c | 2 +- > tests/intel/xe_vm.c | 16 +++++++++------- > 8 files changed, 51 insertions(+), 40 deletions(-) > > diff --git a/include/drm-uapi/xe_drm.h b/include/drm-uapi/xe_drm.h > index f96c84a98..078edd9f8 100644 > --- a/include/drm-uapi/xe_drm.h > +++ b/include/drm-uapi/xe_drm.h > @@ -663,8 +663,10 @@ struct drm_xe_vm_bind_op { > #define XE_VM_BIND_OP_RESTART 0x3 > #define XE_VM_BIND_OP_UNMAP_ALL 0x4 > #define XE_VM_BIND_OP_PREFETCH 0x5 > + /** @op: Bind operation to perform */ > + __u32 op; > > -#define XE_VM_BIND_FLAG_READONLY (0x1 << 16) > +#define XE_VM_BIND_FLAG_READONLY (0x1 << 0) > /* > * A bind ops completions are always async, hence the support for out > * sync. This flag indicates the allocation of the memory for new page > @@ -689,12 +691,12 @@ struct drm_xe_vm_bind_op { > * configured in the VM and must be set if the VM is configured with > * DRM_XE_VM_CREATE_ASYNC_BIND_OPS and not in an error state. > */ > -#define XE_VM_BIND_FLAG_ASYNC (0x1 << 17) > +#define XE_VM_BIND_FLAG_ASYNC (0x1 << 1) > /* > * Valid on a faulting VM only, do the MAP operation immediately rather > * than deferring the MAP to the page fault handler. > */ > -#define XE_VM_BIND_FLAG_IMMEDIATE (0x1 << 18) > +#define XE_VM_BIND_FLAG_IMMEDIATE (0x1 << 2) > /* > * When the NULL flag is set, the page tables are setup with a special > * bit which indicates writes are dropped and all reads return zero. In > @@ -702,9 +704,9 @@ struct drm_xe_vm_bind_op { > * operations, the BO handle MBZ, and the BO offset MBZ. This flag is > * intended to implement VK sparse bindings. > */ > -#define XE_VM_BIND_FLAG_NULL (0x1 << 19) > - /** @op: Operation to perform (lower 16 bits) and flags (upper 16 bits) */ > - __u32 op; > +#define XE_VM_BIND_FLAG_NULL (0x1 << 3) > + /** @flags: Bind flags */ > + __u32 flags; > > /** @mem_region: Memory region to prefetch VMA to, instance not a mask */ > __u32 region; > diff --git a/lib/intel_batchbuffer.c b/lib/intel_batchbuffer.c > index e7b1b755f..6e668d28c 100644 > --- a/lib/intel_batchbuffer.c > +++ b/lib/intel_batchbuffer.c > @@ -1281,7 +1281,8 @@ void intel_bb_destroy(struct intel_bb *ibb) > } > > static struct drm_xe_vm_bind_op *xe_alloc_bind_ops(struct intel_bb *ibb, > - uint32_t op, uint32_t region) > + uint32_t op, uint32_t flags, > + uint32_t region) > { > struct drm_i915_gem_exec_object2 **objects = ibb->objects; > struct drm_xe_vm_bind_op *bind_ops, *ops; > @@ -1298,6 +1299,7 @@ static struct drm_xe_vm_bind_op *xe_alloc_bind_ops(struct intel_bb *ibb, > ops->obj = objects[i]->handle; > > ops->op = op; > + ops->flags = flags; > ops->obj_offset = 0; > ops->addr = objects[i]->offset; > ops->range = objects[i]->rsvd1; > @@ -1323,9 +1325,10 @@ static void __unbind_xe_objects(struct intel_bb *ibb) > > if (ibb->num_objects > 1) { > struct drm_xe_vm_bind_op *bind_ops; > - uint32_t op = XE_VM_BIND_OP_UNMAP | XE_VM_BIND_FLAG_ASYNC; > + uint32_t op = XE_VM_BIND_OP_UNMAP; > + uint32_t flags = XE_VM_BIND_FLAG_ASYNC; > > - bind_ops = xe_alloc_bind_ops(ibb, op, 0); > + bind_ops = xe_alloc_bind_ops(ibb, op, flags, 0); > xe_vm_bind_array(ibb->fd, ibb->vm_id, 0, bind_ops, > ibb->num_objects, syncs, 2); > free(bind_ops); > @@ -2354,7 +2357,7 @@ __xe_bb_exec(struct intel_bb *ibb, uint64_t flags, bool sync) > > syncs[0].handle = syncobj_create(ibb->fd, 0); > if (ibb->num_objects > 1) { > - bind_ops = xe_alloc_bind_ops(ibb, XE_VM_BIND_OP_MAP | XE_VM_BIND_FLAG_ASYNC, 0); > + bind_ops = xe_alloc_bind_ops(ibb, XE_VM_BIND_OP_MAP, XE_VM_BIND_FLAG_ASYNC, 0); > xe_vm_bind_array(ibb->fd, ibb->vm_id, 0, bind_ops, > ibb->num_objects, syncs, 1); > free(bind_ops); > diff --git a/lib/xe/xe_ioctl.c b/lib/xe/xe_ioctl.c > index 730dcfd16..48cd185de 100644 > --- a/lib/xe/xe_ioctl.c > +++ b/lib/xe/xe_ioctl.c > @@ -67,7 +67,7 @@ void xe_vm_unbind_all_async(int fd, uint32_t vm, uint32_t exec_queue, > uint32_t num_syncs) > { > __xe_vm_bind_assert(fd, vm, exec_queue, bo, 0, 0, 0, > - XE_VM_BIND_OP_UNMAP_ALL | XE_VM_BIND_FLAG_ASYNC, > + XE_VM_BIND_OP_UNMAP_ALL, XE_VM_BIND_FLAG_ASYNC, > sync, num_syncs, 0, 0); > } > > @@ -91,8 +91,8 @@ void xe_vm_bind_array(int fd, uint32_t vm, uint32_t exec_queue, > > int __xe_vm_bind(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > uint64_t offset, uint64_t addr, uint64_t size, uint32_t op, > - struct drm_xe_sync *sync, uint32_t num_syncs, uint32_t region, > - uint64_t ext) > + uint32_t flags, struct drm_xe_sync *sync, uint32_t num_syncs, > + uint32_t region, uint64_t ext) > { > struct drm_xe_vm_bind bind = { > .extensions = ext, > @@ -103,6 +103,7 @@ int __xe_vm_bind(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > .bind.range = size, > .bind.addr = addr, > .bind.op = op, > + .bind.flags = flags, > .bind.region = region, > .num_syncs = num_syncs, > .syncs = (uintptr_t)sync, > @@ -117,11 +118,11 @@ int __xe_vm_bind(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > > void __xe_vm_bind_assert(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > uint64_t offset, uint64_t addr, uint64_t size, > - uint32_t op, struct drm_xe_sync *sync, > + uint32_t op, uint32_t flags, struct drm_xe_sync *sync, > uint32_t num_syncs, uint32_t region, uint64_t ext) > { > igt_assert_eq(__xe_vm_bind(fd, vm, exec_queue, bo, offset, addr, size, > - op, sync, num_syncs, region, ext), 0); > + op, flags, sync, num_syncs, region, ext), 0); > } > > void xe_vm_bind(int fd, uint32_t vm, uint32_t bo, uint64_t offset, > @@ -129,7 +130,7 @@ void xe_vm_bind(int fd, uint32_t vm, uint32_t bo, uint64_t offset, > struct drm_xe_sync *sync, uint32_t num_syncs) > { > __xe_vm_bind_assert(fd, vm, 0, bo, offset, addr, size, > - XE_VM_BIND_OP_MAP, sync, num_syncs, 0, 0); > + XE_VM_BIND_OP_MAP, 0, sync, num_syncs, 0, 0); > } > > void xe_vm_unbind(int fd, uint32_t vm, uint64_t offset, > @@ -137,7 +138,7 @@ void xe_vm_unbind(int fd, uint32_t vm, uint64_t offset, > struct drm_xe_sync *sync, uint32_t num_syncs) > { > __xe_vm_bind_assert(fd, vm, 0, 0, offset, addr, size, > - XE_VM_BIND_OP_UNMAP, sync, num_syncs, 0, 0); > + XE_VM_BIND_OP_UNMAP, 0, sync, num_syncs, 0, 0); > } > > void xe_vm_prefetch_async(int fd, uint32_t vm, uint32_t exec_queue, uint64_t offset, > @@ -146,7 +147,7 @@ void xe_vm_prefetch_async(int fd, uint32_t vm, uint32_t exec_queue, uint64_t off > uint32_t region) > { > __xe_vm_bind_assert(fd, vm, exec_queue, 0, offset, addr, size, > - XE_VM_BIND_OP_PREFETCH | XE_VM_BIND_FLAG_ASYNC, > + XE_VM_BIND_OP_PREFETCH, XE_VM_BIND_FLAG_ASYNC, > sync, num_syncs, region, 0); > } > > @@ -155,7 +156,7 @@ void xe_vm_bind_async(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > struct drm_xe_sync *sync, uint32_t num_syncs) > { > __xe_vm_bind_assert(fd, vm, exec_queue, bo, offset, addr, size, > - XE_VM_BIND_OP_MAP | XE_VM_BIND_FLAG_ASYNC, sync, > + XE_VM_BIND_OP_MAP, XE_VM_BIND_FLAG_ASYNC, sync, > num_syncs, 0, 0); > } > > @@ -165,7 +166,7 @@ void xe_vm_bind_async_flags(int fd, uint32_t vm, uint32_t exec_queue, uint32_t b > uint32_t flags) > { > __xe_vm_bind_assert(fd, vm, exec_queue, bo, offset, addr, size, > - XE_VM_BIND_OP_MAP | XE_VM_BIND_FLAG_ASYNC | flags, > + XE_VM_BIND_OP_MAP, XE_VM_BIND_FLAG_ASYNC | flags, > sync, num_syncs, 0, 0); > } > > @@ -174,7 +175,7 @@ void xe_vm_bind_userptr_async(int fd, uint32_t vm, uint32_t exec_queue, > struct drm_xe_sync *sync, uint32_t num_syncs) > { > __xe_vm_bind_assert(fd, vm, exec_queue, 0, userptr, addr, size, > - XE_VM_BIND_OP_MAP_USERPTR | XE_VM_BIND_FLAG_ASYNC, > + XE_VM_BIND_OP_MAP_USERPTR, XE_VM_BIND_FLAG_ASYNC, > sync, num_syncs, 0, 0); > } > > @@ -184,7 +185,7 @@ void xe_vm_bind_userptr_async_flags(int fd, uint32_t vm, uint32_t exec_queue, > uint32_t num_syncs, uint32_t flags) > { > __xe_vm_bind_assert(fd, vm, exec_queue, 0, userptr, addr, size, > - XE_VM_BIND_OP_MAP_USERPTR | XE_VM_BIND_FLAG_ASYNC | > + XE_VM_BIND_OP_MAP_USERPTR, XE_VM_BIND_FLAG_ASYNC | > flags, sync, num_syncs, 0, 0); > } > > @@ -193,7 +194,7 @@ void xe_vm_unbind_async(int fd, uint32_t vm, uint32_t exec_queue, > struct drm_xe_sync *sync, uint32_t num_syncs) > { > __xe_vm_bind_assert(fd, vm, exec_queue, 0, offset, addr, size, > - XE_VM_BIND_OP_UNMAP | XE_VM_BIND_FLAG_ASYNC, sync, > + XE_VM_BIND_OP_UNMAP, XE_VM_BIND_FLAG_ASYNC, sync, > num_syncs, 0, 0); > } > > @@ -205,8 +206,8 @@ static void __xe_vm_bind_sync(int fd, uint32_t vm, uint32_t bo, uint64_t offset, > .handle = syncobj_create(fd, 0), > }; > > - __xe_vm_bind_assert(fd, vm, 0, bo, offset, addr, size, op, &sync, 1, 0, > - 0); > + __xe_vm_bind_assert(fd, vm, 0, bo, offset, addr, size, op, 0, &sync, 1, > + 0, 0); > > igt_assert(syncobj_wait(fd, &sync.handle, 1, INT64_MAX, 0, NULL)); > syncobj_destroy(fd, sync.handle); > diff --git a/lib/xe/xe_ioctl.h b/lib/xe/xe_ioctl.h > index 6c281b3bf..f0e4109dc 100644 > --- a/lib/xe/xe_ioctl.h > +++ b/lib/xe/xe_ioctl.h > @@ -19,11 +19,11 @@ uint32_t xe_cs_prefetch_size(int fd); > uint32_t xe_vm_create(int fd, uint32_t flags, uint64_t ext); > int __xe_vm_bind(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > uint64_t offset, uint64_t addr, uint64_t size, uint32_t op, > - struct drm_xe_sync *sync, uint32_t num_syncs, uint32_t region, > - uint64_t ext); > + uint32_t flags, struct drm_xe_sync *sync, uint32_t num_syncs, > + uint32_t region, uint64_t ext); > void __xe_vm_bind_assert(int fd, uint32_t vm, uint32_t exec_queue, uint32_t bo, > uint64_t offset, uint64_t addr, uint64_t size, > - uint32_t op, struct drm_xe_sync *sync, > + uint32_t op, uint32_t flags, struct drm_xe_sync *sync, > uint32_t num_syncs, uint32_t region, uint64_t ext); > void xe_vm_bind(int fd, uint32_t vm, uint32_t bo, uint64_t offset, > uint64_t addr, uint64_t size, > diff --git a/lib/xe/xe_util.c b/lib/xe/xe_util.c > index 2f9ffe2f1..5fa4d4610 100644 > --- a/lib/xe/xe_util.c > +++ b/lib/xe/xe_util.c > @@ -116,7 +116,7 @@ static struct drm_xe_vm_bind_op *xe_alloc_bind_ops(struct igt_list_head *obj_lis > { > struct drm_xe_vm_bind_op *bind_ops, *ops; > struct xe_object *obj; > - uint32_t num_objects = 0, i = 0, op; > + uint32_t num_objects = 0, i = 0, op, flags; > > igt_list_for_each_entry(obj, obj_list, link) > num_objects++; > @@ -134,13 +134,16 @@ static struct drm_xe_vm_bind_op *xe_alloc_bind_ops(struct igt_list_head *obj_lis > ops = &bind_ops[i]; > > if (obj->bind_op == XE_OBJECT_BIND) { > - op = XE_VM_BIND_OP_MAP | XE_VM_BIND_FLAG_ASYNC; > + op = XE_VM_BIND_OP_MAP; > + flags = XE_VM_BIND_FLAG_ASYNC; > ops->obj = obj->handle; > } else { > - op = XE_VM_BIND_OP_UNMAP | XE_VM_BIND_FLAG_ASYNC; > + op = XE_VM_BIND_OP_UNMAP; > + flags = XE_VM_BIND_FLAG_ASYNC; > } > > ops->op = op; > + ops->flags = flags; > ops->obj_offset = 0; > ops->addr = obj->offset; > ops->range = obj->size; > diff --git a/tests/intel/xe_exec_basic.c b/tests/intel/xe_exec_basic.c > index a4414e052..e29398aaa 100644 > --- a/tests/intel/xe_exec_basic.c > +++ b/tests/intel/xe_exec_basic.c > @@ -170,7 +170,7 @@ test_exec(int fd, struct drm_xe_engine_class_instance *eci, > if (flags & SPARSE) > __xe_vm_bind_assert(fd, vm[i], bind_exec_queues[i], > 0, 0, sparse_addr[i], bo_size, > - XE_VM_BIND_OP_MAP | > + XE_VM_BIND_OP_MAP, > XE_VM_BIND_FLAG_ASYNC | > XE_VM_BIND_FLAG_NULL, sync, > 1, 0, 0); > diff --git a/tests/intel/xe_exec_threads.c b/tests/intel/xe_exec_threads.c > index 12e76874e..1f9af894f 100644 > --- a/tests/intel/xe_exec_threads.c > +++ b/tests/intel/xe_exec_threads.c > @@ -609,7 +609,7 @@ test_legacy_mode(int fd, uint32_t vm, uint64_t addr, uint64_t userptr, > if (rebind_error_inject == i) > __xe_vm_bind_assert(fd, vm, bind_exec_queues[e], > 0, 0, addr, bo_size, > - XE_VM_BIND_OP_UNMAP | > + XE_VM_BIND_OP_UNMAP, > XE_VM_BIND_FLAG_ASYNC | > INJECT_ERROR, sync_all, > n_exec_queues, 0, 0); > diff --git a/tests/intel/xe_vm.c b/tests/intel/xe_vm.c > index 4952ea786..f96305851 100644 > --- a/tests/intel/xe_vm.c > +++ b/tests/intel/xe_vm.c > @@ -316,7 +316,7 @@ static void userptr_invalid(int fd) > vm = xe_vm_create(fd, 0, 0); > munmap(data, size); > ret = __xe_vm_bind(fd, vm, 0, 0, to_user_pointer(data), 0x40000, > - size, XE_VM_BIND_OP_MAP_USERPTR, NULL, 0, 0, 0); > + size, XE_VM_BIND_OP_MAP_USERPTR, 0, NULL, 0, 0, 0); > igt_assert(ret == -EFAULT); > > xe_vm_destroy(fd, vm); > @@ -437,7 +437,7 @@ static void vm_async_ops_err(int fd, bool destroy) > if (i == N_BINDS / 8) /* Inject error on this bind */ > __xe_vm_bind_assert(fd, vm, 0, bo, 0, > addr + i * bo_size * 2, > - bo_size, XE_VM_BIND_OP_MAP | > + bo_size, XE_VM_BIND_OP_MAP, > XE_VM_BIND_FLAG_ASYNC | > INJECT_ERROR, &sync, 1, 0, 0); > else > @@ -451,7 +451,7 @@ static void vm_async_ops_err(int fd, bool destroy) > if (i == N_BINDS / 8) > __xe_vm_bind_assert(fd, vm, 0, 0, 0, > addr + i * bo_size * 2, > - bo_size, XE_VM_BIND_OP_UNMAP | > + bo_size, XE_VM_BIND_OP_UNMAP, > XE_VM_BIND_FLAG_ASYNC | > INJECT_ERROR, &sync, 1, 0, 0); > else > @@ -465,7 +465,7 @@ static void vm_async_ops_err(int fd, bool destroy) > if (i == N_BINDS / 8) > __xe_vm_bind_assert(fd, vm, 0, bo, 0, > addr + i * bo_size * 2, > - bo_size, XE_VM_BIND_OP_MAP | > + bo_size, XE_VM_BIND_OP_MAP, > XE_VM_BIND_FLAG_ASYNC | > INJECT_ERROR, &sync, 1, 0, 0); > else > @@ -479,7 +479,7 @@ static void vm_async_ops_err(int fd, bool destroy) > if (i == N_BINDS / 8) > __xe_vm_bind_assert(fd, vm, 0, 0, 0, > addr + i * bo_size * 2, > - bo_size, XE_VM_BIND_OP_UNMAP | > + bo_size, XE_VM_BIND_OP_UNMAP, > XE_VM_BIND_FLAG_ASYNC | > INJECT_ERROR, &sync, 1, 0, 0); > else > @@ -928,7 +928,8 @@ test_bind_array(int fd, struct drm_xe_engine_class_instance *eci, int n_execs, > bind_ops[i].range = bo_size; > bind_ops[i].addr = addr; > bind_ops[i].tile_mask = 0x1 << eci->gt_id; > - bind_ops[i].op = XE_VM_BIND_OP_MAP | XE_VM_BIND_FLAG_ASYNC; > + bind_ops[i].op = XE_VM_BIND_OP_MAP; > + bind_ops[i].flags = XE_VM_BIND_FLAG_ASYNC; > bind_ops[i].region = 0; > bind_ops[i].reserved[0] = 0; > bind_ops[i].reserved[1] = 0; > @@ -972,7 +973,8 @@ test_bind_array(int fd, struct drm_xe_engine_class_instance *eci, int n_execs, > > for (i = 0; i < n_execs; ++i) { > bind_ops[i].obj = 0; > - bind_ops[i].op = XE_VM_BIND_OP_UNMAP | XE_VM_BIND_FLAG_ASYNC; > + bind_ops[i].op = XE_VM_BIND_OP_UNMAP; > + bind_ops[i].flags = XE_VM_BIND_FLAG_ASYNC; > } > > syncobj_reset(fd, &sync[0].handle, 1); > -- > 2.41.0 >