From: Matthew Brost <matthew.brost@intel.com>
To: <intel-xe@lists.freedesktop.org>
Cc: <thomas.hellstrom@linux.intel.com>,
Matthew Brost <matthew.brost@intel.com>
Subject: [PATCH v3 21/22] drm/xe: Don't use migrate exec queue for page fault binds
Date: Tue, 6 Feb 2024 15:37:28 -0800 [thread overview]
Message-ID: <20240206233729.3173206-22-matthew.brost@intel.com> (raw)
In-Reply-To: <20240206233729.3173206-1-matthew.brost@intel.com>
Now that the CPU is always used for binds even in jobs, CPU bind jobs
can pass GPU jobs in the same exec queue resulting dma-fences signaling
out-of-order. Use a dedicated exec queue for binds issued from page
faults to avoid ordering issues.
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
---
drivers/gpu/drm/xe/xe_gt_pagefault.c | 2 +-
drivers/gpu/drm/xe/xe_migrate.c | 22 +++++++++++++++++++++-
drivers/gpu/drm/xe/xe_migrate.h | 1 +
3 files changed, 23 insertions(+), 2 deletions(-)
diff --git a/drivers/gpu/drm/xe/xe_gt_pagefault.c b/drivers/gpu/drm/xe/xe_gt_pagefault.c
index 2beba426517e..a30bcf314589 100644
--- a/drivers/gpu/drm/xe/xe_gt_pagefault.c
+++ b/drivers/gpu/drm/xe/xe_gt_pagefault.c
@@ -210,7 +210,7 @@ static int handle_pagefault(struct xe_gt *gt, struct pagefault *pf)
if (ret)
goto unlock_dma_resv;
vm->dummy_ops.vops.pt_update_ops[tile->id].q =
- xe_tile_migrate_exec_queue(tile);
+ xe_tile_migrate_bind_exec_queue(tile);
fence = xe_vm_ops_execute(vm, &vm->dummy_ops.vops);
xe_vma_ops_free(&vm->dummy_ops.vops);
if (IS_ERR(fence)) {
diff --git a/drivers/gpu/drm/xe/xe_migrate.c b/drivers/gpu/drm/xe/xe_migrate.c
index 7d752e424b3b..1dd73e2117ea 100644
--- a/drivers/gpu/drm/xe/xe_migrate.c
+++ b/drivers/gpu/drm/xe/xe_migrate.c
@@ -40,6 +40,8 @@
struct xe_migrate {
/** @q: Default exec queue used for migration */
struct xe_exec_queue *q;
+ /** @bind_q: Default exec queue used for binds */
+ struct xe_exec_queue *bind_q;
/** @tile: Backpointer to the tile this struct xe_migrate belongs to. */
struct xe_tile *tile;
/** @job_mutex: Timeline mutex for @eng. */
@@ -96,6 +98,11 @@ struct xe_exec_queue *xe_tile_migrate_exec_queue(struct xe_tile *tile)
return tile->migrate->q;
}
+struct xe_exec_queue *xe_tile_migrate_bind_exec_queue(struct xe_tile *tile)
+{
+ return tile->migrate->bind_q;
+}
+
static void xe_migrate_fini(struct drm_device *dev, void *arg)
{
struct xe_migrate *m = arg;
@@ -110,6 +117,8 @@ static void xe_migrate_fini(struct drm_device *dev, void *arg)
mutex_destroy(&m->job_mutex);
xe_vm_close_and_put(m->q->vm);
xe_exec_queue_put(m->q);
+ if (m->bind_q)
+ xe_exec_queue_put(m->bind_q);
}
static u64 xe_migrate_vm_addr(u64 slot, u32 level)
@@ -367,6 +376,15 @@ struct xe_migrate *xe_migrate_init(struct xe_tile *tile)
if (!hwe || !logical_mask)
return ERR_PTR(-EINVAL);
+ m->bind_q = xe_exec_queue_create(xe, vm, logical_mask, 1, hwe,
+ EXEC_QUEUE_FLAG_KERNEL |
+ EXEC_QUEUE_FLAG_PERMANENT |
+ EXEC_QUEUE_FLAG_HIGH_PRIORITY, 0);
+ if (IS_ERR(m->bind_q)) {
+ xe_vm_close_and_put(vm);
+ return ERR_CAST(m->bind_q);
+ }
+
m->q = xe_exec_queue_create(xe, vm, logical_mask, 1, hwe,
EXEC_QUEUE_FLAG_KERNEL |
EXEC_QUEUE_FLAG_PERMANENT |
@@ -378,6 +396,8 @@ struct xe_migrate *xe_migrate_init(struct xe_tile *tile)
EXEC_QUEUE_FLAG_PERMANENT);
}
if (IS_ERR(m->q)) {
+ if (m->bind_q)
+ xe_exec_queue_put(m->bind_q);
xe_vm_close_and_put(vm);
return ERR_CAST(m->q);
}
@@ -1185,7 +1205,7 @@ __xe_migrate_update_pgtables(struct xe_migrate *m,
struct xe_tile *tile = m->tile;
struct xe_sched_job *job;
struct dma_fence *fence;
- bool is_migrate = pt_update_ops->q == m->q;
+ bool is_migrate = pt_update_ops->q == m->bind_q;
int err;
if (is_migrate)
diff --git a/drivers/gpu/drm/xe/xe_migrate.h b/drivers/gpu/drm/xe/xe_migrate.h
index f80c94bb8f4c..701bb27349b0 100644
--- a/drivers/gpu/drm/xe/xe_migrate.h
+++ b/drivers/gpu/drm/xe/xe_migrate.h
@@ -119,5 +119,6 @@ xe_migrate_update_pgtables(struct xe_migrate *m,
void xe_migrate_wait(struct xe_migrate *m);
struct xe_exec_queue *xe_tile_migrate_exec_queue(struct xe_tile *tile);
+struct xe_exec_queue *xe_tile_migrate_bind_exec_queue(struct xe_tile *tile);
#endif
--
2.34.1
next prev parent reply other threads:[~2024-02-06 23:37 UTC|newest]
Thread overview: 28+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-02-06 23:37 [PATCH v3 00/22] Refactor VM bind code Matthew Brost
2024-02-06 23:37 ` [PATCH v3 01/22] drm/xe: Lock all gpuva ops during VM bind IOCTL Matthew Brost
2024-02-06 23:37 ` [PATCH v3 02/22] drm/xe: Add ops_execute function which returns a fence Matthew Brost
2024-02-06 23:37 ` [PATCH v3 03/22] drm/xe: Move migrate to prefetch to op_lock funtion Matthew Brost
2024-02-06 23:37 ` [PATCH v3 04/22] drm/xe: Add struct xe_vma_ops abstraction Matthew Brost
2024-02-06 23:37 ` [PATCH v3 05/22] drm/xe: Update xe_vm_rebind to use dummy VMA operations Matthew Brost
2024-02-06 23:37 ` [PATCH v3 06/22] drm/xe: Simplify VM bind IOCTL error handling and cleanup Matthew Brost
2024-02-06 23:37 ` [PATCH v3 07/22] drm/xe: Update pagefaults to use dummy VMA operations Matthew Brost
2024-02-06 23:37 ` [PATCH v3 08/22] drm/xe: s/xe_tile_migrate_engine/xe_tile_migrate_exec_queue Matthew Brost
2024-02-06 23:37 ` [PATCH v3 09/22] drm/xe: Add vm_bind_ioctl_ops_install_fences helper Matthew Brost
2024-02-06 23:37 ` [PATCH v3 10/22] drm/xe: Move setting last fence to vm_bind_ioctl_ops_install_fences Matthew Brost
2024-02-06 23:37 ` [PATCH v3 11/22] drm/xe: Add xe_gt_tlb_invalidation_range and convert PT layer to use this Matthew Brost
2024-02-06 23:37 ` [PATCH v3 12/22] drm/xe: Add some members to xe_vma_ops Matthew Brost
2024-02-06 23:37 ` [PATCH v3 13/22] drm/xe: Add xe_vm_pgtable_update_op " Matthew Brost
2024-02-06 23:37 ` [PATCH v3 14/22] drm/xe: Convert multiple bind ops into single job Matthew Brost
2024-02-06 23:37 ` [PATCH v3 15/22] drm/xe: Remove old functions defs in xe_pt.h Matthew Brost
2024-02-06 23:37 ` [PATCH v3 16/22] drm/xe: Update PT layer with better error handling Matthew Brost
2024-02-06 23:37 ` [PATCH v3 17/22] drm/xe: Update VM trace events Matthew Brost
2024-02-06 23:37 ` [PATCH v3 18/22] drm/xe: Update clear / populate arguments Matthew Brost
2024-02-06 23:37 ` [PATCH v3 19/22] drm/xe: Add __xe_migrate_update_pgtables_cpu helper Matthew Brost
2024-02-06 23:37 ` [PATCH v3 20/22] drm/xe: CPU binds for jobs Matthew Brost
2024-03-28 11:28 ` Thomas Hellström
2024-02-06 23:37 ` Matthew Brost [this message]
2024-02-06 23:37 ` [PATCH v3 22/22] drm/xe: Add VM bind IOCTL error injection Matthew Brost
2024-02-07 0:12 ` ✗ CI.Patch_applied: failure for Refactor VM bind code (rev3) Patchwork
2024-02-07 22:11 ` ✓ CI.Patch_applied: success for Refactor VM bind code (rev4) Patchwork
2024-02-07 22:12 ` ✗ CI.checkpatch: warning " Patchwork
2024-02-07 22:13 ` ✗ CI.KUnit: failure " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20240206233729.3173206-22-matthew.brost@intel.com \
--to=matthew.brost@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=thomas.hellstrom@linux.intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox