From: Matthew Brost <matthew.brost@intel.com>
To: intel-xe@lists.freedesktop.org
Cc: dri-devel@lists.freedesktop.org
Subject: [RFC PATCH 06/12] drm/xe: Convert to DRM dep queue scheduler layer
Date: Sun, 15 Mar 2026 21:32:49 -0700 [thread overview]
Message-ID: <20260316043255.226352-7-matthew.brost@intel.com> (raw)
In-Reply-To: <20260316043255.226352-1-matthew.brost@intel.com>
Replace the drm_gpu_scheduler/drm_sched_entity pair used throughout Xe
with the new drm_dep layer (struct drm_dep_queue / struct drm_dep_job).
The conversion spans three submission backends — GuC, execlist, and the
generic dependency scheduler (xe_dep_scheduler) — as well as the job
lifecycle, TDR callbacks, and device teardown sequencing.
xe_gpu_scheduler: struct drm_gpu_scheduler base replaced with struct
drm_dep_queue. xe_sched_init() updated to drm_dep_queue_init() args.
The xe_sched_entity alias now maps to drm_dep_queue (the N:1
entity-to-scheduler distinction disappears entirely since each queue
is its own entity). drm_sched_for_each_pending_job replaced with
drm_dep_queue_for_each_pending_job. drm_sched_tdr_queue_imm replaced
with drm_dep_queue_trigger_timeout.
GuC backend: guc_exec_queue_free_job() removed; job lifetime is now
managed by drm_dep_job refcounting rather than a free_job vfunc.
guc_exec_queue_timedout_job() updated to return drm_dep_timedout_stat
values; the already-signaled check is replaced with
drm_dep_job_is_finished(); vf_recovery paths now return
DRM_DEP_TIMEDOUT_STAT_REQUEUE_JOB. guc_dep_queue_release() added as
the .release vfunc performing kfree_rcu on the embedded
xe_guc_exec_queue.
xe_dep_scheduler: struct drm_gpu_scheduler + struct drm_sched_entity +
struct rcu_head collapsed into a single struct drm_dep_queue (which
carries its own rcu_head). drm_sched_entity_init/fini removed.
xe_dep_scheduler_fini() simplified to drm_dep_queue_put().
xe_dep_scheduler_entity() renamed to xe_dep_scheduler_dep_q() to match
the new naming.
GuC teardown sequencing: the previous wait_event_timeout-based drain is
replaced by the drm_dep module's built-in unload protection. Each
drm_dep_queue holds a drm_dev_get() reference on its owning
struct drm_device; drm_dev_put() is called as the final step of queue
teardown. This ensures the driver module cannot be unloaded while any
queue is still alive without requiring a separate drain API or per-device
hash table. guc_submit_fini() is simplified accordingly.
xe_sched_job: to_xe_sched_job() and drm_job accessors updated for
struct drm_dep_job. Job dependency calls updated throughout
(drm_dep_job_add_dependency, drm_dep_job_arm, drm_dep_job_push).
exec_queue: q->entity renamed to q->dep_q throughout.
Signed-off-by: Matthew Brost <matthew.brost@intel.com>
Assisted-by: GitHub Copilot:claude-sonnet-4.6
Me: WIP
---
drivers/gpu/drm/xe/Kconfig | 2 +-
drivers/gpu/drm/xe/xe_dep_job_types.h | 8 +-
drivers/gpu/drm/xe/xe_dep_scheduler.c | 81 ++++----
drivers/gpu/drm/xe/xe_dep_scheduler.h | 7 +-
drivers/gpu/drm/xe/xe_exec_queue_types.h | 6 +-
drivers/gpu/drm/xe/xe_execlist.c | 43 ++---
drivers/gpu/drm/xe/xe_execlist_types.h | 4 +-
drivers/gpu/drm/xe/xe_gpu_scheduler.c | 38 ++--
drivers/gpu/drm/xe/xe_gpu_scheduler.h | 50 ++---
drivers/gpu/drm/xe/xe_gpu_scheduler_types.h | 9 +-
drivers/gpu/drm/xe/xe_guc_exec_queue_types.h | 8 +-
drivers/gpu/drm/xe/xe_guc_submit.c | 184 ++++++++-----------
drivers/gpu/drm/xe/xe_migrate.c | 2 +-
drivers/gpu/drm/xe/xe_pt.c | 2 +-
drivers/gpu/drm/xe/xe_sched_job.c | 52 +++---
drivers/gpu/drm/xe/xe_sched_job.h | 7 +-
drivers/gpu/drm/xe/xe_sched_job_types.h | 8 +-
drivers/gpu/drm/xe/xe_sync.c | 2 +-
drivers/gpu/drm/xe/xe_tlb_inval_job.c | 86 ++++-----
19 files changed, 255 insertions(+), 344 deletions(-)
diff --git a/drivers/gpu/drm/xe/Kconfig b/drivers/gpu/drm/xe/Kconfig
index 4d7dcaff2b91..9430877d6294 100644
--- a/drivers/gpu/drm/xe/Kconfig
+++ b/drivers/gpu/drm/xe/Kconfig
@@ -41,7 +41,7 @@ config DRM_XE
select DRM_EXEC
select DRM_GPUSVM if !UML
select DRM_GPUVM
- select DRM_SCHED
+ select DRM_DEP
select MMU_NOTIFIER
select WANT_DEV_COREDUMP
select AUXILIARY_BUS
diff --git a/drivers/gpu/drm/xe/xe_dep_job_types.h b/drivers/gpu/drm/xe/xe_dep_job_types.h
index c6a484f24c8c..891fe5cfcf89 100644
--- a/drivers/gpu/drm/xe/xe_dep_job_types.h
+++ b/drivers/gpu/drm/xe/xe_dep_job_types.h
@@ -6,7 +6,7 @@
#ifndef _XE_DEP_JOB_TYPES_H_
#define _XE_DEP_JOB_TYPES_H_
-#include <drm/gpu_scheduler.h>
+#include <drm/drm_dep.h>
struct xe_dep_job;
@@ -14,14 +14,12 @@ struct xe_dep_job;
struct xe_dep_job_ops {
/** @run_job: Run generic Xe dependency job */
struct dma_fence *(*run_job)(struct xe_dep_job *job);
- /** @free_job: Free generic Xe dependency job */
- void (*free_job)(struct xe_dep_job *job);
};
/** struct xe_dep_job - Generic dependency Xe job */
struct xe_dep_job {
- /** @drm: base DRM scheduler job */
- struct drm_sched_job drm;
+ /** @drm: base DRM dependency job */
+ struct drm_dep_job drm;
/** @ops: dependency job operations */
const struct xe_dep_job_ops *ops;
};
diff --git a/drivers/gpu/drm/xe/xe_dep_scheduler.c b/drivers/gpu/drm/xe/xe_dep_scheduler.c
index 51d99fee9aa5..d3fec14d7073 100644
--- a/drivers/gpu/drm/xe/xe_dep_scheduler.c
+++ b/drivers/gpu/drm/xe/xe_dep_scheduler.c
@@ -5,11 +5,12 @@
#include <linux/slab.h>
-#include <drm/gpu_scheduler.h>
+#include <drm/drm_dep.h>
#include "xe_dep_job_types.h"
#include "xe_dep_scheduler.h"
-#include "xe_device_types.h"
+#include "xe_device.h"
+#include "xe_gt_types.h"
/**
* DOC: Xe Dependency Scheduler
@@ -27,15 +28,11 @@
/** struct xe_dep_scheduler - Generic Xe dependency scheduler */
struct xe_dep_scheduler {
- /** @sched: DRM GPU scheduler */
- struct drm_gpu_scheduler sched;
- /** @entity: DRM scheduler entity */
- struct drm_sched_entity entity;
- /** @rcu: For safe freeing of exported dma fences */
- struct rcu_head rcu;
+ /** @queue: DRM dependency queue */
+ struct drm_dep_queue queue;
};
-static struct dma_fence *xe_dep_scheduler_run_job(struct drm_sched_job *drm_job)
+static struct dma_fence *xe_dep_scheduler_run_job(struct drm_dep_job *drm_job)
{
struct xe_dep_job *dep_job =
container_of(drm_job, typeof(*dep_job), drm);
@@ -43,17 +40,21 @@ static struct dma_fence *xe_dep_scheduler_run_job(struct drm_sched_job *drm_job)
return dep_job->ops->run_job(dep_job);
}
-static void xe_dep_scheduler_free_job(struct drm_sched_job *drm_job)
+static void xe_dep_scheduler_release(struct drm_dep_queue *drm_q)
{
- struct xe_dep_job *dep_job =
- container_of(drm_job, typeof(*dep_job), drm);
+ struct xe_dep_scheduler *dep_scheduler =
+ container_of(drm_q, typeof(*dep_scheduler), queue);
- dep_job->ops->free_job(dep_job);
+ /*
+ * RCU free due sched being exported via DRM scheduler fences
+ * (timeline name).
+ */
+ kfree_rcu(dep_scheduler, queue.rcu);
}
-static const struct drm_sched_backend_ops sched_ops = {
+static const struct drm_dep_queue_ops sched_ops = {
.run_job = xe_dep_scheduler_run_job,
- .free_job = xe_dep_scheduler_free_job,
+ .release = xe_dep_scheduler_release,
};
/**
@@ -74,37 +75,28 @@ xe_dep_scheduler_create(struct xe_device *xe,
const char *name, u32 job_limit)
{
struct xe_dep_scheduler *dep_scheduler;
- struct drm_gpu_scheduler *sched;
- const struct drm_sched_init_args args = {
- .ops = &sched_ops,
- .submit_wq = submit_wq,
- .num_rqs = 1,
- .credit_limit = job_limit,
- .timeout = MAX_SCHEDULE_TIMEOUT,
- .name = name,
- .dev = xe->drm.dev,
- };
+ struct xe_gt *gt = xe_device_get_root_tile(xe)->primary_gt;
int err;
dep_scheduler = kzalloc_obj(*dep_scheduler);
if (!dep_scheduler)
return ERR_PTR(-ENOMEM);
- err = drm_sched_init(&dep_scheduler->sched, &args);
+ err = drm_dep_queue_init(&dep_scheduler->queue,
+ &(const struct drm_dep_queue_init_args){
+ .ops = &sched_ops,
+ .submit_wq = submit_wq,
+ .timeout_wq = gt->ordered_wq,
+ .credit_limit = job_limit,
+ .timeout = MAX_SCHEDULE_TIMEOUT,
+ .name = name,
+ .drm = &xe->drm,
+ });
if (err)
goto err_free;
- sched = &dep_scheduler->sched;
- err = drm_sched_entity_init(&dep_scheduler->entity, 0, &sched, 1, NULL);
- if (err)
- goto err_sched;
-
- init_rcu_head(&dep_scheduler->rcu);
-
return dep_scheduler;
-err_sched:
- drm_sched_fini(&dep_scheduler->sched);
err_free:
kfree(dep_scheduler);
@@ -120,24 +112,17 @@ xe_dep_scheduler_create(struct xe_device *xe,
*/
void xe_dep_scheduler_fini(struct xe_dep_scheduler *dep_scheduler)
{
- drm_sched_entity_fini(&dep_scheduler->entity);
- drm_sched_fini(&dep_scheduler->sched);
- /*
- * RCU free due sched being exported via DRM scheduler fences
- * (timeline name).
- */
- kfree_rcu(dep_scheduler, rcu);
+ drm_dep_queue_put(&dep_scheduler->queue);
}
/**
- * xe_dep_scheduler_entity() - Retrieve a generic Xe dependency scheduler
- * DRM scheduler entity
+ * xe_dep_scheduler_dep_q() - Retrieve the dep queue for a generic Xe dependency scheduler
* @dep_scheduler: Generic Xe dependency scheduler object
*
- * Return: The generic Xe dependency scheduler's DRM scheduler entity
+ * Return: The &drm_dep_queue owned by @dep_scheduler.
*/
-struct drm_sched_entity *
-xe_dep_scheduler_entity(struct xe_dep_scheduler *dep_scheduler)
+struct drm_dep_queue *
+xe_dep_scheduler_dep_q(struct xe_dep_scheduler *dep_scheduler)
{
- return &dep_scheduler->entity;
+ return &dep_scheduler->queue;
}
diff --git a/drivers/gpu/drm/xe/xe_dep_scheduler.h b/drivers/gpu/drm/xe/xe_dep_scheduler.h
index 853961eec64b..c32b6f4f8c04 100644
--- a/drivers/gpu/drm/xe/xe_dep_scheduler.h
+++ b/drivers/gpu/drm/xe/xe_dep_scheduler.h
@@ -5,8 +5,9 @@
#include <linux/types.h>
-struct drm_sched_entity;
+struct drm_dep_queue;
struct workqueue_struct;
+struct xe_dep_job;
struct xe_dep_scheduler;
struct xe_device;
@@ -17,5 +18,5 @@ xe_dep_scheduler_create(struct xe_device *xe,
void xe_dep_scheduler_fini(struct xe_dep_scheduler *dep_scheduler);
-struct drm_sched_entity *
-xe_dep_scheduler_entity(struct xe_dep_scheduler *dep_scheduler);
+struct drm_dep_queue *
+xe_dep_scheduler_dep_q(struct xe_dep_scheduler *dep_scheduler);
diff --git a/drivers/gpu/drm/xe/xe_exec_queue_types.h b/drivers/gpu/drm/xe/xe_exec_queue_types.h
index 8ce78e0b1d50..35c7625a2df5 100644
--- a/drivers/gpu/drm/xe/xe_exec_queue_types.h
+++ b/drivers/gpu/drm/xe/xe_exec_queue_types.h
@@ -8,7 +8,7 @@
#include <linux/kref.h>
-#include <drm/gpu_scheduler.h>
+#include <drm/drm_dep.h>
#include "xe_gpu_scheduler_types.h"
#include "xe_hw_engine_types.h"
@@ -245,8 +245,8 @@ struct xe_exec_queue {
/** @ring_ops: ring operations for this exec queue */
const struct xe_ring_ops *ring_ops;
- /** @entity: DRM sched entity for this exec queue (1 to 1 relationship) */
- struct drm_sched_entity *entity;
+ /** @dep_q: dep queue for this exec queue (1 to 1 relationship) */
+ struct drm_dep_queue *dep_q;
#define XE_MAX_JOB_COUNT_PER_EXEC_QUEUE 1000
/** @job_cnt: number of drm jobs in this exec queue */
diff --git a/drivers/gpu/drm/xe/xe_execlist.c b/drivers/gpu/drm/xe/xe_execlist.c
index 755a2bff5d7b..fb948b2c617c 100644
--- a/drivers/gpu/drm/xe/xe_execlist.c
+++ b/drivers/gpu/drm/xe/xe_execlist.c
@@ -307,7 +307,7 @@ void xe_execlist_port_destroy(struct xe_execlist_port *port)
}
static struct dma_fence *
-execlist_run_job(struct drm_sched_job *drm_job)
+execlist_run_job(struct drm_dep_job *drm_job)
{
struct xe_sched_job *job = to_xe_sched_job(drm_job);
struct xe_exec_queue *q = job->q;
@@ -319,30 +319,31 @@ execlist_run_job(struct drm_sched_job *drm_job)
return job->fence;
}
-static void execlist_job_free(struct drm_sched_job *drm_job)
+static void execlist_dep_queue_release(struct drm_dep_queue *q)
{
- struct xe_sched_job *job = to_xe_sched_job(drm_job);
+ struct xe_execlist_exec_queue *exl =
+ container_of(q, typeof(*exl), queue);
- xe_exec_queue_update_run_ticks(job->q);
- xe_sched_job_put(job);
+ /*
+ * RCU free: the dep queue's name may be referenced by exported dma
+ * fences (timeline name). Defer freeing until after any RCU readers.
+ */
+ kfree_rcu(exl, queue.rcu);
}
-static const struct drm_sched_backend_ops drm_sched_ops = {
+static const struct drm_dep_queue_ops execlist_dep_queue_ops = {
.run_job = execlist_run_job,
- .free_job = execlist_job_free,
+ .release = execlist_dep_queue_release,
};
static int execlist_exec_queue_init(struct xe_exec_queue *q)
{
- struct drm_gpu_scheduler *sched;
- const struct drm_sched_init_args args = {
- .ops = &drm_sched_ops,
- .num_rqs = 1,
+ const struct drm_dep_queue_init_args args = {
+ .ops = &execlist_dep_queue_ops,
.credit_limit = xe_lrc_ring_size() / MAX_JOB_SIZE_BYTES,
- .hang_limit = XE_SCHED_HANG_LIMIT,
.timeout = XE_SCHED_JOB_TIMEOUT,
.name = q->hwe->name,
- .dev = gt_to_xe(q->gt)->drm.dev,
+ .drm = >_to_xe(q->gt)->drm,
};
struct xe_execlist_exec_queue *exl;
struct xe_device *xe = gt_to_xe(q->gt);
@@ -358,27 +359,20 @@ static int execlist_exec_queue_init(struct xe_exec_queue *q)
exl->q = q;
- err = drm_sched_init(&exl->sched, &args);
+ err = drm_dep_queue_init(&exl->queue, &args);
if (err)
goto err_free;
- sched = &exl->sched;
- err = drm_sched_entity_init(&exl->entity, 0, &sched, 1, NULL);
- if (err)
- goto err_sched;
-
exl->port = q->hwe->exl_port;
exl->has_run = false;
exl->active_priority = XE_EXEC_QUEUE_PRIORITY_UNSET;
q->execlist = exl;
- q->entity = &exl->entity;
+ q->dep_q = &exl->queue;
xe_exec_queue_assign_name(q, ffs(q->logical_mask) - 1);
return 0;
-err_sched:
- drm_sched_fini(&exl->sched);
err_free:
kfree(exl);
return err;
@@ -388,10 +382,7 @@ static void execlist_exec_queue_fini(struct xe_exec_queue *q)
{
struct xe_execlist_exec_queue *exl = q->execlist;
- drm_sched_entity_fini(&exl->entity);
- drm_sched_fini(&exl->sched);
-
- kfree(exl);
+ drm_dep_queue_put(&exl->queue);
}
static void execlist_exec_queue_destroy_async(struct work_struct *w)
diff --git a/drivers/gpu/drm/xe/xe_execlist_types.h b/drivers/gpu/drm/xe/xe_execlist_types.h
index 92c4ba52db0c..c2c8218db350 100644
--- a/drivers/gpu/drm/xe/xe_execlist_types.h
+++ b/drivers/gpu/drm/xe/xe_execlist_types.h
@@ -34,9 +34,7 @@ struct xe_execlist_port {
struct xe_execlist_exec_queue {
struct xe_exec_queue *q;
- struct drm_gpu_scheduler sched;
-
- struct drm_sched_entity entity;
+ struct drm_dep_queue queue;
struct xe_execlist_port *port;
diff --git a/drivers/gpu/drm/xe/xe_gpu_scheduler.c b/drivers/gpu/drm/xe/xe_gpu_scheduler.c
index 9c8004d5dd91..a8e6384dffe8 100644
--- a/drivers/gpu/drm/xe/xe_gpu_scheduler.c
+++ b/drivers/gpu/drm/xe/xe_gpu_scheduler.c
@@ -7,8 +7,7 @@
static void xe_sched_process_msg_queue(struct xe_gpu_scheduler *sched)
{
- if (!drm_sched_is_stopped(&sched->base))
- queue_work(sched->base.submit_wq, &sched->work_process_msg);
+ drm_dep_queue_work_enqueue(&sched->base, &sched->work_process_msg);
}
static void xe_sched_process_msg_queue_if_ready(struct xe_gpu_scheduler *sched)
@@ -43,7 +42,9 @@ static void xe_sched_process_msg_work(struct work_struct *w)
container_of(w, struct xe_gpu_scheduler, work_process_msg);
struct xe_sched_msg *msg;
- if (drm_sched_is_stopped(&sched->base))
+ drm_dep_queue_sched_guard(&sched->base);
+
+ if (drm_dep_queue_is_stopped(&sched->base))
return;
msg = xe_sched_get_msg(sched);
@@ -55,25 +56,23 @@ static void xe_sched_process_msg_work(struct work_struct *w)
}
int xe_sched_init(struct xe_gpu_scheduler *sched,
- const struct drm_sched_backend_ops *ops,
+ const struct drm_dep_queue_ops *ops,
const struct xe_sched_backend_ops *xe_ops,
struct workqueue_struct *submit_wq,
- uint32_t hw_submission, unsigned hang_limit,
- long timeout, struct workqueue_struct *timeout_wq,
- atomic_t *score, const char *name,
- struct device *dev)
+ uint32_t hw_submission, long timeout,
+ struct workqueue_struct *timeout_wq,
+ enum drm_dep_queue_flags flags,
+ const char *name, struct drm_device *drm)
{
- const struct drm_sched_init_args args = {
+ const struct drm_dep_queue_init_args args = {
.ops = ops,
.submit_wq = submit_wq,
- .num_rqs = 1,
.credit_limit = hw_submission,
- .hang_limit = hang_limit,
.timeout = timeout,
.timeout_wq = timeout_wq,
- .score = score,
.name = name,
- .dev = dev,
+ .drm = drm,
+ .flags = flags,
};
sched->ops = xe_ops;
@@ -81,30 +80,29 @@ int xe_sched_init(struct xe_gpu_scheduler *sched,
INIT_LIST_HEAD(&sched->msgs);
INIT_WORK(&sched->work_process_msg, xe_sched_process_msg_work);
- return drm_sched_init(&sched->base, &args);
+ return drm_dep_queue_init(&sched->base, &args);
}
void xe_sched_fini(struct xe_gpu_scheduler *sched)
{
- xe_sched_submission_stop(sched);
- drm_sched_fini(&sched->base);
+ drm_dep_queue_put(&sched->base);
}
void xe_sched_submission_start(struct xe_gpu_scheduler *sched)
{
- drm_sched_wqueue_start(&sched->base);
- queue_work(sched->base.submit_wq, &sched->work_process_msg);
+ drm_dep_queue_start(&sched->base);
+ drm_dep_queue_work_enqueue(&sched->base, &sched->work_process_msg);
}
void xe_sched_submission_stop(struct xe_gpu_scheduler *sched)
{
- drm_sched_wqueue_stop(&sched->base);
+ drm_dep_queue_stop(&sched->base);
cancel_work_sync(&sched->work_process_msg);
}
void xe_sched_submission_resume_tdr(struct xe_gpu_scheduler *sched)
{
- drm_sched_resume_timeout(&sched->base, sched->base.timeout);
+ drm_dep_queue_resume_timeout(&sched->base);
}
void xe_sched_add_msg(struct xe_gpu_scheduler *sched,
diff --git a/drivers/gpu/drm/xe/xe_gpu_scheduler.h b/drivers/gpu/drm/xe/xe_gpu_scheduler.h
index 664c2db56af3..4086aafb0a9a 100644
--- a/drivers/gpu/drm/xe/xe_gpu_scheduler.h
+++ b/drivers/gpu/drm/xe/xe_gpu_scheduler.h
@@ -10,13 +10,13 @@
#include "xe_sched_job.h"
int xe_sched_init(struct xe_gpu_scheduler *sched,
- const struct drm_sched_backend_ops *ops,
+ const struct drm_dep_queue_ops *ops,
const struct xe_sched_backend_ops *xe_ops,
struct workqueue_struct *submit_wq,
- uint32_t hw_submission, unsigned hang_limit,
- long timeout, struct workqueue_struct *timeout_wq,
- atomic_t *score, const char *name,
- struct device *dev);
+ uint32_t hw_submission, long timeout,
+ struct workqueue_struct *timeout_wq,
+ enum drm_dep_queue_flags flags,
+ const char *name, struct drm_device *drm);
void xe_sched_fini(struct xe_gpu_scheduler *sched);
void xe_sched_submission_start(struct xe_gpu_scheduler *sched);
@@ -41,32 +41,29 @@ static inline void xe_sched_msg_unlock(struct xe_gpu_scheduler *sched)
spin_unlock(&sched->msg_lock);
}
-static inline void xe_sched_stop(struct xe_gpu_scheduler *sched)
-{
- drm_sched_stop(&sched->base, NULL);
-}
-
static inline void xe_sched_tdr_queue_imm(struct xe_gpu_scheduler *sched)
{
- drm_sched_tdr_queue_imm(&sched->base);
+ drm_dep_queue_trigger_timeout(&sched->base);
}
static inline void xe_sched_resubmit_jobs(struct xe_gpu_scheduler *sched)
{
- struct drm_sched_job *s_job;
+ struct drm_dep_job *drm_job;
+ struct xe_sched_job *job;
bool restore_replay = false;
- drm_sched_for_each_pending_job(s_job, &sched->base, NULL) {
- restore_replay |= to_xe_sched_job(s_job)->restore_replay;
- if (restore_replay || !drm_sched_job_is_signaled(s_job))
- sched->base.ops->run_job(s_job);
+ drm_dep_queue_for_each_pending_job(drm_job, &sched->base) {
+ job = to_xe_sched_job(drm_job);
+ restore_replay |= job->restore_replay;
+ if (restore_replay || !drm_dep_job_is_signaled(drm_job))
+ sched->base.ops->run_job(drm_job);
}
}
static inline bool
xe_sched_invalidate_job(struct xe_sched_job *job, int threshold)
{
- return drm_sched_invalidate_job(&job->drm, threshold);
+ return drm_dep_job_invalidate_job(&job->drm, threshold);
}
/**
@@ -78,24 +75,13 @@ xe_sched_invalidate_job(struct xe_sched_job *job, int threshold)
static inline
struct xe_sched_job *xe_sched_first_pending_job(struct xe_gpu_scheduler *sched)
{
- struct drm_sched_job *job;
+ struct drm_dep_job *drm_job;
- drm_sched_for_each_pending_job(job, &sched->base, NULL)
- if (!drm_sched_job_is_signaled(job))
- return to_xe_sched_job(job);
+ drm_dep_queue_for_each_pending_job(drm_job, &sched->base)
+ if (!drm_dep_job_is_signaled(drm_job))
+ return to_xe_sched_job(drm_job);
return NULL;
}
-static inline int
-xe_sched_entity_init(struct xe_sched_entity *entity,
- struct xe_gpu_scheduler *sched)
-{
- return drm_sched_entity_init(entity, 0,
- (struct drm_gpu_scheduler **)&sched,
- 1, NULL);
-}
-
-#define xe_sched_entity_fini drm_sched_entity_fini
-
#endif
diff --git a/drivers/gpu/drm/xe/xe_gpu_scheduler_types.h b/drivers/gpu/drm/xe/xe_gpu_scheduler_types.h
index 63d9bf92583c..ff89d36d3b2a 100644
--- a/drivers/gpu/drm/xe/xe_gpu_scheduler_types.h
+++ b/drivers/gpu/drm/xe/xe_gpu_scheduler_types.h
@@ -6,7 +6,7 @@
#ifndef _XE_GPU_SCHEDULER_TYPES_H_
#define _XE_GPU_SCHEDULER_TYPES_H_
-#include <drm/gpu_scheduler.h>
+#include <drm/drm_dep.h>
/**
* struct xe_sched_msg - an in-band (relative to GPU scheduler run queue)
@@ -41,8 +41,8 @@ struct xe_sched_backend_ops {
* struct xe_gpu_scheduler - Xe GPU scheduler
*/
struct xe_gpu_scheduler {
- /** @base: DRM GPU scheduler */
- struct drm_gpu_scheduler base;
+ /** @base: DRM dependency queue */
+ struct drm_dep_queue base;
/** @ops: Xe scheduler ops */
const struct xe_sched_backend_ops *ops;
/** @msgs: list of messages to be processed in @work_process_msg */
@@ -53,7 +53,6 @@ struct xe_gpu_scheduler {
struct work_struct work_process_msg;
};
-#define xe_sched_entity drm_sched_entity
-#define xe_sched_policy drm_sched_policy
+#define xe_sched_entity drm_dep_queue
#endif
diff --git a/drivers/gpu/drm/xe/xe_guc_exec_queue_types.h b/drivers/gpu/drm/xe/xe_guc_exec_queue_types.h
index fd0915ed8eb1..42ba4892ff71 100644
--- a/drivers/gpu/drm/xe/xe_guc_exec_queue_types.h
+++ b/drivers/gpu/drm/xe/xe_guc_exec_queue_types.h
@@ -18,14 +18,10 @@ struct xe_exec_queue;
* struct xe_guc_exec_queue - GuC specific state for an xe_exec_queue
*/
struct xe_guc_exec_queue {
- /** @q: Backpointer to parent xe_exec_queue */
- struct xe_exec_queue *q;
- /** @rcu: For safe freeing of exported dma fences */
- struct rcu_head rcu;
/** @sched: GPU scheduler for this xe_exec_queue */
struct xe_gpu_scheduler sched;
- /** @entity: Scheduler entity for this xe_exec_queue */
- struct xe_sched_entity entity;
+ /** @q: Backpointer to parent xe_exec_queue */
+ struct xe_exec_queue *q;
/**
* @static_msgs: Static messages for this xe_exec_queue, used when
* a message needs to sent through the GPU scheduler but memory
diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c b/drivers/gpu/drm/xe/xe_guc_submit.c
index a145234f662b..fc9704fad177 100644
--- a/drivers/gpu/drm/xe/xe_guc_submit.c
+++ b/drivers/gpu/drm/xe/xe_guc_submit.c
@@ -244,17 +244,8 @@ static void guc_submit_sw_fini(struct drm_device *drm, void *arg)
{
struct xe_guc *guc = arg;
struct xe_device *xe = guc_to_xe(guc);
- struct xe_gt *gt = guc_to_gt(guc);
- int ret;
-
- ret = wait_event_timeout(guc->submission_state.fini_wq,
- xa_empty(&guc->submission_state.exec_queue_lookup),
- HZ * 5);
drain_workqueue(xe->destroy_wq);
-
- xe_gt_assert(gt, ret);
-
xa_destroy(&guc->submission_state.exec_queue_lookup);
}
@@ -1203,7 +1194,7 @@ static void submit_exec_queue(struct xe_exec_queue *q, struct xe_sched_job *job)
}
static struct dma_fence *
-guc_exec_queue_run_job(struct drm_sched_job *drm_job)
+guc_exec_queue_run_job(struct drm_dep_job *drm_job)
{
struct xe_sched_job *job = to_xe_sched_job(drm_job);
struct xe_exec_queue *q = job->q;
@@ -1242,13 +1233,6 @@ guc_exec_queue_run_job(struct drm_sched_job *drm_job)
return job->fence;
}
-static void guc_exec_queue_free_job(struct drm_sched_job *drm_job)
-{
- struct xe_sched_job *job = to_xe_sched_job(drm_job);
-
- trace_xe_sched_job_free(job);
- xe_sched_job_put(job);
-}
int xe_guc_read_stopped(struct xe_guc *guc)
{
@@ -1486,11 +1470,11 @@ static void disable_scheduling(struct xe_exec_queue *q, bool immediate)
G2H_LEN_DW_SCHED_CONTEXT_MODE_SET, 1);
}
-static enum drm_gpu_sched_stat
-guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
+static enum drm_dep_timedout_stat
+guc_exec_queue_timedout_job(struct drm_dep_job *drm_job)
{
struct xe_sched_job *job = to_xe_sched_job(drm_job);
- struct drm_sched_job *tmp_job;
+ struct drm_dep_job *tmp_job;
struct xe_exec_queue *q = job->q, *primary;
struct xe_gpu_scheduler *sched = &q->guc->sched;
struct xe_guc *guc = exec_queue_to_guc(q);
@@ -1502,17 +1486,13 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
xe_gt_assert(guc_to_gt(guc), !exec_queue_destroyed(q));
- primary = xe_exec_queue_multi_queue_primary(q);
+ if (drm_dep_job_is_finished(&job->drm))
+ return DRM_DEP_TIMEDOUT_STAT_JOB_SIGNALED;
- /*
- * TDR has fired before free job worker. Common if exec queue
- * immediately closed after last fence signaled. Add back to pending
- * list so job can be freed and kick scheduler ensuring free job is not
- * lost.
- */
- if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &job->fence->flags) ||
- vf_recovery(guc))
- return DRM_GPU_SCHED_STAT_NO_HANG;
+ if (vf_recovery(guc))
+ return DRM_DEP_TIMEDOUT_STAT_REQUEUE_JOB;
+
+ primary = xe_exec_queue_multi_queue_primary(q);
/* Kill the run_job entry point */
if (xe_exec_queue_is_multi_queue(q))
@@ -1577,7 +1557,7 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
xe_guc_read_stopped(guc) ||
vf_recovery(guc), HZ * 5);
if (vf_recovery(guc))
- goto handle_vf_resume;
+ return DRM_DEP_TIMEDOUT_STAT_REQUEUE_JOB;
if (!ret || xe_guc_read_stopped(guc))
goto trigger_reset;
@@ -1599,7 +1579,7 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
xe_guc_read_stopped(guc) ||
vf_recovery(guc), HZ * 5);
if (vf_recovery(guc))
- goto handle_vf_resume;
+ return DRM_DEP_TIMEDOUT_STAT_REQUEUE_JOB;
if (!ret || xe_guc_read_stopped(guc)) {
trigger_reset:
if (!ret)
@@ -1644,15 +1624,13 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
"VM job timed out on non-killed execqueue\n");
if (!wedged && (q->flags & EXEC_QUEUE_FLAG_KERNEL ||
(q->flags & EXEC_QUEUE_FLAG_VM && !exec_queue_killed(q)))) {
- if (!xe_sched_invalidate_job(job, 2)) {
+ if (!xe_sched_invalidate_job(job, 2))
xe_gt_reset_async(q->gt);
- goto rearm;
- }
}
/* Mark all outstanding jobs as bad, thus completing them */
xe_sched_job_set_error(job, err);
- drm_sched_for_each_pending_job(tmp_job, &sched->base, NULL)
+ drm_dep_queue_for_each_pending_job(tmp_job, &sched->base)
xe_sched_job_set_error(to_xe_sched_job(tmp_job), -ECANCELED);
if (xe_exec_queue_is_multi_queue(q)) {
@@ -1663,11 +1641,7 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
xe_guc_exec_queue_trigger_cleanup(q);
}
- /*
- * We want the job added back to the pending list so it gets freed; this
- * is what DRM_GPU_SCHED_STAT_NO_HANG does.
- */
- return DRM_GPU_SCHED_STAT_NO_HANG;
+ return DRM_DEP_TIMEDOUT_STAT_REQUEUE_JOB;
rearm:
/*
@@ -1679,8 +1653,8 @@ guc_exec_queue_timedout_job(struct drm_sched_job *drm_job)
xe_guc_exec_queue_group_start(q);
else
xe_sched_submission_start(sched);
-handle_vf_resume:
- return DRM_GPU_SCHED_STAT_NO_HANG;
+
+ return DRM_DEP_TIMEDOUT_STAT_REQUEUE_JOB;
}
static void guc_exec_queue_fini(struct xe_exec_queue *q)
@@ -1689,24 +1663,11 @@ static void guc_exec_queue_fini(struct xe_exec_queue *q)
struct xe_guc *guc = exec_queue_to_guc(q);
release_guc_id(guc, q);
- xe_sched_entity_fini(&ge->entity);
xe_sched_fini(&ge->sched);
-
- /*
- * RCU free due sched being exported via DRM scheduler fences
- * (timeline name).
- */
- kfree_rcu(ge, rcu);
}
-static void __guc_exec_queue_destroy_async(struct work_struct *w)
+static void __guc_exec_queue_destroy(struct xe_exec_queue *q)
{
- struct xe_guc_exec_queue *ge =
- container_of(w, struct xe_guc_exec_queue, destroy_async);
- struct xe_exec_queue *q = ge->q;
- struct xe_guc *guc = exec_queue_to_guc(q);
-
- guard(xe_pm_runtime)(guc_to_xe(guc));
trace_xe_exec_queue_destroy(q);
if (xe_exec_queue_is_multi_queue_secondary(q)) {
@@ -1717,36 +1678,25 @@ static void __guc_exec_queue_destroy_async(struct work_struct *w)
mutex_unlock(&group->list_lock);
}
- /* Confirm no work left behind accessing device structures */
- cancel_delayed_work_sync(&ge->sched.base.work_tdr);
-
xe_exec_queue_fini(q);
}
+static void __guc_exec_queue_destroy_async(struct work_struct *w)
+{
+ struct xe_guc_exec_queue *ge =
+ container_of(w, struct xe_guc_exec_queue, destroy_async);
+ struct xe_exec_queue *q = ge->q;
+
+ __guc_exec_queue_destroy(q);
+}
+
static void guc_exec_queue_destroy_async(struct xe_exec_queue *q)
{
struct xe_guc *guc = exec_queue_to_guc(q);
struct xe_device *xe = guc_to_xe(guc);
INIT_WORK(&q->guc->destroy_async, __guc_exec_queue_destroy_async);
-
- /* We must block on kernel engines so slabs are empty on driver unload */
- if (q->flags & EXEC_QUEUE_FLAG_PERMANENT || exec_queue_wedged(q))
- __guc_exec_queue_destroy_async(&q->guc->destroy_async);
- else
- queue_work(xe->destroy_wq, &q->guc->destroy_async);
-}
-
-static void __guc_exec_queue_destroy(struct xe_guc *guc, struct xe_exec_queue *q)
-{
- /*
- * Might be done from within the GPU scheduler, need to do async as we
- * fini the scheduler when the engine is fini'd, the scheduler can't
- * complete fini within itself (circular dependency). Async resolves
- * this we and don't really care when everything is fini'd, just that it
- * is.
- */
- guc_exec_queue_destroy_async(q);
+ queue_work(xe->destroy_wq, &q->guc->destroy_async);
}
static void __guc_exec_queue_process_msg_cleanup(struct xe_sched_msg *msg)
@@ -1770,7 +1720,7 @@ static void __guc_exec_queue_process_msg_cleanup(struct xe_sched_msg *msg)
if (exec_queue_registered(q) && xe_uc_fw_is_running(&guc->fw))
disable_scheduling_deregister(guc, q);
else
- __guc_exec_queue_destroy(guc, q);
+ guc_exec_queue_destroy_async(q);
}
static bool guc_exec_queue_allowed_to_change_state(struct xe_exec_queue *q)
@@ -1961,10 +1911,24 @@ static void guc_exec_queue_process_msg(struct xe_sched_msg *msg)
xe_pm_runtime_put(xe);
}
-static const struct drm_sched_backend_ops drm_sched_ops = {
+static void guc_dep_queue_release(struct drm_dep_queue *q)
+{
+ struct xe_gpu_scheduler *sched =
+ container_of(q, typeof(*sched), base);
+ struct xe_guc_exec_queue *ge =
+ container_of(sched, typeof(*ge), sched);
+
+ /*
+ * RCU free: the dep queue's name may be referenced by exported dma
+ * fences (timeline name). Defer freeing until after any RCU readers.
+ */
+ kfree_rcu(ge, sched.base.rcu);
+}
+
+static const struct drm_dep_queue_ops guc_dep_queue_ops = {
.run_job = guc_exec_queue_run_job,
- .free_job = guc_exec_queue_free_job,
.timedout_job = guc_exec_queue_timedout_job,
+ .release = guc_dep_queue_release,
};
static const struct xe_sched_backend_ops xe_sched_ops = {
@@ -1977,6 +1941,7 @@ static int guc_exec_queue_init(struct xe_exec_queue *q)
struct xe_guc *guc = exec_queue_to_guc(q);
struct workqueue_struct *submit_wq = NULL;
struct xe_guc_exec_queue *ge;
+ enum drm_dep_queue_flags flags = DRM_DEP_QUEUE_FLAGS_BYPASS_SUPPORTED;
long timeout;
int err, i;
@@ -1988,7 +1953,6 @@ static int guc_exec_queue_init(struct xe_exec_queue *q)
q->guc = ge;
ge->q = q;
- init_rcu_head(&ge->rcu);
init_waitqueue_head(&ge->suspend_wait);
for (i = 0; i < MAX_STATIC_MSG_TYPE; ++i)
@@ -2005,35 +1969,29 @@ static int guc_exec_queue_init(struct xe_exec_queue *q)
if (xe_exec_queue_is_multi_queue_secondary(q)) {
struct xe_exec_queue *primary = xe_exec_queue_multi_queue_primary(q);
- submit_wq = primary->guc->sched.base.submit_wq;
+ submit_wq = drm_dep_queue_submit_wq(&primary->guc->sched.base);
}
- err = xe_sched_init(&ge->sched, &drm_sched_ops, &xe_sched_ops,
- submit_wq, xe_lrc_ring_size() / MAX_JOB_SIZE_BYTES, 64,
- timeout, guc_to_gt(guc)->ordered_wq, NULL,
- q->name, gt_to_xe(q->gt)->drm.dev);
+ err = xe_sched_init(&ge->sched, &guc_dep_queue_ops, &xe_sched_ops,
+ submit_wq, xe_lrc_ring_size() / MAX_JOB_SIZE_BYTES,
+ timeout, guc_to_gt(guc)->ordered_wq, flags,
+ q->name, >_to_xe(q->gt)->drm);
if (err)
goto err_free;
sched = &ge->sched;
- err = xe_sched_entity_init(&ge->entity, sched);
- if (err)
- goto err_sched;
mutex_lock(&guc->submission_state.lock);
err = alloc_guc_id(guc, q);
if (err)
- goto err_entity;
+ goto err_sched;
- q->entity = &ge->entity;
+ /* dep_q IS the queue: ge->sched.base is the drm_dep_queue */
+ q->dep_q = &ge->sched.base;
if (xe_guc_read_stopped(guc) || vf_recovery(guc))
- xe_sched_stop(sched);
-
- mutex_unlock(&guc->submission_state.lock);
-
- xe_exec_queue_assign_name(q, q->guc->id);
+ xe_sched_submission_stop(sched);
/*
* Maintain secondary queues of the multi queue group in a list
@@ -2045,11 +2003,15 @@ static int guc_exec_queue_init(struct xe_exec_queue *q)
INIT_LIST_HEAD(&q->multi_queue.link);
mutex_lock(&group->list_lock);
if (group->stopped)
- WRITE_ONCE(q->guc->sched.base.pause_submit, true);
+ drm_dep_queue_set_stopped(&q->guc->sched.base);
list_add_tail(&q->multi_queue.link, &group->list);
mutex_unlock(&group->list_lock);
}
+ mutex_unlock(&guc->submission_state.lock);
+
+ xe_exec_queue_assign_name(q, q->guc->id);
+
if (xe_exec_queue_is_multi_queue(q))
trace_xe_exec_queue_create_multi_queue(q);
else
@@ -2057,11 +2019,11 @@ static int guc_exec_queue_init(struct xe_exec_queue *q)
return 0;
-err_entity:
- mutex_unlock(&guc->submission_state.lock);
- xe_sched_entity_fini(&ge->entity);
err_sched:
+ mutex_unlock(&guc->submission_state.lock);
xe_sched_fini(&ge->sched);
+
+ return err;
err_free:
kfree(ge);
@@ -2126,7 +2088,7 @@ static void guc_exec_queue_destroy(struct xe_exec_queue *q)
if (!(q->flags & EXEC_QUEUE_FLAG_PERMANENT) && !exec_queue_wedged(q))
guc_exec_queue_add_msg(q, msg, CLEANUP);
else
- __guc_exec_queue_destroy(exec_queue_to_guc(q), q);
+ __guc_exec_queue_destroy(q);
}
static int guc_exec_queue_set_priority(struct xe_exec_queue *q,
@@ -2373,7 +2335,7 @@ static void guc_exec_queue_stop(struct xe_guc *guc, struct xe_exec_queue *q)
}
if (do_destroy)
- __guc_exec_queue_destroy(guc, q);
+ guc_exec_queue_destroy_async(q);
}
static int guc_submit_reset_prepare(struct xe_guc *guc)
@@ -2519,7 +2481,7 @@ static void guc_exec_queue_pause(struct xe_guc *guc, struct xe_exec_queue *q)
/* Stop scheduling + flush any DRM scheduler operations */
xe_sched_submission_stop(sched);
- cancel_delayed_work_sync(&sched->base.work_tdr);
+ drm_dep_queue_cancel_tdr_sync(&sched->base);
guc_exec_queue_revert_pending_state_change(guc, q);
@@ -2647,11 +2609,11 @@ static void guc_exec_queue_unpause_prepare(struct xe_guc *guc,
{
struct xe_gpu_scheduler *sched = &q->guc->sched;
struct xe_sched_job *job = NULL;
- struct drm_sched_job *s_job;
+ struct drm_dep_job *dep_job;
bool restore_replay = false;
- drm_sched_for_each_pending_job(s_job, &sched->base, NULL) {
- job = to_xe_sched_job(s_job);
+ drm_dep_queue_for_each_pending_job(dep_job, &sched->base) {
+ job = to_xe_sched_job(dep_job);
restore_replay |= job->restore_replay;
if (restore_replay) {
xe_gt_dbg(guc_to_gt(guc), "Replay JOB - guc_id=%d, seqno=%d",
@@ -2775,7 +2737,7 @@ void xe_guc_submit_unpause_vf(struct xe_guc *guc)
* created after resfix done.
*/
if (q->guc->id != index ||
- !drm_sched_is_stopped(&q->guc->sched.base))
+ !drm_dep_queue_is_stopped(&q->guc->sched.base))
continue;
guc_exec_queue_unpause(guc, q);
@@ -2938,7 +2900,7 @@ static void handle_deregister_done(struct xe_guc *guc, struct xe_exec_queue *q)
trace_xe_exec_queue_deregister_done(q);
clear_exec_queue_registered(q);
- __guc_exec_queue_destroy(guc, q);
+ guc_exec_queue_destroy_async(q);
}
int xe_guc_deregister_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
@@ -3243,8 +3205,8 @@ xe_guc_exec_queue_snapshot_capture(struct xe_exec_queue *q)
snapshot->class = q->class;
snapshot->logical_mask = q->logical_mask;
snapshot->width = q->width;
- snapshot->refcount = kref_read(&q->refcount);
- snapshot->sched_timeout = sched->base.timeout;
+ snapshot->refcount = drm_dep_queue_refcount(&sched->base);
+ snapshot->sched_timeout = drm_dep_queue_timeout(&sched->base);
snapshot->sched_props.timeslice_us = q->sched_props.timeslice_us;
snapshot->sched_props.preempt_timeout_us =
q->sched_props.preempt_timeout_us;
diff --git a/drivers/gpu/drm/xe/xe_migrate.c b/drivers/gpu/drm/xe/xe_migrate.c
index 519f7c70abfb..565054ba0c34 100644
--- a/drivers/gpu/drm/xe/xe_migrate.c
+++ b/drivers/gpu/drm/xe/xe_migrate.c
@@ -2279,7 +2279,7 @@ static struct dma_fence *xe_migrate_vram(struct xe_migrate *m,
if (deps && !dma_fence_is_signaled(deps)) {
dma_fence_get(deps);
- err = drm_sched_job_add_dependency(&job->drm, deps);
+ err = drm_dep_job_add_dependency(&job->drm, deps);
if (err)
dma_fence_wait(deps, false);
err = 0;
diff --git a/drivers/gpu/drm/xe/xe_pt.c b/drivers/gpu/drm/xe/xe_pt.c
index 13b355fadd58..24374a3459c2 100644
--- a/drivers/gpu/drm/xe/xe_pt.c
+++ b/drivers/gpu/drm/xe/xe_pt.c
@@ -1322,7 +1322,7 @@ static int xe_pt_vm_dependencies(struct xe_sched_job *job,
return -ETIME;
dma_fence_get(fence);
- err = drm_sched_job_add_dependency(&job->drm, fence);
+ err = drm_dep_job_add_dependency(&job->drm, fence);
if (err)
return err;
}
diff --git a/drivers/gpu/drm/xe/xe_sched_job.c b/drivers/gpu/drm/xe/xe_sched_job.c
index 99f11bb4d2b9..6b83618e82aa 100644
--- a/drivers/gpu/drm/xe/xe_sched_job.c
+++ b/drivers/gpu/drm/xe/xe_sched_job.c
@@ -21,6 +21,12 @@
#include "xe_trace.h"
#include "xe_vm.h"
+static void xe_sched_job_release(struct drm_dep_job *dep_job);
+
+static const struct drm_dep_job_ops xe_sched_job_dep_ops = {
+ .release = xe_sched_job_release,
+};
+
static struct kmem_cache *xe_sched_job_slab;
static struct kmem_cache *xe_sched_job_parallel_slab;
@@ -109,15 +115,20 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q,
if (!job)
return ERR_PTR(-ENOMEM);
+ err = drm_dep_job_init(&job->drm,
+ &(const struct drm_dep_job_init_args){
+ .ops = &xe_sched_job_dep_ops,
+ .q = q->dep_q,
+ .credits = 1,
+ });
+ if (err)
+ goto err_free;
+
job->q = q;
job->sample_timestamp = U64_MAX;
- kref_init(&job->refcount);
xe_exec_queue_get(job->q);
-
- err = drm_sched_job_init(&job->drm, q->entity, 1, NULL,
- q->xef ? q->xef->drm->client_id : 0);
- if (err)
- goto err_free;
+ atomic_inc(&q->job_cnt);
+ xe_pm_runtime_get_noresume(job_to_xe(job));
for (i = 0; i < q->width; ++i) {
struct dma_fence *fence = xe_lrc_alloc_seqno_fence();
@@ -147,37 +158,34 @@ struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q,
for (i = 0; i < width; ++i)
job->ptrs[i].batch_addr = batch_addr[i];
- atomic_inc(&q->job_cnt);
- xe_pm_runtime_get_noresume(job_to_xe(job));
trace_xe_sched_job_create(job);
return job;
err_sched_job:
- xe_sched_job_free_fences(job);
- drm_sched_job_cleanup(&job->drm);
+ drm_dep_job_put(&job->drm);
+ return ERR_PTR(err);
+
err_free:
- xe_exec_queue_put(q);
job_free(job);
return ERR_PTR(err);
}
/**
- * xe_sched_job_destroy - Destroy Xe schedule job
- * @ref: reference to Xe schedule job
+ * xe_sched_job_release - Release Xe schedule job
+ * @dep_job: base DRM dependency job
*
* Called when ref == 0, drop a reference to job's xe_engine + fence, cleanup
- * base DRM schedule job, and free memory for Xe schedule job.
+ * and free memory for Xe schedule job.
*/
-void xe_sched_job_destroy(struct kref *ref)
+static void xe_sched_job_release(struct drm_dep_job *dep_job)
{
struct xe_sched_job *job =
- container_of(ref, struct xe_sched_job, refcount);
+ container_of(dep_job, struct xe_sched_job, drm);
struct xe_device *xe = job_to_xe(job);
struct xe_exec_queue *q = job->q;
xe_sched_job_free_fences(job);
dma_fence_put(job->fence);
- drm_sched_job_cleanup(&job->drm);
job_free(job);
atomic_dec(&q->job_cnt);
xe_exec_queue_put(q);
@@ -214,7 +222,6 @@ void xe_sched_job_set_error(struct xe_sched_job *job, int error)
trace_xe_sched_job_set_error(job);
- dma_fence_enable_sw_signaling(job->fence);
xe_hw_fence_irq_run(job->q->fence_irq);
}
@@ -287,16 +294,15 @@ struct dma_fence *xe_sched_job_arm(struct xe_sched_job *job)
}
job->fence = dma_fence_get(fence); /* Pairs with put in scheduler */
- drm_sched_job_arm(&job->drm);
+ drm_dep_job_arm(&job->drm);
- return &job->drm.s_fence->finished;
+ return drm_dep_job_finished_fence(&job->drm);
}
void xe_sched_job_push(struct xe_sched_job *job)
{
- xe_sched_job_get(job);
trace_xe_sched_job_exec(job);
- drm_sched_entity_push_job(&job->drm);
+ drm_dep_job_push(&job->drm);
}
/**
@@ -357,5 +363,5 @@ xe_sched_job_snapshot_print(struct xe_sched_job_snapshot *snapshot,
int xe_sched_job_add_deps(struct xe_sched_job *job, struct dma_resv *resv,
enum dma_resv_usage usage)
{
- return drm_sched_job_add_resv_dependencies(&job->drm, resv, usage);
+ return drm_dep_job_add_resv_dependencies(&job->drm, resv, usage);
}
diff --git a/drivers/gpu/drm/xe/xe_sched_job.h b/drivers/gpu/drm/xe/xe_sched_job.h
index a39cc4ab980b..bdd0305970b0 100644
--- a/drivers/gpu/drm/xe/xe_sched_job.h
+++ b/drivers/gpu/drm/xe/xe_sched_job.h
@@ -20,7 +20,6 @@ void xe_sched_job_module_exit(void);
struct xe_sched_job *xe_sched_job_create(struct xe_exec_queue *q,
u64 *batch_addr);
-void xe_sched_job_destroy(struct kref *ref);
/**
* xe_sched_job_get - get reference to Xe schedule job
@@ -30,7 +29,7 @@ void xe_sched_job_destroy(struct kref *ref);
*/
static inline struct xe_sched_job *xe_sched_job_get(struct xe_sched_job *job)
{
- kref_get(&job->refcount);
+ drm_dep_job_get(&job->drm);
return job;
}
@@ -43,7 +42,7 @@ static inline struct xe_sched_job *xe_sched_job_get(struct xe_sched_job *job)
*/
static inline void xe_sched_job_put(struct xe_sched_job *job)
{
- kref_put(&job->refcount, xe_sched_job_destroy);
+ drm_dep_job_put(&job->drm);
}
void xe_sched_job_set_error(struct xe_sched_job *job, int error);
@@ -62,7 +61,7 @@ void xe_sched_job_init_user_fence(struct xe_sched_job *job,
struct xe_sync_entry *sync);
static inline struct xe_sched_job *
-to_xe_sched_job(struct drm_sched_job *drm)
+to_xe_sched_job(struct drm_dep_job *drm)
{
return container_of(drm, struct xe_sched_job, drm);
}
diff --git a/drivers/gpu/drm/xe/xe_sched_job_types.h b/drivers/gpu/drm/xe/xe_sched_job_types.h
index 13c2970e81a8..6b6189f58fd2 100644
--- a/drivers/gpu/drm/xe/xe_sched_job_types.h
+++ b/drivers/gpu/drm/xe/xe_sched_job_types.h
@@ -8,7 +8,7 @@
#include <linux/kref.h>
-#include <drm/gpu_scheduler.h>
+#include <drm/drm_dep.h>
struct xe_exec_queue;
struct dma_fence;
@@ -35,12 +35,10 @@ struct xe_job_ptrs {
* struct xe_sched_job - Xe schedule job (batch buffer tracking)
*/
struct xe_sched_job {
- /** @drm: base DRM scheduler job */
- struct drm_sched_job drm;
+ /** @drm: base DRM dependency job */
+ struct drm_dep_job drm;
/** @q: Exec queue */
struct xe_exec_queue *q;
- /** @refcount: ref count of this job */
- struct kref refcount;
/**
* @fence: dma fence to indicate completion. 1 way relationship - job
* can safely reference fence, fence cannot safely reference job.
diff --git a/drivers/gpu/drm/xe/xe_sync.c b/drivers/gpu/drm/xe/xe_sync.c
index 24d6d9af20d6..cbe9c0adc9a2 100644
--- a/drivers/gpu/drm/xe/xe_sync.c
+++ b/drivers/gpu/drm/xe/xe_sync.c
@@ -234,7 +234,7 @@ ALLOW_ERROR_INJECTION(xe_sync_entry_parse, ERRNO);
int xe_sync_entry_add_deps(struct xe_sync_entry *sync, struct xe_sched_job *job)
{
if (sync->fence)
- return drm_sched_job_add_dependency(&job->drm,
+ return drm_dep_job_add_dependency(&job->drm,
dma_fence_get(sync->fence));
return 0;
diff --git a/drivers/gpu/drm/xe/xe_tlb_inval_job.c b/drivers/gpu/drm/xe/xe_tlb_inval_job.c
index 04d21015cd5d..71cf8fcd99ba 100644
--- a/drivers/gpu/drm/xe/xe_tlb_inval_job.c
+++ b/drivers/gpu/drm/xe/xe_tlb_inval_job.c
@@ -65,18 +65,14 @@ static struct dma_fence *xe_tlb_inval_job_run(struct xe_dep_job *dep_job)
return job->fence;
}
-static void xe_tlb_inval_job_free(struct xe_dep_job *dep_job)
-{
- struct xe_tlb_inval_job *job =
- container_of(dep_job, typeof(*job), dep);
-
- /* Pairs with get in xe_tlb_inval_job_push */
- xe_tlb_inval_job_put(job);
-}
-
static const struct xe_dep_job_ops dep_job_ops = {
.run_job = xe_tlb_inval_job_run,
- .free_job = xe_tlb_inval_job_free,
+};
+
+static void xe_tlb_inval_job_destroy(struct drm_dep_job *drm_job);
+
+static const struct drm_dep_job_ops xe_tlb_inval_job_dep_ops = {
+ .release = xe_tlb_inval_job_destroy,
};
/**
@@ -100,8 +96,8 @@ xe_tlb_inval_job_create(struct xe_exec_queue *q, struct xe_tlb_inval *tlb_inval,
struct xe_vm *vm, u64 start, u64 end, int type)
{
struct xe_tlb_inval_job *job;
- struct drm_sched_entity *entity =
- xe_dep_scheduler_entity(dep_scheduler);
+ struct drm_dep_queue *dep_q =
+ xe_dep_scheduler_dep_q(dep_scheduler);
struct xe_tlb_inval_fence *ifence;
int err;
@@ -121,7 +117,6 @@ xe_tlb_inval_job_create(struct xe_exec_queue *q, struct xe_tlb_inval *tlb_inval,
xe_page_reclaim_list_init(&job->prl);
job->dep.ops = &dep_job_ops;
job->type = type;
- kref_init(&job->refcount);
xe_exec_queue_get(q); /* Pairs with put in xe_tlb_inval_job_destroy */
xe_vm_get(vm); /* Pairs with put in xe_tlb_inval_job_destroy */
@@ -132,8 +127,12 @@ xe_tlb_inval_job_create(struct xe_exec_queue *q, struct xe_tlb_inval *tlb_inval,
}
job->fence = &ifence->base;
- err = drm_sched_job_init(&job->dep.drm, entity, 1, NULL,
- q->xef ? q->xef->drm->client_id : 0);
+ err = drm_dep_job_init(&job->dep.drm,
+ &(const struct drm_dep_job_init_args){
+ .ops = &xe_tlb_inval_job_dep_ops,
+ .q = dep_q,
+ .credits = 1,
+ });
if (err)
goto err_fence;
@@ -171,10 +170,10 @@ void xe_tlb_inval_job_add_page_reclaim(struct xe_tlb_inval_job *job,
xe_page_reclaim_entries_get(job->prl.entries);
}
-static void xe_tlb_inval_job_destroy(struct kref *ref)
+static void xe_tlb_inval_job_destroy(struct drm_dep_job *drm_job)
{
- struct xe_tlb_inval_job *job = container_of(ref, typeof(*job),
- refcount);
+ struct xe_tlb_inval_job *job = container_of(drm_job, typeof(*job),
+ dep.drm);
struct xe_tlb_inval_fence *ifence =
container_of(job->fence, typeof(*ifence), base);
struct xe_exec_queue *q = job->q;
@@ -190,7 +189,6 @@ static void xe_tlb_inval_job_destroy(struct kref *ref)
/* Ref from xe_tlb_inval_fence_init */
dma_fence_put(job->fence);
- drm_sched_job_cleanup(&job->dep.drm);
kfree(job);
xe_vm_put(vm); /* Pairs with get from xe_tlb_inval_job_create */
xe_exec_queue_put(q); /* Pairs with get from xe_tlb_inval_job_create */
@@ -209,11 +207,19 @@ static void xe_tlb_inval_job_destroy(struct kref *ref)
*/
int xe_tlb_inval_job_alloc_dep(struct xe_tlb_inval_job *job)
{
- xe_assert(gt_to_xe(job->q->gt), !xa_load(&job->dep.drm.dependencies, 0));
+ int ret;
+
might_alloc(GFP_KERNEL);
- return drm_sched_job_add_dependency(&job->dep.drm,
- dma_fence_get_stub());
+ ret = drm_dep_job_add_dependency(&job->dep.drm,
+ DRM_DEP_JOB_FENCE_PREALLOC);
+ if (ret < 0)
+ return ret;
+
+ /* Assert allocation slot is zero */
+ xe_assert(gt_to_xe(job->q->gt), !ret);
+
+ return 0;
}
/**
@@ -236,26 +242,14 @@ struct dma_fence *xe_tlb_inval_job_push(struct xe_tlb_inval_job *job,
{
struct xe_tlb_inval_fence *ifence =
container_of(job->fence, typeof(*ifence), base);
+ struct dma_fence *stub = dma_fence_get_stub();
- if (!dma_fence_is_signaled(fence)) {
- void *ptr;
-
- /*
- * Can be in path of reclaim, hence the preallocation of fence
- * storage in xe_tlb_inval_job_alloc_dep. Verify caller did
- * this correctly.
- */
- xe_assert(gt_to_xe(job->q->gt),
- xa_load(&job->dep.drm.dependencies, 0) ==
- dma_fence_get_stub());
-
+ if (fence != stub) {
dma_fence_get(fence); /* ref released once dependency processed by scheduler */
- ptr = xa_store(&job->dep.drm.dependencies, 0, fence,
- GFP_ATOMIC);
- xe_assert(gt_to_xe(job->q->gt), !xa_is_err(ptr));
+ drm_dep_job_replace_dependency(&job->dep.drm, 0, fence);
}
+ dma_fence_put(stub);
- xe_tlb_inval_job_get(job); /* Pairs with put in free_job */
job->fence_armed = true;
/*
@@ -269,17 +263,17 @@ struct dma_fence *xe_tlb_inval_job_push(struct xe_tlb_inval_job *job,
xe_tlb_inval_fence_init(job->tlb_inval, ifence, false);
dma_fence_get(job->fence); /* Pairs with put in DRM scheduler */
- drm_sched_job_arm(&job->dep.drm);
+ drm_dep_job_arm(&job->dep.drm);
+ fence = drm_dep_job_finished_fence(&job->dep.drm);
/*
* caller ref, get must be done before job push as it could immediately
* signal and free.
*/
- dma_fence_get(&job->dep.drm.s_fence->finished);
- drm_sched_entity_push_job(&job->dep.drm);
+ dma_fence_get(fence);
+ drm_dep_job_push(&job->dep.drm);
/* Let the upper layers fish this out */
- xe_exec_queue_tlb_inval_last_fence_set(job->q, job->vm,
- &job->dep.drm.s_fence->finished,
+ xe_exec_queue_tlb_inval_last_fence_set(job->q, job->vm, fence,
job->type);
xe_migrate_job_unlock(m, job->q);
@@ -290,7 +284,7 @@ struct dma_fence *xe_tlb_inval_job_push(struct xe_tlb_inval_job *job,
* be squashed in dma-resv/DRM scheduler. Instead, we use the DRM scheduler
* context and job's finished fence, which enables squashing.
*/
- return &job->dep.drm.s_fence->finished;
+ return fence;
}
/**
@@ -301,7 +295,7 @@ struct dma_fence *xe_tlb_inval_job_push(struct xe_tlb_inval_job *job,
*/
void xe_tlb_inval_job_get(struct xe_tlb_inval_job *job)
{
- kref_get(&job->refcount);
+ drm_dep_job_get(&job->dep.drm);
}
/**
@@ -315,5 +309,5 @@ void xe_tlb_inval_job_get(struct xe_tlb_inval_job *job)
void xe_tlb_inval_job_put(struct xe_tlb_inval_job *job)
{
if (!IS_ERR_OR_NULL(job))
- kref_put(&job->refcount, xe_tlb_inval_job_destroy);
+ drm_dep_job_put(&job->dep.drm);
}
--
2.34.1
next prev parent reply other threads:[~2026-03-16 4:33 UTC|newest]
Thread overview: 65+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-03-16 4:32 [RFC PATCH 00/12] Introduce DRM dep queue Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 01/12] workqueue: Add interface to teach lockdep to warn on reclaim violations Matthew Brost
2026-03-25 15:59 ` Tejun Heo
2026-03-26 1:49 ` Matthew Brost
2026-03-26 2:19 ` Tejun Heo
2026-03-27 4:33 ` Matthew Brost
2026-03-27 17:25 ` Tejun Heo
2026-03-16 4:32 ` [RFC PATCH 02/12] drm/dep: Add DRM dependency queue layer Matthew Brost
2026-03-16 9:16 ` Boris Brezillon
2026-03-17 5:22 ` Matthew Brost
2026-03-17 8:48 ` Boris Brezillon
2026-03-16 10:25 ` Danilo Krummrich
2026-03-17 5:10 ` Matthew Brost
2026-03-17 12:19 ` Danilo Krummrich
2026-03-18 23:02 ` Matthew Brost
2026-03-17 2:47 ` Daniel Almeida
2026-03-17 5:45 ` Matthew Brost
2026-03-17 7:17 ` Miguel Ojeda
2026-03-17 8:26 ` Matthew Brost
2026-03-17 12:04 ` Daniel Almeida
2026-03-17 19:41 ` Miguel Ojeda
2026-03-23 17:31 ` Matthew Brost
2026-03-23 17:42 ` Miguel Ojeda
2026-03-17 18:14 ` Matthew Brost
2026-03-17 19:48 ` Daniel Almeida
2026-03-17 20:43 ` Boris Brezillon
2026-03-18 22:40 ` Matthew Brost
2026-03-19 9:57 ` Boris Brezillon
2026-03-22 6:43 ` Matthew Brost
2026-03-23 7:58 ` Matthew Brost
2026-03-23 10:06 ` Boris Brezillon
2026-03-23 17:11 ` Matthew Brost
2026-03-17 12:31 ` Danilo Krummrich
2026-03-17 14:25 ` Daniel Almeida
2026-03-17 14:33 ` Danilo Krummrich
2026-03-18 22:50 ` Matthew Brost
2026-03-17 8:47 ` Christian König
2026-03-17 14:55 ` Boris Brezillon
2026-03-18 23:28 ` Matthew Brost
2026-03-19 9:11 ` Boris Brezillon
2026-03-23 4:50 ` Matthew Brost
2026-03-23 9:55 ` Boris Brezillon
2026-03-23 17:08 ` Matthew Brost
2026-03-23 18:38 ` Matthew Brost
2026-03-24 9:23 ` Boris Brezillon
2026-03-24 16:06 ` Matthew Brost
2026-03-25 2:33 ` Matthew Brost
2026-03-24 8:49 ` Boris Brezillon
2026-03-24 16:51 ` Matthew Brost
2026-03-17 16:30 ` Shashank Sharma
2026-03-16 4:32 ` [RFC PATCH 03/12] drm/xe: Use WQ_MEM_WARN_ON_RECLAIM on all workqueues in the reclaim path Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 04/12] drm/xe: Issue GGTT invalidation under lock in ggtt_node_remove Matthew Brost
2026-03-26 5:45 ` Bhadane, Dnyaneshwar
2026-03-16 4:32 ` [RFC PATCH 05/12] drm/xe: Return fence from xe_sched_job_arm and adjust job references Matthew Brost
2026-03-16 4:32 ` Matthew Brost [this message]
2026-03-16 4:32 ` [RFC PATCH 07/12] drm/xe: Make scheduler message lock IRQ-safe Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 08/12] drm/xe: Rework exec queue object on top of DRM dep Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 09/12] drm/xe: Enable IRQ job put in " Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 10/12] drm/xe: Use DRM dep queue kill semantics Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 11/12] accel/amdxdna: Convert to drm_dep scheduler layer Matthew Brost
2026-03-16 4:32 ` [RFC PATCH 12/12] drm/panthor: " Matthew Brost
2026-03-16 4:52 ` ✗ CI.checkpatch: warning for Introduce DRM dep queue Patchwork
2026-03-16 4:53 ` ✓ CI.KUnit: success " Patchwork
2026-03-16 5:28 ` ✓ Xe.CI.BAT: " Patchwork
2026-03-16 8:09 ` ✗ Xe.CI.FULL: failure " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260316043255.226352-7-matthew.brost@intel.com \
--to=matthew.brost@intel.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=intel-xe@lists.freedesktop.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox