From: Matthew Brost <matthew.brost@intel.com>
To: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
Cc: <intel-xe@lists.freedesktop.org>
Subject: Re: [PATCH v3 02/18] drm/xe/multi_queue: Add user interface for multi queue support
Date: Fri, 21 Nov 2025 14:51:21 -0800 [thread overview]
Message-ID: <aSDs6c+wt5B2E61C@lstrano-desk.jf.intel.com> (raw)
In-Reply-To: <20251121035147.766072-22-niranjana.vishwanathapura@intel.com>
On Thu, Nov 20, 2025 at 07:51:36PM -0800, Niranjana Vishwanathapura wrote:
> Multi Queue is a new mode of execution supported by the compute and
> blitter copy command streamers (CCS and BCS, respectively). It is an
> enhancement of the existing hardware architecture and leverages the
> same submission model. It enables support for efficient, parallel
> execution of multiple queues within a single context. All the queues
> of a group must use the same address space (VM).
>
> The new DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_QUEUE execution queue
> property supports creating a multi queue group and adding queues to
> a queue group. All queues of a multi queue group share the same
> context.
>
> A exec queue create ioctl call with above property specified with value
> DRM_XE_SUPER_GROUP_CREATE will create a new multi queue group with the
> queue being created as the primary queue (aka q0) of the group. To add
> secondary queues to the group, they need to be created with the above
> property with id of the primary queue as the value. The properties of
> the primary queue (like priority, timeslice) applies to the whole group.
> So, these properties can't be set for secondary queues of a group.
>
> Once destroyed, the secondary queues of a multi queue group can't be
> replaced. However, they can be dynamically added to the group up to a
> total of 64 queues per group. Once the primary queue is destroyed,
> secondary queues can't be added to the queue group.
>
> v2: Remove group->lock, fix xe_exec_queue_group_add()/delete()
> function semantics, add additional comments, remove unused
> group->list_lock, add XE_BO_FLAG_GGTT_INVALIDATE for cgp bo,
> Assert LRC is valid, update uapi kernel doc.
> (Matt Brost)
> v3: Use XE_BO_FLAG_PINNED_LATE_RESTORE/USER_VRAM/GGTT_INVALIDATE
> flags for cgp bo (Matt)
>
> Signed-off-by: Stuart Summers <stuart.summers@intel.com>
> Signed-off-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
Reviewed-by: Matthew Brost <matthew.brost@intel.com>
As this is new uAPI we will need a UMD PR and ack from a UMD team.
Matt
> ---
> drivers/gpu/drm/xe/xe_exec_queue.c | 194 ++++++++++++++++++++++-
> drivers/gpu/drm/xe/xe_exec_queue.h | 47 ++++++
> drivers/gpu/drm/xe/xe_exec_queue_types.h | 26 +++
> include/uapi/drm/xe_drm.h | 10 ++
> 4 files changed, 275 insertions(+), 2 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.c b/drivers/gpu/drm/xe/xe_exec_queue.c
> index 8724f8de67e2..330b5103222c 100644
> --- a/drivers/gpu/drm/xe/xe_exec_queue.c
> +++ b/drivers/gpu/drm/xe/xe_exec_queue.c
> @@ -13,6 +13,7 @@
> #include <drm/drm_syncobj.h>
> #include <uapi/drm/xe_drm.h>
>
> +#include "xe_bo.h"
> #include "xe_dep_scheduler.h"
> #include "xe_device.h"
> #include "xe_gt.h"
> @@ -63,6 +64,33 @@ enum xe_exec_queue_sched_prop {
> static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q,
> u64 extensions, int ext_number);
>
> +static void xe_exec_queue_group_cleanup(struct xe_exec_queue *q)
> +{
> + struct xe_exec_queue_group *group = q->multi_queue.group;
> + struct xe_lrc *lrc;
> + unsigned long idx;
> +
> + if (xe_exec_queue_is_multi_queue_secondary(q)) {
> + /*
> + * Put pairs with get from xe_exec_queue_lookup() call
> + * in xe_exec_queue_group_validate().
> + */
> + xe_exec_queue_put(xe_exec_queue_multi_queue_primary(q));
> + return;
> + }
> +
> + if (!group)
> + return;
> +
> + /* Primary queue cleanup */
> + xa_for_each(&group->xa, idx, lrc)
> + xe_lrc_put(lrc);
> +
> + xa_destroy(&group->xa);
> + xe_bo_unpin_map_no_vm(group->cgp_bo);
> + kfree(group);
> +}
> +
> static void __xe_exec_queue_free(struct xe_exec_queue *q)
> {
> int i;
> @@ -73,6 +101,10 @@ static void __xe_exec_queue_free(struct xe_exec_queue *q)
>
> if (xe_exec_queue_uses_pxp(q))
> xe_pxp_exec_queue_remove(gt_to_xe(q->gt)->pxp, q);
> +
> + if (xe_exec_queue_is_multi_queue(q))
> + xe_exec_queue_group_cleanup(q);
> +
> if (q->vm)
> xe_vm_put(q->vm);
>
> @@ -567,6 +599,147 @@ exec_queue_set_pxp_type(struct xe_device *xe, struct xe_exec_queue *q, u64 value
> return xe_pxp_exec_queue_set_type(xe->pxp, q, DRM_XE_PXP_TYPE_HWDRM);
> }
>
> +static int xe_exec_queue_group_init(struct xe_device *xe, struct xe_exec_queue *q)
> +{
> + struct xe_tile *tile = gt_to_tile(q->gt);
> + struct xe_exec_queue_group *group;
> + struct xe_bo *bo;
> +
> + group = kzalloc(sizeof(*group), GFP_KERNEL);
> + if (!group)
> + return -ENOMEM;
> +
> + bo = xe_bo_create_pin_map_novm(xe, tile, SZ_4K, ttm_bo_type_kernel,
> + XE_BO_FLAG_VRAM_IF_DGFX(tile) |
> + XE_BO_FLAG_PINNED_LATE_RESTORE |
> + XE_BO_FLAG_FORCE_USER_VRAM |
> + XE_BO_FLAG_GGTT_INVALIDATE |
> + XE_BO_FLAG_GGTT, false);
> + if (IS_ERR(bo)) {
> + drm_err(&xe->drm, "CGP bo allocation for queue group failed: %ld\n",
> + PTR_ERR(bo));
> + kfree(group);
> + return PTR_ERR(bo);
> + }
> +
> + xe_map_memset(xe, &bo->vmap, 0, 0, SZ_4K);
> +
> + group->primary = q;
> + group->cgp_bo = bo;
> + xa_init_flags(&group->xa, XA_FLAGS_ALLOC1);
> + q->multi_queue.group = group;
> +
> + return 0;
> +}
> +
> +static inline bool xe_exec_queue_supports_multi_queue(struct xe_exec_queue *q)
> +{
> + return q->gt->info.multi_queue_engine_class_mask & BIT(q->class);
> +}
> +
> +static int xe_exec_queue_group_validate(struct xe_device *xe, struct xe_exec_queue *q,
> + u32 primary_id)
> +{
> + struct xe_exec_queue_group *group;
> + struct xe_exec_queue *primary;
> + int ret;
> +
> + /*
> + * Get from below xe_exec_queue_lookup() pairs with put
> + * in xe_exec_queue_group_cleanup().
> + */
> + primary = xe_exec_queue_lookup(q->vm->xef, primary_id);
> + if (XE_IOCTL_DBG(xe, !primary))
> + return -ENOENT;
> +
> + if (XE_IOCTL_DBG(xe, !xe_exec_queue_is_multi_queue_primary(primary)) ||
> + XE_IOCTL_DBG(xe, q->vm != primary->vm) ||
> + XE_IOCTL_DBG(xe, q->logical_mask != primary->logical_mask)) {
> + ret = -EINVAL;
> + goto put_primary;
> + }
> +
> + group = primary->multi_queue.group;
> + q->multi_queue.valid = true;
> + q->multi_queue.group = group;
> +
> + return 0;
> +put_primary:
> + xe_exec_queue_put(primary);
> + return ret;
> +}
> +
> +#define XE_MAX_GROUP_SIZE 64
> +static int xe_exec_queue_group_add(struct xe_device *xe, struct xe_exec_queue *q)
> +{
> + struct xe_exec_queue_group *group = q->multi_queue.group;
> + u32 pos;
> + int err;
> +
> + xe_assert(xe, xe_exec_queue_is_multi_queue_secondary(q));
> +
> + /* Primary queue holds a reference to LRCs of all secondary queues */
> + err = xa_alloc(&group->xa, &pos, xe_lrc_get(q->lrc[0]),
> + XA_LIMIT(1, XE_MAX_GROUP_SIZE - 1), GFP_KERNEL);
> + if (XE_IOCTL_DBG(xe, err)) {
> + xe_lrc_put(q->lrc[0]);
> +
> + /* It is invalid if queue group limit is exceeded */
> + if (err == -EBUSY)
> + err = -EINVAL;
> +
> + return err;
> + }
> +
> + q->multi_queue.pos = pos;
> +
> + return 0;
> +}
> +
> +static void xe_exec_queue_group_delete(struct xe_device *xe, struct xe_exec_queue *q)
> +{
> + struct xe_exec_queue_group *group = q->multi_queue.group;
> + struct xe_lrc *lrc;
> +
> + xe_assert(xe, xe_exec_queue_is_multi_queue_secondary(q));
> +
> + lrc = xa_erase(&group->xa, q->multi_queue.pos);
> + xe_assert(xe, lrc);
> + xe_lrc_put(lrc);
> +}
> +
> +static int exec_queue_set_multi_group(struct xe_device *xe, struct xe_exec_queue *q,
> + u64 value)
> +{
> + if (XE_IOCTL_DBG(xe, !xe_exec_queue_supports_multi_queue(q)))
> + return -ENODEV;
> +
> + if (XE_IOCTL_DBG(xe, !xe_device_uc_enabled(xe)))
> + return -EOPNOTSUPP;
> +
> + if (XE_IOCTL_DBG(xe, xe_exec_queue_is_parallel(q)))
> + return -EINVAL;
> +
> + if (XE_IOCTL_DBG(xe, xe_exec_queue_is_multi_queue(q)))
> + return -EINVAL;
> +
> + if (value & DRM_XE_MULTI_GROUP_CREATE) {
> + if (XE_IOCTL_DBG(xe, value & ~DRM_XE_MULTI_GROUP_CREATE))
> + return -EINVAL;
> +
> + q->multi_queue.valid = true;
> + q->multi_queue.is_primary = true;
> + q->multi_queue.pos = 0;
> + return 0;
> + }
> +
> + /* While adding secondary queues, the upper 32 bits must be 0 */
> + if (XE_IOCTL_DBG(xe, value & (~0ull << 32)))
> + return -EINVAL;
> +
> + return xe_exec_queue_group_validate(xe, q, value);
> +}
> +
> typedef int (*xe_exec_queue_set_property_fn)(struct xe_device *xe,
> struct xe_exec_queue *q,
> u64 value);
> @@ -575,6 +748,7 @@ static const xe_exec_queue_set_property_fn exec_queue_set_property_funcs[] = {
> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY] = exec_queue_set_priority,
> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE] = exec_queue_set_timeslice,
> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE] = exec_queue_set_pxp_type,
> + [DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP] = exec_queue_set_multi_group,
> };
>
> static int exec_queue_user_ext_set_property(struct xe_device *xe,
> @@ -595,7 +769,8 @@ static int exec_queue_user_ext_set_property(struct xe_device *xe,
> XE_IOCTL_DBG(xe, ext.pad) ||
> XE_IOCTL_DBG(xe, ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY &&
> ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE &&
> - ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE))
> + ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE &&
> + ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP))
> return -EINVAL;
>
> idx = array_index_nospec(ext.property, ARRAY_SIZE(exec_queue_set_property_funcs));
> @@ -644,6 +819,12 @@ static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue
> return exec_queue_user_extensions(xe, q, ext.next_extension,
> ++ext_number);
>
> + if (xe_exec_queue_is_multi_queue_primary(q)) {
> + err = xe_exec_queue_group_init(xe, q);
> + if (XE_IOCTL_DBG(xe, err))
> + return err;
> + }
> +
> return 0;
> }
>
> @@ -798,12 +979,18 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
> if (IS_ERR(q))
> return PTR_ERR(q);
>
> + if (xe_exec_queue_is_multi_queue_secondary(q)) {
> + err = xe_exec_queue_group_add(xe, q);
> + if (XE_IOCTL_DBG(xe, err))
> + goto put_exec_queue;
> + }
> +
> if (xe_vm_in_preempt_fence_mode(vm)) {
> q->lr.context = dma_fence_context_alloc(1);
>
> err = xe_vm_add_compute_exec_queue(vm, q);
> if (XE_IOCTL_DBG(xe, err))
> - goto put_exec_queue;
> + goto delete_queue_group;
> }
>
> if (q->vm && q->hwe->hw_engine_group) {
> @@ -826,6 +1013,9 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
>
> kill_exec_queue:
> xe_exec_queue_kill(q);
> +delete_queue_group:
> + if (xe_exec_queue_is_multi_queue_secondary(q))
> + xe_exec_queue_group_delete(xe, q);
> put_exec_queue:
> xe_exec_queue_put(q);
> return err;
> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.h b/drivers/gpu/drm/xe/xe_exec_queue.h
> index fda4d4f9bda8..e6daa40003f2 100644
> --- a/drivers/gpu/drm/xe/xe_exec_queue.h
> +++ b/drivers/gpu/drm/xe/xe_exec_queue.h
> @@ -66,6 +66,53 @@ static inline bool xe_exec_queue_uses_pxp(struct xe_exec_queue *q)
> return q->pxp.type;
> }
>
> +/**
> + * xe_exec_queue_is_multi_queue() - Whether an exec_queue is part of a queue group.
> + * @q: The exec_queue
> + *
> + * Return: True if the exec_queue is part of a queue group, false otherwise.
> + */
> +static inline bool xe_exec_queue_is_multi_queue(struct xe_exec_queue *q)
> +{
> + return q->multi_queue.valid;
> +}
> +
> +/**
> + * xe_exec_queue_is_multi_queue_primary() - Whether an exec_queue is primary queue
> + * of a multi queue group.
> + * @q: The exec_queue
> + *
> + * Return: True if @q is primary queue of a queue group, false otherwise.
> + */
> +static inline bool xe_exec_queue_is_multi_queue_primary(struct xe_exec_queue *q)
> +{
> + return q->multi_queue.is_primary;
> +}
> +
> +/**
> + * xe_exec_queue_is_multi_queue_secondary() - Whether an exec_queue is secondary queue
> + * of a multi queue group.
> + * @q: The exec_queue
> + *
> + * Return: True if @q is secondary queue of a queue group, false otherwise.
> + */
> +static inline bool xe_exec_queue_is_multi_queue_secondary(struct xe_exec_queue *q)
> +{
> + return xe_exec_queue_is_multi_queue(q) && !xe_exec_queue_is_multi_queue_primary(q);
> +}
> +
> +/**
> + * xe_exec_queue_multi_queue_primary() - Get multi queue group's primary queue
> + * @q: The exec_queue
> + *
> + * If @q belongs to a multi queue group, then the primary queue of the group will
> + * be returned. Otherwise, @q will be returned.
> + */
> +static inline struct xe_exec_queue *xe_exec_queue_multi_queue_primary(struct xe_exec_queue *q)
> +{
> + return xe_exec_queue_is_multi_queue(q) ? q->multi_queue.group->primary : q;
> +}
> +
> bool xe_exec_queue_is_lr(struct xe_exec_queue *q);
>
> bool xe_exec_queue_is_idle(struct xe_exec_queue *q);
> diff --git a/drivers/gpu/drm/xe/xe_exec_queue_types.h b/drivers/gpu/drm/xe/xe_exec_queue_types.h
> index 771ffe35cd0c..f429b1952be9 100644
> --- a/drivers/gpu/drm/xe/xe_exec_queue_types.h
> +++ b/drivers/gpu/drm/xe/xe_exec_queue_types.h
> @@ -32,6 +32,20 @@ enum xe_exec_queue_priority {
> XE_EXEC_QUEUE_PRIORITY_COUNT
> };
>
> +/**
> + * struct xe_exec_queue_group - Execution multi queue group
> + *
> + * Contains multi queue group information.
> + */
> +struct xe_exec_queue_group {
> + /** @primary: Primary queue of this group */
> + struct xe_exec_queue *primary;
> + /** @cgp_bo: BO for the Context Group Page */
> + struct xe_bo *cgp_bo;
> + /** @xa: xarray to store LRCs */
> + struct xarray xa;
> +};
> +
> /**
> * struct xe_exec_queue - Execution queue
> *
> @@ -111,6 +125,18 @@ struct xe_exec_queue {
> struct xe_guc_exec_queue *guc;
> };
>
> + /** @multi_queue: Multi queue information */
> + struct {
> + /** @multi_queue.group: Queue group information */
> + struct xe_exec_queue_group *group;
> + /** @multi_queue.pos: Position of queue within the multi-queue group */
> + u8 pos;
> + /** @multi_queue.valid: Queue belongs to a multi queue group */
> + u8 valid:1;
> + /** @multi_queue.is_primary: Is primary queue (Q0) of the group */
> + u8 is_primary:1;
> + } multi_queue;
> +
> /** @sched_props: scheduling properties */
> struct {
> /** @sched_props.timeslice_us: timeslice period in micro-seconds */
> diff --git a/include/uapi/drm/xe_drm.h b/include/uapi/drm/xe_drm.h
> index 47853659a705..4de21e0a4fcf 100644
> --- a/include/uapi/drm/xe_drm.h
> +++ b/include/uapi/drm/xe_drm.h
> @@ -1252,6 +1252,14 @@ struct drm_xe_vm_bind {
> * Given that going into a power-saving state kills PXP HWDRM sessions,
> * runtime PM will be blocked while queues of this type are alive.
> * All PXP queues will be killed if a PXP invalidation event occurs.
> + * - %DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP - Create a multi-queue group
> + * or add secondary queues to a multi-queue group.
> + * If the extension's 'value' field has %DRM_XE_MULTI_GROUP_CREATE flag set,
> + * then a new multi-queue group is created with this queue as the primary queue
> + * (Q0). Otherwise, the queue gets added to the multi-queue group whose primary
> + * queue's exec_queue_id is specified in the lower 32 bits of the 'value' field.
> + * All the other non-relevant bits of extension's 'value' field while adding the
> + * primary or the secondary queues of the group must be set to 0.
> *
> * The example below shows how to use @drm_xe_exec_queue_create to create
> * a simple exec_queue (no parallel submission) of class
> @@ -1292,6 +1300,8 @@ struct drm_xe_exec_queue_create {
> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY 0
> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE 1
> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE 2
> +#define DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP 3
> +#define DRM_XE_MULTI_GROUP_CREATE (1ull << 63)
> /** @extensions: Pointer to the first extension struct, if any */
> __u64 extensions;
>
> --
> 2.43.0
>
next prev parent reply other threads:[~2025-11-21 22:51 UTC|newest]
Thread overview: 41+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-21 3:51 [PATCH v3 00/18] drm/xe: Multi Queue feature support Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 01/18] drm/xe/multi_queue: Add multi_queue_enable_mask to gt information Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 02/18] drm/xe/multi_queue: Add user interface for multi queue support Niranjana Vishwanathapura
2025-11-21 22:51 ` Matthew Brost [this message]
2025-11-22 4:35 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 03/18] drm/xe/multi_queue: Add GuC " Niranjana Vishwanathapura
2025-11-22 22:16 ` Matthew Brost
2025-12-03 3:40 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 04/18] drm/xe/multi_queue: Add multi queue priority property Niranjana Vishwanathapura
2025-11-21 22:57 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 05/18] drm/xe/multi_queue: Handle invalid exec queue property setting Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 06/18] drm/xe/multi_queue: Add exec_queue set_property ioctl support Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 07/18] drm/xe/multi_queue: Add support for multi queue dynamic priority change Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 08/18] drm/xe/multi_queue: Add multi queue information to guc_info dump Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 09/18] drm/xe/multi_queue: Handle tearing down of a multi queue Niranjana Vishwanathapura
2025-11-21 23:03 ` Matthew Brost
2025-11-22 4:40 ` Niranjana Vishwanathapura
2025-11-22 5:47 ` Matthew Brost
2025-12-09 3:31 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 10/18] drm/xe/multi_queue: Set QUEUE_DRAIN_MODE for Multi Queue batches Niranjana Vishwanathapura
2025-11-24 18:49 ` Matt Roper
2025-12-02 21:28 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 11/18] drm/xe/multi_queue: Handle CGP context error Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 12/18] drm/xe/multi_queue: Reset GT upon CGP_SYNC failure Niranjana Vishwanathapura
2025-11-21 23:08 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 13/18] drm/xe/multi_queue: Tracepoint support Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 14/18] drm/xe/multi_queue: Support active group after primary is destroyed Niranjana Vishwanathapura
2025-11-22 5:57 ` Matthew Brost
2025-11-22 6:08 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 15/18] drm/xe/xe3p: Disable GuC Dynamic ICS for Xe3p Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 16/18] drm/xe/doc: Add documentation for Multi Queue Group Niranjana Vishwanathapura
2025-11-22 6:02 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 17/18] drm/xe/doc: Add documentation for Multi Queue Group GuC interface Niranjana Vishwanathapura
2025-11-22 6:10 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 18/18] drm/xe/multi_queue: Enable multi_queue on xe3p_xpc Niranjana Vishwanathapura
2025-11-21 4:01 ` ✗ CI.checkpatch: warning for drm/xe: Multi Queue feature support (rev3) Patchwork
2025-11-21 4:02 ` ✓ CI.KUnit: success " Patchwork
2025-11-21 4:51 ` ✗ Xe.CI.BAT: failure " Patchwork
2025-11-21 8:10 ` ✗ Xe.CI.Full: " Patchwork
2025-11-24 14:04 ` Patchwork
2025-11-27 9:38 ` [PATCH v3 00/18] drm/xe: Multi Queue feature support Hoppe, Mateusz
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aSDs6c+wt5B2E61C@lstrano-desk.jf.intel.com \
--to=matthew.brost@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=niranjana.vishwanathapura@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox