From: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
To: Matthew Brost <matthew.brost@intel.com>
Cc: <intel-xe@lists.freedesktop.org>
Subject: Re: [PATCH v3 02/18] drm/xe/multi_queue: Add user interface for multi queue support
Date: Fri, 21 Nov 2025 20:35:16 -0800 [thread overview]
Message-ID: <aSE9hPYxCNub4RHd@nvishwa1-desk> (raw)
In-Reply-To: <aSDs6c+wt5B2E61C@lstrano-desk.jf.intel.com>
On Fri, Nov 21, 2025 at 02:51:21PM -0800, Matthew Brost wrote:
>On Thu, Nov 20, 2025 at 07:51:36PM -0800, Niranjana Vishwanathapura wrote:
>> Multi Queue is a new mode of execution supported by the compute and
>> blitter copy command streamers (CCS and BCS, respectively). It is an
>> enhancement of the existing hardware architecture and leverages the
>> same submission model. It enables support for efficient, parallel
>> execution of multiple queues within a single context. All the queues
>> of a group must use the same address space (VM).
>>
>> The new DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_QUEUE execution queue
>> property supports creating a multi queue group and adding queues to
>> a queue group. All queues of a multi queue group share the same
>> context.
>>
>> A exec queue create ioctl call with above property specified with value
>> DRM_XE_SUPER_GROUP_CREATE will create a new multi queue group with the
>> queue being created as the primary queue (aka q0) of the group. To add
>> secondary queues to the group, they need to be created with the above
>> property with id of the primary queue as the value. The properties of
>> the primary queue (like priority, timeslice) applies to the whole group.
>> So, these properties can't be set for secondary queues of a group.
>>
>> Once destroyed, the secondary queues of a multi queue group can't be
>> replaced. However, they can be dynamically added to the group up to a
>> total of 64 queues per group. Once the primary queue is destroyed,
>> secondary queues can't be added to the queue group.
>>
>> v2: Remove group->lock, fix xe_exec_queue_group_add()/delete()
>> function semantics, add additional comments, remove unused
>> group->list_lock, add XE_BO_FLAG_GGTT_INVALIDATE for cgp bo,
>> Assert LRC is valid, update uapi kernel doc.
>> (Matt Brost)
>> v3: Use XE_BO_FLAG_PINNED_LATE_RESTORE/USER_VRAM/GGTT_INVALIDATE
>> flags for cgp bo (Matt)
>>
>> Signed-off-by: Stuart Summers <stuart.summers@intel.com>
>> Signed-off-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
>
>Reviewed-by: Matthew Brost <matthew.brost@intel.com>
>
>As this is new uAPI we will need a UMD PR and ack from a UMD team.
>
Thanks.
I have added the link to UMD PR in the cover letter.
https://github.com/intel/compute-runtime/pull/862.
Will check with UMD team for ack.
>Matt
>
>> ---
>> drivers/gpu/drm/xe/xe_exec_queue.c | 194 ++++++++++++++++++++++-
>> drivers/gpu/drm/xe/xe_exec_queue.h | 47 ++++++
>> drivers/gpu/drm/xe/xe_exec_queue_types.h | 26 +++
>> include/uapi/drm/xe_drm.h | 10 ++
>> 4 files changed, 275 insertions(+), 2 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.c b/drivers/gpu/drm/xe/xe_exec_queue.c
>> index 8724f8de67e2..330b5103222c 100644
>> --- a/drivers/gpu/drm/xe/xe_exec_queue.c
>> +++ b/drivers/gpu/drm/xe/xe_exec_queue.c
>> @@ -13,6 +13,7 @@
>> #include <drm/drm_syncobj.h>
>> #include <uapi/drm/xe_drm.h>
>>
>> +#include "xe_bo.h"
>> #include "xe_dep_scheduler.h"
>> #include "xe_device.h"
>> #include "xe_gt.h"
>> @@ -63,6 +64,33 @@ enum xe_exec_queue_sched_prop {
>> static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q,
>> u64 extensions, int ext_number);
>>
>> +static void xe_exec_queue_group_cleanup(struct xe_exec_queue *q)
>> +{
>> + struct xe_exec_queue_group *group = q->multi_queue.group;
>> + struct xe_lrc *lrc;
>> + unsigned long idx;
>> +
>> + if (xe_exec_queue_is_multi_queue_secondary(q)) {
>> + /*
>> + * Put pairs with get from xe_exec_queue_lookup() call
>> + * in xe_exec_queue_group_validate().
>> + */
>> + xe_exec_queue_put(xe_exec_queue_multi_queue_primary(q));
>> + return;
>> + }
>> +
>> + if (!group)
>> + return;
>> +
>> + /* Primary queue cleanup */
>> + xa_for_each(&group->xa, idx, lrc)
>> + xe_lrc_put(lrc);
>> +
>> + xa_destroy(&group->xa);
>> + xe_bo_unpin_map_no_vm(group->cgp_bo);
>> + kfree(group);
>> +}
>> +
>> static void __xe_exec_queue_free(struct xe_exec_queue *q)
>> {
>> int i;
>> @@ -73,6 +101,10 @@ static void __xe_exec_queue_free(struct xe_exec_queue *q)
>>
>> if (xe_exec_queue_uses_pxp(q))
>> xe_pxp_exec_queue_remove(gt_to_xe(q->gt)->pxp, q);
>> +
>> + if (xe_exec_queue_is_multi_queue(q))
>> + xe_exec_queue_group_cleanup(q);
>> +
>> if (q->vm)
>> xe_vm_put(q->vm);
>>
>> @@ -567,6 +599,147 @@ exec_queue_set_pxp_type(struct xe_device *xe, struct xe_exec_queue *q, u64 value
>> return xe_pxp_exec_queue_set_type(xe->pxp, q, DRM_XE_PXP_TYPE_HWDRM);
>> }
>>
>> +static int xe_exec_queue_group_init(struct xe_device *xe, struct xe_exec_queue *q)
>> +{
>> + struct xe_tile *tile = gt_to_tile(q->gt);
>> + struct xe_exec_queue_group *group;
>> + struct xe_bo *bo;
>> +
>> + group = kzalloc(sizeof(*group), GFP_KERNEL);
>> + if (!group)
>> + return -ENOMEM;
>> +
>> + bo = xe_bo_create_pin_map_novm(xe, tile, SZ_4K, ttm_bo_type_kernel,
>> + XE_BO_FLAG_VRAM_IF_DGFX(tile) |
>> + XE_BO_FLAG_PINNED_LATE_RESTORE |
>> + XE_BO_FLAG_FORCE_USER_VRAM |
>> + XE_BO_FLAG_GGTT_INVALIDATE |
>> + XE_BO_FLAG_GGTT, false);
>> + if (IS_ERR(bo)) {
>> + drm_err(&xe->drm, "CGP bo allocation for queue group failed: %ld\n",
>> + PTR_ERR(bo));
>> + kfree(group);
>> + return PTR_ERR(bo);
>> + }
>> +
>> + xe_map_memset(xe, &bo->vmap, 0, 0, SZ_4K);
>> +
>> + group->primary = q;
>> + group->cgp_bo = bo;
>> + xa_init_flags(&group->xa, XA_FLAGS_ALLOC1);
>> + q->multi_queue.group = group;
>> +
>> + return 0;
>> +}
>> +
>> +static inline bool xe_exec_queue_supports_multi_queue(struct xe_exec_queue *q)
>> +{
>> + return q->gt->info.multi_queue_engine_class_mask & BIT(q->class);
>> +}
>> +
>> +static int xe_exec_queue_group_validate(struct xe_device *xe, struct xe_exec_queue *q,
>> + u32 primary_id)
>> +{
>> + struct xe_exec_queue_group *group;
>> + struct xe_exec_queue *primary;
>> + int ret;
>> +
>> + /*
>> + * Get from below xe_exec_queue_lookup() pairs with put
>> + * in xe_exec_queue_group_cleanup().
>> + */
>> + primary = xe_exec_queue_lookup(q->vm->xef, primary_id);
>> + if (XE_IOCTL_DBG(xe, !primary))
>> + return -ENOENT;
>> +
>> + if (XE_IOCTL_DBG(xe, !xe_exec_queue_is_multi_queue_primary(primary)) ||
>> + XE_IOCTL_DBG(xe, q->vm != primary->vm) ||
>> + XE_IOCTL_DBG(xe, q->logical_mask != primary->logical_mask)) {
>> + ret = -EINVAL;
>> + goto put_primary;
>> + }
>> +
>> + group = primary->multi_queue.group;
>> + q->multi_queue.valid = true;
>> + q->multi_queue.group = group;
>> +
>> + return 0;
>> +put_primary:
>> + xe_exec_queue_put(primary);
>> + return ret;
>> +}
>> +
>> +#define XE_MAX_GROUP_SIZE 64
>> +static int xe_exec_queue_group_add(struct xe_device *xe, struct xe_exec_queue *q)
>> +{
>> + struct xe_exec_queue_group *group = q->multi_queue.group;
>> + u32 pos;
>> + int err;
>> +
>> + xe_assert(xe, xe_exec_queue_is_multi_queue_secondary(q));
>> +
>> + /* Primary queue holds a reference to LRCs of all secondary queues */
>> + err = xa_alloc(&group->xa, &pos, xe_lrc_get(q->lrc[0]),
>> + XA_LIMIT(1, XE_MAX_GROUP_SIZE - 1), GFP_KERNEL);
>> + if (XE_IOCTL_DBG(xe, err)) {
>> + xe_lrc_put(q->lrc[0]);
>> +
>> + /* It is invalid if queue group limit is exceeded */
>> + if (err == -EBUSY)
>> + err = -EINVAL;
>> +
>> + return err;
>> + }
>> +
>> + q->multi_queue.pos = pos;
>> +
>> + return 0;
>> +}
>> +
>> +static void xe_exec_queue_group_delete(struct xe_device *xe, struct xe_exec_queue *q)
>> +{
>> + struct xe_exec_queue_group *group = q->multi_queue.group;
>> + struct xe_lrc *lrc;
>> +
>> + xe_assert(xe, xe_exec_queue_is_multi_queue_secondary(q));
>> +
>> + lrc = xa_erase(&group->xa, q->multi_queue.pos);
>> + xe_assert(xe, lrc);
>> + xe_lrc_put(lrc);
>> +}
>> +
>> +static int exec_queue_set_multi_group(struct xe_device *xe, struct xe_exec_queue *q,
>> + u64 value)
>> +{
>> + if (XE_IOCTL_DBG(xe, !xe_exec_queue_supports_multi_queue(q)))
>> + return -ENODEV;
>> +
>> + if (XE_IOCTL_DBG(xe, !xe_device_uc_enabled(xe)))
>> + return -EOPNOTSUPP;
>> +
>> + if (XE_IOCTL_DBG(xe, xe_exec_queue_is_parallel(q)))
>> + return -EINVAL;
>> +
>> + if (XE_IOCTL_DBG(xe, xe_exec_queue_is_multi_queue(q)))
>> + return -EINVAL;
>> +
>> + if (value & DRM_XE_MULTI_GROUP_CREATE) {
>> + if (XE_IOCTL_DBG(xe, value & ~DRM_XE_MULTI_GROUP_CREATE))
>> + return -EINVAL;
>> +
>> + q->multi_queue.valid = true;
>> + q->multi_queue.is_primary = true;
>> + q->multi_queue.pos = 0;
>> + return 0;
>> + }
>> +
>> + /* While adding secondary queues, the upper 32 bits must be 0 */
>> + if (XE_IOCTL_DBG(xe, value & (~0ull << 32)))
>> + return -EINVAL;
>> +
>> + return xe_exec_queue_group_validate(xe, q, value);
>> +}
>> +
>> typedef int (*xe_exec_queue_set_property_fn)(struct xe_device *xe,
>> struct xe_exec_queue *q,
>> u64 value);
>> @@ -575,6 +748,7 @@ static const xe_exec_queue_set_property_fn exec_queue_set_property_funcs[] = {
>> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY] = exec_queue_set_priority,
>> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE] = exec_queue_set_timeslice,
>> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE] = exec_queue_set_pxp_type,
>> + [DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP] = exec_queue_set_multi_group,
>> };
>>
>> static int exec_queue_user_ext_set_property(struct xe_device *xe,
>> @@ -595,7 +769,8 @@ static int exec_queue_user_ext_set_property(struct xe_device *xe,
>> XE_IOCTL_DBG(xe, ext.pad) ||
>> XE_IOCTL_DBG(xe, ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY &&
>> ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE &&
>> - ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE))
>> + ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE &&
>> + ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP))
>> return -EINVAL;
>>
>> idx = array_index_nospec(ext.property, ARRAY_SIZE(exec_queue_set_property_funcs));
>> @@ -644,6 +819,12 @@ static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue
>> return exec_queue_user_extensions(xe, q, ext.next_extension,
>> ++ext_number);
>>
>> + if (xe_exec_queue_is_multi_queue_primary(q)) {
>> + err = xe_exec_queue_group_init(xe, q);
>> + if (XE_IOCTL_DBG(xe, err))
>> + return err;
>> + }
>> +
>> return 0;
>> }
>>
>> @@ -798,12 +979,18 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
>> if (IS_ERR(q))
>> return PTR_ERR(q);
>>
>> + if (xe_exec_queue_is_multi_queue_secondary(q)) {
>> + err = xe_exec_queue_group_add(xe, q);
>> + if (XE_IOCTL_DBG(xe, err))
>> + goto put_exec_queue;
>> + }
>> +
>> if (xe_vm_in_preempt_fence_mode(vm)) {
>> q->lr.context = dma_fence_context_alloc(1);
>>
>> err = xe_vm_add_compute_exec_queue(vm, q);
>> if (XE_IOCTL_DBG(xe, err))
>> - goto put_exec_queue;
>> + goto delete_queue_group;
>> }
>>
>> if (q->vm && q->hwe->hw_engine_group) {
>> @@ -826,6 +1013,9 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
>>
>> kill_exec_queue:
>> xe_exec_queue_kill(q);
>> +delete_queue_group:
>> + if (xe_exec_queue_is_multi_queue_secondary(q))
>> + xe_exec_queue_group_delete(xe, q);
>> put_exec_queue:
>> xe_exec_queue_put(q);
>> return err;
>> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.h b/drivers/gpu/drm/xe/xe_exec_queue.h
>> index fda4d4f9bda8..e6daa40003f2 100644
>> --- a/drivers/gpu/drm/xe/xe_exec_queue.h
>> +++ b/drivers/gpu/drm/xe/xe_exec_queue.h
>> @@ -66,6 +66,53 @@ static inline bool xe_exec_queue_uses_pxp(struct xe_exec_queue *q)
>> return q->pxp.type;
>> }
>>
>> +/**
>> + * xe_exec_queue_is_multi_queue() - Whether an exec_queue is part of a queue group.
>> + * @q: The exec_queue
>> + *
>> + * Return: True if the exec_queue is part of a queue group, false otherwise.
>> + */
>> +static inline bool xe_exec_queue_is_multi_queue(struct xe_exec_queue *q)
>> +{
>> + return q->multi_queue.valid;
>> +}
>> +
>> +/**
>> + * xe_exec_queue_is_multi_queue_primary() - Whether an exec_queue is primary queue
>> + * of a multi queue group.
>> + * @q: The exec_queue
>> + *
>> + * Return: True if @q is primary queue of a queue group, false otherwise.
>> + */
>> +static inline bool xe_exec_queue_is_multi_queue_primary(struct xe_exec_queue *q)
>> +{
>> + return q->multi_queue.is_primary;
>> +}
>> +
>> +/**
>> + * xe_exec_queue_is_multi_queue_secondary() - Whether an exec_queue is secondary queue
>> + * of a multi queue group.
>> + * @q: The exec_queue
>> + *
>> + * Return: True if @q is secondary queue of a queue group, false otherwise.
>> + */
>> +static inline bool xe_exec_queue_is_multi_queue_secondary(struct xe_exec_queue *q)
>> +{
>> + return xe_exec_queue_is_multi_queue(q) && !xe_exec_queue_is_multi_queue_primary(q);
>> +}
>> +
>> +/**
>> + * xe_exec_queue_multi_queue_primary() - Get multi queue group's primary queue
>> + * @q: The exec_queue
>> + *
>> + * If @q belongs to a multi queue group, then the primary queue of the group will
>> + * be returned. Otherwise, @q will be returned.
>> + */
>> +static inline struct xe_exec_queue *xe_exec_queue_multi_queue_primary(struct xe_exec_queue *q)
>> +{
>> + return xe_exec_queue_is_multi_queue(q) ? q->multi_queue.group->primary : q;
>> +}
>> +
>> bool xe_exec_queue_is_lr(struct xe_exec_queue *q);
>>
>> bool xe_exec_queue_is_idle(struct xe_exec_queue *q);
>> diff --git a/drivers/gpu/drm/xe/xe_exec_queue_types.h b/drivers/gpu/drm/xe/xe_exec_queue_types.h
>> index 771ffe35cd0c..f429b1952be9 100644
>> --- a/drivers/gpu/drm/xe/xe_exec_queue_types.h
>> +++ b/drivers/gpu/drm/xe/xe_exec_queue_types.h
>> @@ -32,6 +32,20 @@ enum xe_exec_queue_priority {
>> XE_EXEC_QUEUE_PRIORITY_COUNT
>> };
>>
>> +/**
>> + * struct xe_exec_queue_group - Execution multi queue group
>> + *
>> + * Contains multi queue group information.
>> + */
>> +struct xe_exec_queue_group {
>> + /** @primary: Primary queue of this group */
>> + struct xe_exec_queue *primary;
>> + /** @cgp_bo: BO for the Context Group Page */
>> + struct xe_bo *cgp_bo;
>> + /** @xa: xarray to store LRCs */
>> + struct xarray xa;
>> +};
>> +
>> /**
>> * struct xe_exec_queue - Execution queue
>> *
>> @@ -111,6 +125,18 @@ struct xe_exec_queue {
>> struct xe_guc_exec_queue *guc;
>> };
>>
>> + /** @multi_queue: Multi queue information */
>> + struct {
>> + /** @multi_queue.group: Queue group information */
>> + struct xe_exec_queue_group *group;
>> + /** @multi_queue.pos: Position of queue within the multi-queue group */
>> + u8 pos;
>> + /** @multi_queue.valid: Queue belongs to a multi queue group */
>> + u8 valid:1;
>> + /** @multi_queue.is_primary: Is primary queue (Q0) of the group */
>> + u8 is_primary:1;
>> + } multi_queue;
>> +
>> /** @sched_props: scheduling properties */
>> struct {
>> /** @sched_props.timeslice_us: timeslice period in micro-seconds */
>> diff --git a/include/uapi/drm/xe_drm.h b/include/uapi/drm/xe_drm.h
>> index 47853659a705..4de21e0a4fcf 100644
>> --- a/include/uapi/drm/xe_drm.h
>> +++ b/include/uapi/drm/xe_drm.h
>> @@ -1252,6 +1252,14 @@ struct drm_xe_vm_bind {
>> * Given that going into a power-saving state kills PXP HWDRM sessions,
>> * runtime PM will be blocked while queues of this type are alive.
>> * All PXP queues will be killed if a PXP invalidation event occurs.
>> + * - %DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP - Create a multi-queue group
>> + * or add secondary queues to a multi-queue group.
>> + * If the extension's 'value' field has %DRM_XE_MULTI_GROUP_CREATE flag set,
>> + * then a new multi-queue group is created with this queue as the primary queue
>> + * (Q0). Otherwise, the queue gets added to the multi-queue group whose primary
>> + * queue's exec_queue_id is specified in the lower 32 bits of the 'value' field.
>> + * All the other non-relevant bits of extension's 'value' field while adding the
>> + * primary or the secondary queues of the group must be set to 0.
>> *
>> * The example below shows how to use @drm_xe_exec_queue_create to create
>> * a simple exec_queue (no parallel submission) of class
>> @@ -1292,6 +1300,8 @@ struct drm_xe_exec_queue_create {
>> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY 0
>> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE 1
>> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE 2
>> +#define DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP 3
>> +#define DRM_XE_MULTI_GROUP_CREATE (1ull << 63)
>> /** @extensions: Pointer to the first extension struct, if any */
>> __u64 extensions;
>>
>> --
>> 2.43.0
>>
next prev parent reply other threads:[~2025-11-22 4:35 UTC|newest]
Thread overview: 41+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-21 3:51 [PATCH v3 00/18] drm/xe: Multi Queue feature support Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 01/18] drm/xe/multi_queue: Add multi_queue_enable_mask to gt information Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 02/18] drm/xe/multi_queue: Add user interface for multi queue support Niranjana Vishwanathapura
2025-11-21 22:51 ` Matthew Brost
2025-11-22 4:35 ` Niranjana Vishwanathapura [this message]
2025-11-21 3:51 ` [PATCH v3 03/18] drm/xe/multi_queue: Add GuC " Niranjana Vishwanathapura
2025-11-22 22:16 ` Matthew Brost
2025-12-03 3:40 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 04/18] drm/xe/multi_queue: Add multi queue priority property Niranjana Vishwanathapura
2025-11-21 22:57 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 05/18] drm/xe/multi_queue: Handle invalid exec queue property setting Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 06/18] drm/xe/multi_queue: Add exec_queue set_property ioctl support Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 07/18] drm/xe/multi_queue: Add support for multi queue dynamic priority change Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 08/18] drm/xe/multi_queue: Add multi queue information to guc_info dump Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 09/18] drm/xe/multi_queue: Handle tearing down of a multi queue Niranjana Vishwanathapura
2025-11-21 23:03 ` Matthew Brost
2025-11-22 4:40 ` Niranjana Vishwanathapura
2025-11-22 5:47 ` Matthew Brost
2025-12-09 3:31 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 10/18] drm/xe/multi_queue: Set QUEUE_DRAIN_MODE for Multi Queue batches Niranjana Vishwanathapura
2025-11-24 18:49 ` Matt Roper
2025-12-02 21:28 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 11/18] drm/xe/multi_queue: Handle CGP context error Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 12/18] drm/xe/multi_queue: Reset GT upon CGP_SYNC failure Niranjana Vishwanathapura
2025-11-21 23:08 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 13/18] drm/xe/multi_queue: Tracepoint support Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 14/18] drm/xe/multi_queue: Support active group after primary is destroyed Niranjana Vishwanathapura
2025-11-22 5:57 ` Matthew Brost
2025-11-22 6:08 ` Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 15/18] drm/xe/xe3p: Disable GuC Dynamic ICS for Xe3p Niranjana Vishwanathapura
2025-11-21 3:51 ` [PATCH v3 16/18] drm/xe/doc: Add documentation for Multi Queue Group Niranjana Vishwanathapura
2025-11-22 6:02 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 17/18] drm/xe/doc: Add documentation for Multi Queue Group GuC interface Niranjana Vishwanathapura
2025-11-22 6:10 ` Matthew Brost
2025-11-21 3:51 ` [PATCH v3 18/18] drm/xe/multi_queue: Enable multi_queue on xe3p_xpc Niranjana Vishwanathapura
2025-11-21 4:01 ` ✗ CI.checkpatch: warning for drm/xe: Multi Queue feature support (rev3) Patchwork
2025-11-21 4:02 ` ✓ CI.KUnit: success " Patchwork
2025-11-21 4:51 ` ✗ Xe.CI.BAT: failure " Patchwork
2025-11-21 8:10 ` ✗ Xe.CI.Full: " Patchwork
2025-11-24 14:04 ` Patchwork
2025-11-27 9:38 ` [PATCH v3 00/18] drm/xe: Multi Queue feature support Hoppe, Mateusz
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aSE9hPYxCNub4RHd@nvishwa1-desk \
--to=niranjana.vishwanathapura@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=matthew.brost@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox