From: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
To: Matthew Brost <matthew.brost@intel.com>
Cc: <intel-xe@lists.freedesktop.org>
Subject: Re: [PATCH 02/16] drm/xe/multi_queue: Add user interface for multi queue support
Date: Mon, 3 Nov 2025 14:58:24 -0800 [thread overview]
Message-ID: <aQkzkPLxukAwNzbi@nvishwa1-desk> (raw)
In-Reply-To: <aQUOjCwmhFKsR1tD@lstrano-desk.jf.intel.com>
On Fri, Oct 31, 2025 at 12:31:24PM -0700, Matthew Brost wrote:
>On Fri, Oct 31, 2025 at 11:29:22AM -0700, Niranjana Vishwanathapura wrote:
>> Multi Queue is a new mode of execution supported by the compute and
>> blitter copy command streamers (CCS and BCS, respectively). It is an
>> enhancement of the existing hardware architecture and leverages the
>> same submission model. It enables support for efficient, parallel
>> execution of multiple queues within a single context. All the queues
>> of a group must use the same address space (VM).
>>
>> The new DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_QUEUE execution queue
>> property supports creating a multi queue group and adding queues to
>> a queue group. All queues of a multi queue group share the same
>> context.
>>
>> A exec queue create ioctl call with above property specified with value
>> DRM_XE_SUPER_GROUP_CREATE will create a new multi queue group with the
>> queue being created as the primary queue (aka q0) of the group. To add
>> secondary queues to the group, they need to be created with the above
>> property with id of the primary queue as the value. The properties of
>> the primary queue (like priority, timeslice) applies to the whole group.
>> So, these properties can't be set for secondary queues of a group.
>>
>> Once destroyed, the secondary queues of a multi queue group can't be
>> replaced. However, they can be dynamically added to the group up to a
>> total of 64 queues per group. Once the primary queue is destroyed,
>> secondary queues can't be added to the queue group.
>>
>> Signed-off-by: Stuart Summers <stuart.summers@intel.com>
>> Signed-off-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
>> ---
>> drivers/gpu/drm/xe/xe_exec_queue.c | 191 ++++++++++++++++++++++-
>> drivers/gpu/drm/xe/xe_exec_queue.h | 47 ++++++
>> drivers/gpu/drm/xe/xe_exec_queue_types.h | 30 ++++
>> include/uapi/drm/xe_drm.h | 8 +
>> 4 files changed, 274 insertions(+), 2 deletions(-)
>>
>> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.c b/drivers/gpu/drm/xe/xe_exec_queue.c
>> index 1b57d7c2cc94..86404a7c9fe4 100644
>> --- a/drivers/gpu/drm/xe/xe_exec_queue.c
>> +++ b/drivers/gpu/drm/xe/xe_exec_queue.c
>> @@ -12,6 +12,7 @@
>> #include <drm/drm_file.h>
>> #include <uapi/drm/xe_drm.h>
>>
>> +#include "xe_bo.h"
>> #include "xe_dep_scheduler.h"
>> #include "xe_device.h"
>> #include "xe_gt.h"
>> @@ -62,6 +63,32 @@ enum xe_exec_queue_sched_prop {
>> static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue *q,
>> u64 extensions, int ext_number);
>>
>> +static void xe_exec_queue_group_cleanup(struct xe_exec_queue *q)
>> +{
>> + struct xe_exec_queue_group *group = q->multi_queue.group;
>> + struct xe_lrc *lrc;
>> + unsigned long idx;
>> +
>> + if (xe_exec_queue_is_multi_queue_secondary(q)) {
>> + xe_exec_queue_put(xe_exec_queue_multi_queue_primary(q));
>> + return;
>> + }
>> +
>> + if (!group)
>> + return;
>> +
>> + /* Primary queue cleanup */
>> + mutex_lock(&group->lock);
>
>I don't think you need the group->lock here. Xarrays have their own
>internal locking.
>
>We do use mutexes around xarrays in Xe, but that's to protect the object
>reference—not the xarray itself.
>
>For example, we follow this pattern:
>
>lock();
>obj = xa_find();
>if (obj)
> xe_obj_get(obj);
>unlock();
>
>Similarly, we apply a lock on the removal side. This prevents the object
>from being removed and a reference being dropped in parallel with a
>lookup (i.e., it avoids a use-after-free).
>
>We don’t always use this pattern correctly—some of that is legacy code
>we haven’t cleaned up yet—but we should.
>
>In your case, you're not protecting any object references (i.e., there's
>no lookup function involved), as far as I can tell. So there's no need
>for a lock here.
>
Ok, will remove.
Niranjana
>Matt
>
>> + xa_for_each(&group->xa, idx, lrc)
>> + xe_lrc_put(lrc);
>> + mutex_unlock(&group->lock);
>> +
>> + xa_destroy(&group->xa);
>> + mutex_destroy(&group->lock);
>> + xe_bo_unpin_map_no_vm(group->cgp_bo);
>> + kfree(group);
>> +}
>> +
>> static void __xe_exec_queue_free(struct xe_exec_queue *q)
>> {
>> int i;
>> @@ -72,6 +99,10 @@ static void __xe_exec_queue_free(struct xe_exec_queue *q)
>>
>> if (xe_exec_queue_uses_pxp(q))
>> xe_pxp_exec_queue_remove(gt_to_xe(q->gt)->pxp, q);
>> +
>> + if (xe_exec_queue_is_multi_queue(q))
>> + xe_exec_queue_group_cleanup(q);
>> +
>> if (q->vm)
>> xe_vm_put(q->vm);
>>
>> @@ -549,6 +580,148 @@ exec_queue_set_pxp_type(struct xe_device *xe, struct xe_exec_queue *q, u64 value
>> return xe_pxp_exec_queue_set_type(xe->pxp, q, DRM_XE_PXP_TYPE_HWDRM);
>> }
>>
>> +static int xe_exec_queue_group_init(struct xe_device *xe, struct xe_exec_queue *q)
>> +{
>> + struct xe_tile *tile = gt_to_tile(q->gt);
>> + struct xe_exec_queue_group *group;
>> + struct xe_bo *bo;
>> +
>> + group = kzalloc(sizeof(*group), GFP_KERNEL);
>> + if (!group)
>> + return -ENOMEM;
>> +
>> + bo = xe_bo_create_pin_map_novm(xe, tile, SZ_4K, ttm_bo_type_kernel,
>> + XE_BO_FLAG_VRAM_IF_DGFX(tile) |
>> + XE_BO_FLAG_GGTT, false);
>> + if (IS_ERR(bo)) {
>> + drm_err(&xe->drm, "CGP bo allocation for queue group failed: %ld\n",
>> + PTR_ERR(bo));
>> + kfree(group);
>> + return PTR_ERR(bo);
>> + }
>> +
>> + xe_map_memset(xe, &bo->vmap, 0, 0, SZ_4K);
>> +
>> + group->primary = q;
>> + group->cgp_bo = bo;
>> + xa_init_flags(&group->xa, XA_FLAGS_ALLOC1);
>> + mutex_init(&group->lock);
>> + mutex_init(&group->list_lock);
>> + q->multi_queue.group = group;
>> +
>> + return 0;
>> +}
>> +
>> +static inline bool xe_exec_queue_supports_multi_queue(struct xe_exec_queue *q)
>> +{
>> + return q->gt->info.multi_queue_enable_mask & BIT(q->class);
>> +}
>> +
>> +static int xe_exec_queue_group_validate(struct xe_device *xe, struct xe_exec_queue *q,
>> + u32 primary_id)
>> +{
>> + struct xe_exec_queue_group *group;
>> + struct xe_exec_queue *primary;
>> + int ret;
>> +
>> + primary = xe_exec_queue_lookup(q->vm->xef, primary_id);
>> + if (XE_IOCTL_DBG(xe, !primary))
>> + return -ENOENT;
>> +
>> + if (XE_IOCTL_DBG(xe, !xe_exec_queue_is_multi_queue_primary(primary)) ||
>> + XE_IOCTL_DBG(xe, q->vm != primary->vm) ||
>> + XE_IOCTL_DBG(xe, q->logical_mask != primary->logical_mask)) {
>> + ret = -EINVAL;
>> + goto put_primary;
>> + }
>> +
>> + group = primary->multi_queue.group;
>> + q->multi_queue.valid = true;
>> + q->multi_queue.group = group;
>> +
>> + return 0;
>> +put_primary:
>> + xe_exec_queue_put(primary);
>> + return ret;
>> +}
>> +
>> +#define XE_MAX_GROUP_SIZE 64
>> +static int xe_exec_queue_group_add(struct xe_device *xe, struct xe_exec_queue *q)
>> +{
>> + struct xe_exec_queue_group *group = q->multi_queue.group;
>> + u32 pos;
>> + int err;
>> +
>> + if (!xe_exec_queue_is_multi_queue_secondary(q))
>> + return 0;
>> +
>> + mutex_lock(&group->lock);
>> + err = xa_alloc(&group->xa, &pos, xe_lrc_get(q->lrc[0]),
>> + XA_LIMIT(1, XE_MAX_GROUP_SIZE - 1), GFP_KERNEL);
>> + if (XE_IOCTL_DBG(xe, err)) {
>> + xe_lrc_put(q->lrc[0]);
>> + mutex_unlock(&group->lock);
>> +
>> + /* It is invalid if queue group limit is exceeded */
>> + if (err == -EBUSY)
>> + err = -EINVAL;
>> +
>> + return err;
>> + }
>> +
>> + q->multi_queue.pos = pos;
>> + mutex_unlock(&group->lock);
>> +
>> + return 0;
>> +}
>> +
>> +static void xe_exec_queue_group_delete(struct xe_exec_queue *q)
>> +{
>> + struct xe_exec_queue_group *group = q->multi_queue.group;
>> + struct xe_lrc *lrc;
>> +
>> + if (!xe_exec_queue_is_multi_queue_secondary(q))
>> + return;
>> +
>> + mutex_lock(&group->lock);
>> + lrc = xa_erase(&group->xa, q->multi_queue.pos);
>> + if (lrc)
>> + xe_lrc_put(lrc);
>> + mutex_unlock(&group->lock);
>> +}
>> +
>> +static int exec_queue_set_multi_group(struct xe_device *xe, struct xe_exec_queue *q,
>> + u64 value)
>> +{
>> + if (XE_IOCTL_DBG(xe, !xe_exec_queue_supports_multi_queue(q)))
>> + return -ENODEV;
>> +
>> + if (XE_IOCTL_DBG(xe, !xe_device_uc_enabled(xe)))
>> + return -EOPNOTSUPP;
>> +
>> + if (XE_IOCTL_DBG(xe, xe_exec_queue_is_parallel(q)))
>> + return -EINVAL;
>> +
>> + if (XE_IOCTL_DBG(xe, xe_exec_queue_is_multi_queue(q)))
>> + return -EINVAL;
>> +
>> + if (value & DRM_XE_MULTI_GROUP_CREATE) {
>> + if (XE_IOCTL_DBG(xe, value & ~DRM_XE_MULTI_GROUP_CREATE))
>> + return -EINVAL;
>> +
>> + q->multi_queue.valid = true;
>> + q->multi_queue.is_primary = true;
>> + q->multi_queue.pos = 0;
>> + return 0;
>> + }
>> +
>> + /* While adding secondary queues, the upper 32 bits must be 0 */
>> + if (XE_IOCTL_DBG(xe, value & (~0ull << 32)))
>> + return -EINVAL;
>> +
>> + return xe_exec_queue_group_validate(xe, q, value);
>> +}
>> +
>> typedef int (*xe_exec_queue_set_property_fn)(struct xe_device *xe,
>> struct xe_exec_queue *q,
>> u64 value);
>> @@ -557,6 +730,7 @@ static const xe_exec_queue_set_property_fn exec_queue_set_property_funcs[] = {
>> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY] = exec_queue_set_priority,
>> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE] = exec_queue_set_timeslice,
>> [DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE] = exec_queue_set_pxp_type,
>> + [DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP] = exec_queue_set_multi_group,
>> };
>>
>> static int exec_queue_user_ext_set_property(struct xe_device *xe,
>> @@ -577,7 +751,8 @@ static int exec_queue_user_ext_set_property(struct xe_device *xe,
>> XE_IOCTL_DBG(xe, ext.pad) ||
>> XE_IOCTL_DBG(xe, ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY &&
>> ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE &&
>> - ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE))
>> + ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE &&
>> + ext.property != DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP))
>> return -EINVAL;
>>
>> idx = array_index_nospec(ext.property, ARRAY_SIZE(exec_queue_set_property_funcs));
>> @@ -626,6 +801,12 @@ static int exec_queue_user_extensions(struct xe_device *xe, struct xe_exec_queue
>> return exec_queue_user_extensions(xe, q, ext.next_extension,
>> ++ext_number);
>>
>> + if (xe_exec_queue_is_multi_queue_primary(q)) {
>> + err = xe_exec_queue_group_init(xe, q);
>> + if (XE_IOCTL_DBG(xe, err))
>> + return err;
>> + }
>> +
>> return 0;
>> }
>>
>> @@ -780,12 +961,16 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
>> if (IS_ERR(q))
>> return PTR_ERR(q);
>>
>> + err = xe_exec_queue_group_add(xe, q);
>> + if (XE_IOCTL_DBG(xe, err))
>> + goto put_exec_queue;
>> +
>> if (xe_vm_in_preempt_fence_mode(vm)) {
>> q->lr.context = dma_fence_context_alloc(1);
>>
>> err = xe_vm_add_compute_exec_queue(vm, q);
>> if (XE_IOCTL_DBG(xe, err))
>> - goto put_exec_queue;
>> + goto delete_queue_group;
>> }
>>
>> if (q->vm && q->hwe->hw_engine_group) {
>> @@ -808,6 +993,8 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
>>
>> kill_exec_queue:
>> xe_exec_queue_kill(q);
>> +delete_queue_group:
>> + xe_exec_queue_group_delete(q);
>> put_exec_queue:
>> xe_exec_queue_put(q);
>> return err;
>> diff --git a/drivers/gpu/drm/xe/xe_exec_queue.h b/drivers/gpu/drm/xe/xe_exec_queue.h
>> index a4dfbe858bda..8cd6487018fa 100644
>> --- a/drivers/gpu/drm/xe/xe_exec_queue.h
>> +++ b/drivers/gpu/drm/xe/xe_exec_queue.h
>> @@ -62,6 +62,53 @@ static inline bool xe_exec_queue_uses_pxp(struct xe_exec_queue *q)
>> return q->pxp.type;
>> }
>>
>> +/**
>> + * xe_exec_queue_is_multi_queue() - Whether an exec_queue is part of a queue group.
>> + * @q: The exec_queue
>> + *
>> + * Return: True if the exec_queue is part of a queue group, false otherwise.
>> + */
>> +static inline bool xe_exec_queue_is_multi_queue(struct xe_exec_queue *q)
>> +{
>> + return q->multi_queue.valid;
>> +}
>> +
>> +/**
>> + * xe_exec_queue_is_multi_queue_primary() - Whether an exec_queue is primary queue
>> + * of a multi queue group.
>> + * @q: The exec_queue
>> + *
>> + * Return: True if @q is primary queue of a queue group, false otherwise.
>> + */
>> +static inline bool xe_exec_queue_is_multi_queue_primary(struct xe_exec_queue *q)
>> +{
>> + return q->multi_queue.is_primary;
>> +}
>> +
>> +/**
>> + * xe_exec_queue_is_multi_queue_secondary() - Whether an exec_queue is secondary queue
>> + * of a multi queue group.
>> + * @q: The exec_queue
>> + *
>> + * Return: True if @q is secondary queue of a queue group, false otherwise.
>> + */
>> +static inline bool xe_exec_queue_is_multi_queue_secondary(struct xe_exec_queue *q)
>> +{
>> + return xe_exec_queue_is_multi_queue(q) && !q->multi_queue.is_primary;
>> +}
>> +
>> +/**
>> + * xe_exec_queue_multi_queue_primary() - Get multi queue group's primary queue
>> + * @q: The exec_queue
>> + *
>> + * If @q belongs to a multi queue group, then the primary queue of the group will
>> + * be returned. Otherwise, @q will be returned.
>> + */
>> +static inline struct xe_exec_queue *xe_exec_queue_multi_queue_primary(struct xe_exec_queue *q)
>> +{
>> + return xe_exec_queue_is_multi_queue(q) ? q->multi_queue.group->primary : q;
>> +}
>> +
>> bool xe_exec_queue_is_lr(struct xe_exec_queue *q);
>>
>> bool xe_exec_queue_is_idle(struct xe_exec_queue *q);
>> diff --git a/drivers/gpu/drm/xe/xe_exec_queue_types.h b/drivers/gpu/drm/xe/xe_exec_queue_types.h
>> index c8807268ec6c..3856776df5c4 100644
>> --- a/drivers/gpu/drm/xe/xe_exec_queue_types.h
>> +++ b/drivers/gpu/drm/xe/xe_exec_queue_types.h
>> @@ -31,6 +31,24 @@ enum xe_exec_queue_priority {
>> XE_EXEC_QUEUE_PRIORITY_COUNT
>> };
>>
>> +/**
>> + * struct xe_exec_queue_group - Execution multi queue group
>> + *
>> + * Contains multi queue group information.
>> + */
>> +struct xe_exec_queue_group {
>> + /** @primary: Primary queue of this group */
>> + struct xe_exec_queue *primary;
>> + /** @lock: Queue group update lock */
>> + struct mutex lock;
>> + /** @cgp_bo: BO for the Context Group Page */
>> + struct xe_bo *cgp_bo;
>> + /** @xa: xarray to store LRCs */
>> + struct xarray xa;
>> + /** @list_lock: Secondary queue list lock */
>> + struct mutex list_lock;
>> +};
>> +
>> /**
>> * struct xe_exec_queue - Execution queue
>> *
>> @@ -110,6 +128,18 @@ struct xe_exec_queue {
>> struct xe_guc_exec_queue *guc;
>> };
>>
>> + /** @multi_queue: Multi queue information */
>> + struct {
>> + /** @multi_queue.group: Queue group information */
>> + struct xe_exec_queue_group *group;
>> + /** @multi_queue.pos: Position of queue within the multi-queue group */
>> + u8 pos;
>> + /** @multi_queue.valid: Queue belongs to a multi queue group */
>> + u8 valid:1;
>> + /** @multi_queue.is_primary: Is primary queue (Q0) of the group */
>> + u8 is_primary:1;
>> + } multi_queue;
>> +
>> /** @sched_props: scheduling properties */
>> struct {
>> /** @sched_props.timeslice_us: timeslice period in micro-seconds */
>> diff --git a/include/uapi/drm/xe_drm.h b/include/uapi/drm/xe_drm.h
>> index 47853659a705..d903b3a55ec1 100644
>> --- a/include/uapi/drm/xe_drm.h
>> +++ b/include/uapi/drm/xe_drm.h
>> @@ -1252,6 +1252,12 @@ struct drm_xe_vm_bind {
>> * Given that going into a power-saving state kills PXP HWDRM sessions,
>> * runtime PM will be blocked while queues of this type are alive.
>> * All PXP queues will be killed if a PXP invalidation event occurs.
>> + * - %DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP - Create a multi-queue group
>> + * or add secondary queues to a multi-queue group.
>> + * If the extension's 'value' field has %DRM_XE_MULTI_GROUP_CREATE flag set,
>> + * then a new multi-queue group is created with this queue as the primary queue
>> + * (Q0). Otherwise, the queue gets added to the multi-queue group whose primary
>> + * queue id is specified in the 'value' field.
>> *
>> * The example below shows how to use @drm_xe_exec_queue_create to create
>> * a simple exec_queue (no parallel submission) of class
>> @@ -1292,6 +1298,8 @@ struct drm_xe_exec_queue_create {
>> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_PRIORITY 0
>> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_TIMESLICE 1
>> #define DRM_XE_EXEC_QUEUE_SET_PROPERTY_PXP_TYPE 2
>> +#define DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP 3
>> +#define DRM_XE_MULTI_GROUP_CREATE (1ull << 63)
>> /** @extensions: Pointer to the first extension struct, if any */
>> __u64 extensions;
>>
>> --
>> 2.43.0
>>
next prev parent reply other threads:[~2025-11-03 22:58 UTC|newest]
Thread overview: 61+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-10-31 18:29 [PATCH 00/16] drm/xe: Multi Queue feature support Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 01/16] drm/xe/multi_queue: Add multi_queue_enable_mask to gt information Niranjana Vishwanathapura
2025-11-02 0:01 ` Matthew Brost
2025-11-03 1:25 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 02/16] drm/xe/multi_queue: Add user interface for multi queue support Niranjana Vishwanathapura
2025-10-31 19:31 ` Matthew Brost
2025-11-03 22:58 ` Niranjana Vishwanathapura [this message]
2025-11-02 0:23 ` Matthew Brost
2025-11-03 22:59 ` Niranjana Vishwanathapura
2025-11-02 17:37 ` Matthew Brost
2025-11-03 23:06 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 03/16] drm/xe/multi_queue: Add GuC " Niranjana Vishwanathapura
2025-11-01 18:07 ` Matthew Brost
2025-11-04 4:56 ` Niranjana Vishwanathapura
2025-11-04 17:41 ` Matthew Brost
2025-11-04 18:55 ` Niranjana Vishwanathapura
2025-11-04 19:26 ` Matthew Brost
2025-11-02 18:02 ` Matthew Brost
2025-11-04 5:02 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 04/16] drm/xe/multi_queue: Add multi queue priority property Niranjana Vishwanathapura
2025-11-01 23:59 ` Matthew Brost
2025-11-03 4:45 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 05/16] drm/xe/multi_queue: Handle invalid exec queue property setting Niranjana Vishwanathapura
2025-11-03 22:41 ` Matthew Brost
2025-10-31 18:29 ` [PATCH 06/16] drm/xe/multi_queue: Add exec_queue set_property ioctl support Niranjana Vishwanathapura
2025-11-02 16:53 ` Matthew Brost
2025-11-03 1:49 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 07/16] drm/xe/multi_queue: Add support for multi queue dynamic priority change Niranjana Vishwanathapura
2025-11-01 23:23 ` Matthew Brost
2025-11-03 18:06 ` Niranjana Vishwanathapura
2025-11-01 23:41 ` Matthew Brost
2025-11-03 18:14 ` Niranjana Vishwanathapura
2025-11-03 19:05 ` Matthew Brost
2025-10-31 18:29 ` [PATCH 08/16] drm/xe/multi_queue: Add multi queue information to guc_info dump Niranjana Vishwanathapura
2025-11-01 18:31 ` Matthew Brost
2025-11-03 1:15 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 09/16] drm/xe/multi_queue: Handle tearing down of a multi queue Niranjana Vishwanathapura
2025-11-02 0:39 ` Matthew Brost
2025-11-04 3:35 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 10/16] drm/xe/multi_queue: Set QUEUE_DRAIN_MODE for Multi Queue batches Niranjana Vishwanathapura
2025-11-02 18:22 ` Matthew Brost
2025-11-03 17:09 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 11/16] drm/xe/multi_queue: Handle CGP context error Niranjana Vishwanathapura
2025-11-02 18:29 ` Matthew Brost
2025-11-03 16:44 ` Niranjana Vishwanathapura
2025-11-03 17:18 ` Matthew Brost
2025-10-31 18:29 ` [PATCH 12/16] drm/xe/multi_queue: Tracepoint support Niranjana Vishwanathapura
2025-11-01 18:32 ` Matthew Brost
2025-10-31 18:29 ` [PATCH 13/16] drm/xe/multi_queue: Support active group after primary is destroyed Niranjana Vishwanathapura
2025-11-03 22:05 ` Matthew Brost
2025-11-04 17:24 ` Niranjana Vishwanathapura
2025-11-04 17:30 ` Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 14/16] drm/xe/doc: Add documentation for Multi Queue Group Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 15/16] drm/xe/doc: Add documentation for Multi Queue Group GuC interface Niranjana Vishwanathapura
2025-10-31 18:29 ` [PATCH 16/16] drm/xe/multi_queue: Enable multi_queue on xe3p_xpc Niranjana Vishwanathapura
2025-11-02 0:05 ` Matthew Brost
2025-10-31 18:47 ` [PATCH 00/16] drm/xe: Multi Queue feature support Niranjana Vishwanathapura
2025-10-31 21:15 ` ✗ CI.checkpatch: warning for " Patchwork
2025-10-31 21:16 ` ✓ CI.KUnit: success " Patchwork
2025-10-31 22:19 ` ✗ Xe.CI.BAT: failure " Patchwork
2025-11-01 11:25 ` ✗ Xe.CI.Full: " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aQkzkPLxukAwNzbi@nvishwa1-desk \
--to=niranjana.vishwanathapura@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=matthew.brost@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox