From: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
To: Stuart Summers <stuart.summers@intel.com>
Cc: <igt-dev@lists.freedesktop.org>,
Apoorva Singh <apoorva.singh@intel.com>,
Fei Yang <fei.yang@intel.com>,
Katarzyna Piecielska <katarzyna.piecielska@intel.com>,
Priyanka Dandamudi <priyanka.dandamudi@intel.com>,
Daniel Charles <daniel.charles@intel.com>,
Kamil Konieczny <kamil.konieczny@linux.intel.com>
Subject: Re: [PATCH i-g-t 3/3] tests/intel/xe_exec_reset: Add multi queue subtests
Date: Sun, 12 Apr 2026 19:32:11 -0700 [thread overview]
Message-ID: <adxVq9WK-jdbqZQf@nvishwa1-desk> (raw)
In-Reply-To: <20260409203702.4350-4-stuart.summers@intel.com>
On Thu, Apr 09, 2026 at 08:37:01PM +0000, Stuart Summers wrote:
>From: Apoorva Singh <apoorva.singh@intel.com>
>
>Extend the existing test cases in tests/intel/xe_exec_reset.c
>to include testing of reset flows for both primary queue
>and secondary queues.
>
>Engine resets without CAT faults are triggered via the *-cancel
>cases. These don't include the CANCEL flag as this causes a spinner
>on each queue which isn't adding any extra coverage for multi queue
>over non multi queue.
>
>Since the *-cancel cases are currently implemented only for the
>legacy cases, do the same for multi queue.
>
>New MULTI_QUEUE and SECONDARY_QUEUE flags are added to cover
>the general multi queue cases and the cases where we are triggering
>engine resets and/or cat faults on secondary queues specifically.
>
>Note for multi queue it is interesting to test these secondary
>queue reset scenarios since these are communicated to the driver
>from GuC via the primary queue and after this, the entire queue
>group is torn down. The test cases here are to ensure nothing
>breaks when we hit a scenario like this.
>
>Signed-off-by: Apoorva Singh <apoorva.singh@intel.com>
>Signed-off-by: Fei Yang <fei.yang@intel.com>
>Signed-off-by: Katarzyna Piecielska <katarzyna.piecielska@intel.com>
>Signed-off-by: Priyanka Dandamudi <priyanka.dandamudi@intel.com>
>Signed-off-by: Daniel Charles <daniel.charles@intel.com>
>Signed-off-by: Kamil Konieczny <kamil.konieczny@linux.intel.com>
>Signed-off-by: Stuart Summers <stuart.summers@intel.com>
>---
> lib/xe/xe_legacy.c | 71 ++++++++++--
> tests/intel/xe_exec_reset.c | 222 ++++++++++++++++++++++++++++++++++--
> 2 files changed, 274 insertions(+), 19 deletions(-)
>
>diff --git a/lib/xe/xe_legacy.c b/lib/xe/xe_legacy.c
>index 3371a91ac..f9bd5bcb6 100644
>--- a/lib/xe/xe_legacy.c
>+++ b/lib/xe/xe_legacy.c
>@@ -13,6 +13,8 @@
>
> /* Batch buffer element count, in number of dwords(u32) */
> #define BATCH_DW_COUNT 16
>+#define SECONDARY_QUEUE (0x1 << 15)
>+#define MULTI_QUEUE (0x1 << 14)
> #define COMPRESSION (0x1 << 13)
> #define SYSTEM (0x1 << 12)
> #define LONG_SPIN_REUSE_QUEUE (0x1 << 11)
>@@ -70,10 +72,14 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> xe_spin_nsec_to_ticks(fd, 0, THREE_SEC) : 0,
> };
> int i, b;
>+ int hang_position = flags & SECONDARY_QUEUE ? 1 : 0;
> int extra_execs = (flags & LONG_SPIN_REUSE_QUEUE) ? n_exec_queues : 0;
>
> igt_assert_lte(n_exec_queues, MAX_N_EXECQUEUES);
>
>+ igt_assert_f(!(flags & SECONDARY_QUEUE) || (flags & MULTI_QUEUE),
>+ "SECONDARY_QUEUE requires MULTI_QUEUE to be set");
>+
> if (flags & COMPRESSION)
> igt_require(intel_gen(intel_get_drm_devid(fd)) >= 20);
>
>@@ -101,7 +107,20 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> data = xe_bo_map(fd, bo, bo_size);
>
> for (i = 0; i < n_exec_queues; i++) {
>- exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+ if (flags & MULTI_QUEUE) {
>+ struct drm_xe_ext_set_property multi_queue = {
>+ .base.next_extension = 0,
>+ .base.name = DRM_XE_EXEC_QUEUE_EXTENSION_SET_PROPERTY,
>+ .property = DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP,
>+ };
>+
>+ uint64_t ext = to_user_pointer(&multi_queue);
>+
>+ multi_queue.value = i ? exec_queues[0] : DRM_XE_MULTI_GROUP_CREATE;
>+ exec_queues[i] = xe_exec_queue_create(fd, vm, eci, ext);
>+ } else {
>+ exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+ }
> syncobjs[i] = syncobj_create(fd, 0);
> }
>
>@@ -123,17 +142,22 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> }
>
> for (i = 0; i < n_execs; i++) {
>- u64 base_addr = (!use_capture_mode && (flags & CAT_ERROR) && !i)
>- ? (addr + bo_size * 128) : addr;
>+ u64 base_addr = (!use_capture_mode && flags & CAT_ERROR &&
>+ i == hang_position) ?
>+ (addr + bo_size * 128) : addr;
> u64 batch_offset = (char *)&data[i].batch - (char *)data;
> u64 batch_addr = base_addr + batch_offset;
> u64 spin_offset = (char *)&data[i].spin - (char *)data;
> u64 sdi_offset = (char *)&data[i].data - (char *)data;
> u64 sdi_addr = base_addr + sdi_offset;
> u64 exec_addr;
>- int e = i % n_exec_queues;
>+ int err, e = i % n_exec_queues;
>
>- if (!i || flags & CANCEL ||
>+ /*
>+ * For cat fault on a secondary queue the fault will
>+ * be on the spinner.
>+ */
>+ if (i == hang_position || flags & CANCEL ||
> (flags & LONG_SPIN && i < n_exec_queues)) {
> spin_opts.addr = base_addr + spin_offset;
> xe_spin_init(&data[i].spin, &spin_opts);
>@@ -160,10 +184,17 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> if (e != i)
> syncobj_reset(fd, &syncobjs[e], 1);
>
>- xe_exec(fd, &exec);
>+ /*
>+ * Secondary queues are reset when the primary queue
>+ * is reset. The submission can race here and it is
>+ * expected for those to fail submission if the primary
>+ * reset has already happened.
>+ */
>+ err = __xe_exec(fd, &exec);
>+ igt_assert(!err || ((flags & MULTI_QUEUE) && err == -ECANCELED));
>
>- if (!i && !(flags & CAT_ERROR) && !use_capture_mode &&
>- !(flags & COMPRESSION))
>+ if (i == hang_position && !(flags & CAT_ERROR) &&
>+ !use_capture_mode && !(flags & COMPRESSION))
> xe_spin_wait_started(&data[i].spin);
> }
>
>@@ -186,7 +217,21 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> * Expectation here is that on reset, submissions will
> * still satisfy the syncobj_wait.
> */
>- igt_assert(syncobj_wait(fd, &syncobjs[i], 1, INT64_MAX, 0, NULL));
>+ int err = syncobj_wait_err(fd, &syncobjs[i], 1, INT64_MAX, 0);
>+
>+ /*
>+ * Currently any time GuC resets a queue which is part of a
>+ * multi queue queue group submitted by the KMD, the KMD
>+ * will tear down the entire group. This means we don't know
>+ * whether a particular queue submitted prior to the hanging
>+ * queue will complete or not. So we have to check all possible
>+ * return values here.
>+ *
>+ * In the event we get an -ECANCELED at the exec above and the
>+ * syncobj was not installed, we expect this to return -EINVAL
>+ * here instead.
>+ */
>+ igt_assert(!err || ((flags & MULTI_QUEUE) && err == -EINVAL));
> }
>
> igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
>@@ -232,7 +277,13 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> if (!use_capture_mode && !(flags & (GT_RESET | CANCEL | COMPRESSION))) {
> for (i = flags & LONG_SPIN ? n_exec_queues : 0;
> i < n_execs + extra_execs; i++) {
>- if (!i)
>+ /*
>+ * For multi-queue there is no guarantee which
>+ * queue will be scheduled first as they are all
>+ * submitted at the same priority in this test.
>+ * So we can't guarantee any data integrity here.
>+ */
>+ if (i == hang_position || flags & MULTI_QUEUE)
> continue;
>
> igt_assert_eq(data[i].data, 0xc0ffee);
>diff --git a/tests/intel/xe_exec_reset.c b/tests/intel/xe_exec_reset.c
>index 6b71d9144..ea50567c0 100644
>--- a/tests/intel/xe_exec_reset.c
>+++ b/tests/intel/xe_exec_reset.c
>@@ -112,7 +112,7 @@ static void test_spin(int fd, struct drm_xe_engine_class_instance *eci,
> #define MAX_N_EXECQUEUES 16
> #define GT_RESET (0x1 << 0)
> #define CLOSE_FD (0x1 << 1)
>-#define CLOSE_EXEC_QUEUES (0x1 << 2)
>+#define CLOSE_EXEC_QUEUES (0x1 << 2)
> #define VIRTUAL (0x1 << 3)
> #define PARALLEL (0x1 << 4)
> #define CAT_ERROR (0x1 << 5)
>@@ -124,6 +124,8 @@ static void test_spin(int fd, struct drm_xe_engine_class_instance *eci,
> #define LONG_SPIN_REUSE_QUEUE (0x1 << 11)
> #define SYSTEM (0x1 << 12)
> #define COMPRESSION (0x1 << 13)
>+#define MULTI_QUEUE (0x1 << 14)
>+#define SECONDARY_QUEUE (0x1 << 15)
>
> /**
> * SUBTEST: %s-cat-error
>@@ -353,6 +355,57 @@ test_balancer(int fd, int gt, int class, int n_exec_queues, int n_execs,
> *
> * SUBTEST: cm-close-execqueues-close-fd
> * Description: Test compute mode close exec_queues close fd
>+ *
>+ * SUBTEST: multi-queue-cat-error
>+ * Sub-category: MultiQ tests
>+ * Description: Test cat error with multi_queue
>+ *
>+ * SUBTEST: multi-queue-cat-error-on-secondary
>+ * Sub-category: MultiQ tests
>+ * Description: Test cat error with multi_queue
>+ * on a secondary queue
>+ *
>+ * SUBTEST: multi-queue-gt-reset
>+ * Sub-category: MultiQ tests
>+ * Description: Test GT reset with multi_queue
>+ *
>+ * SUBTEST: multi-queue-cancel
>+ * Sub-category: MultiQ tests
>+ * Description: Test engine reset with multi_queue
>+ *
>+ * SUBTEST: multi-queue-cancel-on-secondary
>+ * Sub-category: MultiQ tests
>+ * Description: Test engine reset with multi_queue
>+ * on a secondary queue
>+ *
>+ * SUBTEST: multi-queue-close-fd
>+ * Sub-category: MultiQ tests
>+ * Description: Test close fd with multi_queue
>+ *
>+ * SUBTEST: multi-queue-close-execqueues
>+ * Sub-category: MultiQ tests
>+ * Description: Test close execqueues with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-cat-error
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode cat error with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-cat-error-on-secondary
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode cat error with multi_queue
>+ * on a secondary queue
>+ *
>+ * SUBTEST: cm-multi-queue-gt-reset
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode GT reset with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-close-fd
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode close fd with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-close-execqueues
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode close execqueues with multi_queue
Remove 'Sub-category' tags here as it is not consistent with other
multi-queue tests (Besides, MultiQ is not a consistent naming used
either).
Other than that, patch LGTM
Reviewed-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
> */
>
> static void
>@@ -384,9 +437,14 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
> } *data;
> struct xe_spin_opts spin_opts = { .preempt = flags & PREEMPT };
> int i, b;
>+ int hang_position = flags & SECONDARY_QUEUE ? 1 : 0;
>
> igt_assert_lte(n_exec_queues, MAX_N_EXECQUEUES);
>
>+ igt_assert_f(!(flags & SECONDARY_QUEUE) ||
>+ ((flags & MULTI_QUEUE) && (flags & CAT_ERROR)),
>+ "SECONDARY_QUEUE requires MULTI_QUEUE and CAT_ERROR to be set");
>+
> if (flags & CLOSE_FD)
> fd = drm_open_driver(DRIVER_XE);
>
>@@ -401,7 +459,20 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
> memset(data, 0, bo_size);
>
> for (i = 0; i < n_exec_queues; i++) {
>- exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+ if (flags & MULTI_QUEUE) {
>+ struct drm_xe_ext_set_property multi_queue = {
>+ .base.next_extension = 0,
>+ .base.name = DRM_XE_EXEC_QUEUE_EXTENSION_SET_PROPERTY,
>+ .property = DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP,
>+ };
>+
>+ uint64_t ext = to_user_pointer(&multi_queue);
>+
>+ multi_queue.value = i ? exec_queues[0] : DRM_XE_MULTI_GROUP_CREATE;
>+ exec_queues[i] = xe_exec_queue_create(fd, vm, eci, ext);
>+ } else {
>+ exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+ }
> };
>
> sync[0].addr = to_user_pointer(&data[0].vm_sync);
>@@ -411,17 +482,21 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
> data[0].vm_sync = 0;
>
> for (i = 0; i < n_execs; i++) {
>- uint64_t base_addr = flags & CAT_ERROR && !i ?
>- addr + bo_size * 128 : addr;
>+ uint64_t base_addr = (flags & CAT_ERROR && i == hang_position) ?
>+ (addr + bo_size * 128) : addr;
> uint64_t batch_offset = (char *)&data[i].batch - (char *)data;
> uint64_t batch_addr = base_addr + batch_offset;
> uint64_t spin_offset = (char *)&data[i].spin - (char *)data;
> uint64_t sdi_offset = (char *)&data[i].data - (char *)data;
> uint64_t sdi_addr = base_addr + sdi_offset;
> uint64_t exec_addr;
>- int e = i % n_exec_queues;
>+ int err, e = i % n_exec_queues;
>
>- if (!i || flags & CANCEL) {
>+ /*
>+ * For cat fault on a secondary queue the fault will
>+ * be on the spinner.
>+ */
>+ if (i == hang_position || flags & CANCEL) {
> spin_opts.addr = base_addr + spin_offset;
> xe_spin_init(&data[i].spin, &spin_opts);
> exec_addr = spin_opts.addr;
>@@ -442,7 +517,18 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
>
> exec.exec_queue_id = exec_queues[e];
> exec.address = exec_addr;
>- xe_exec(fd, &exec);
>+
>+ /*
>+ * Secondary queues are reset when the primary queue
>+ * is reset. The submission can race here and it is
>+ * expected for those to fail submission if the primary
>+ * reset has already happened.
>+ */
>+ err = __xe_exec(fd, &exec);
>+ igt_assert(!err || ((flags & MULTI_QUEUE) && err == -ECANCELED));
>+
>+ if (i == hang_position && !(flags & CAT_ERROR))
>+ xe_spin_wait_started(&data[i].spin);
> }
>
> if (flags & GT_RESET) {
>@@ -467,8 +553,18 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
>
> err = __xe_wait_ufence(fd, &data[i].exec_sync, USER_FENCE_VALUE,
> exec_queues[i % n_exec_queues], &timeout);
>- if (!i) {
>+ if (i == hang_position) {
> igt_assert(err == -ETIME || err == -EIO);
>+ } else if (flags & MULTI_QUEUE) {
>+ /*
>+ * Currently any time GuC resets a queue submitted
>+ * by the KMD, the KMD will tear down the entire
>+ * queue group. This means we don't know whether
>+ * a particular queue submitted prior to the hanging
>+ * queue will complete or not. So we have to check
>+ * all possible return values here.
>+ */
>+ igt_assert(err == -ETIME || err == -EIO || !err);
> } else if (flags & GT_RESET || flags & CAT_ERROR) {
> /* exec races with reset: may return -EIO or complete */
> igt_assert(err == -EIO || !err);
>@@ -483,7 +579,13 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
>
> if (!(flags & (GT_RESET | CANCEL))) {
> for (i = 0; i < n_execs; i++) {
>- if (!i)
>+ /*
>+ * For multi-queue there is no guarantee which
>+ * queue will be scheduled first as they are all
>+ * submitted at the same priority in this test.
>+ * So we can't guarantee any data integrity here.
>+ */
>+ if (i == hang_position || flags & MULTI_QUEUE)
> continue;
>
> igt_assert_eq(data[i].data, 0xc0ffee);
>@@ -986,6 +1088,108 @@ int igt_main()
> xe_for_each_gt(fd, gt)
> gt_mocs_reset(fd, gt);
>
>+ igt_subtest("multi-queue-cat-error") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(fd, hwe, 16, 16,
>+ CAT_ERROR | MULTI_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("multi-queue-cat-error-on-secondary") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(fd, hwe, 16, 16,
>+ CAT_ERROR | MULTI_QUEUE |
>+ SECONDARY_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("multi-queue-gt-reset") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(fd, hwe, 16, 16,
>+ GT_RESET | MULTI_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("multi-queue-cancel") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(fd, hwe, 16, 16,
>+ MULTI_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("multi-queue-cancel-on-secondary") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(fd, hwe, 16, 16,
>+ MULTI_QUEUE | SECONDARY_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("multi-queue-close-fd") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(-1, hwe, 16, 256,
>+ CLOSE_FD | MULTI_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("multi-queue-close-execqueues") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ xe_legacy_test_mode(-1, hwe, 16, 256,
>+ CLOSE_EXEC_QUEUES | CLOSE_FD |
>+ MULTI_QUEUE,
>+ LEGACY_MODE_ADDR,
>+ false);
>+ }
>+
>+ igt_subtest("cm-multi-queue-cat-error") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ test_compute_mode(fd, hwe, 16, 16,
>+ CAT_ERROR | MULTI_QUEUE);
>+ }
>+
>+ igt_subtest("cm-multi-queue-cat-error-on-secondary") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ test_compute_mode(fd, hwe, 16, 16,
>+ CAT_ERROR | MULTI_QUEUE |
>+ SECONDARY_QUEUE);
>+ }
>+
>+ igt_subtest("cm-multi-queue-gt-reset") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ test_compute_mode(fd, hwe, 16, 16,
>+ GT_RESET | MULTI_QUEUE);
>+ }
>+
>+ igt_subtest("cm-multi-queue-close-fd") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ test_compute_mode(-1, hwe, 16, 256,
>+ CLOSE_FD | MULTI_QUEUE);
>+ }
>+
>+ igt_subtest("cm-multi-queue-close-execqueues") {
>+ igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+ xe_for_each_multi_queue_engine(fd, hwe)
>+ test_compute_mode(-1, hwe, 16, 256,
>+ CLOSE_EXEC_QUEUES | CLOSE_FD |
>+ MULTI_QUEUE);
>+ }
>+
> igt_fixture()
> drm_close_driver(fd);
> }
>--
>2.43.0
>
next prev parent reply other threads:[~2026-04-13 2:32 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-09 20:36 [PATCH i-g-t 0/3] Add test coverage for multi queue reset Stuart Summers
2026-04-09 20:36 ` [PATCH i-g-t 1/3] tests/intel/xe_exec_reset: Add a comment about return for syncobj wait Stuart Summers
2026-04-13 2:29 ` Niranjana Vishwanathapura
2026-04-09 20:37 ` [PATCH i-g-t 2/3] tests/intel/xe_exec_reset: Add checks for hanging queue wait_ufence return Stuart Summers
2026-04-13 2:30 ` Niranjana Vishwanathapura
2026-04-09 20:37 ` [PATCH i-g-t 3/3] tests/intel/xe_exec_reset: Add multi queue subtests Stuart Summers
2026-04-13 2:32 ` Niranjana Vishwanathapura [this message]
2026-04-13 21:00 ` Summers, Stuart
2026-04-13 2:34 ` Niranjana Vishwanathapura
2026-04-13 21:01 ` Summers, Stuart
2026-04-13 21:17 ` Summers, Stuart
2026-04-10 4:15 ` ✓ Xe.CI.BAT: success for Add test coverage for multi queue reset Patchwork
2026-04-10 4:29 ` ✓ i915.CI.BAT: " Patchwork
2026-04-10 9:54 ` ✗ Xe.CI.FULL: failure " Patchwork
2026-04-13 21:02 ` Summers, Stuart
2026-04-11 0:11 ` ✗ i915.CI.Full: " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=adxVq9WK-jdbqZQf@nvishwa1-desk \
--to=niranjana.vishwanathapura@intel.com \
--cc=apoorva.singh@intel.com \
--cc=daniel.charles@intel.com \
--cc=fei.yang@intel.com \
--cc=igt-dev@lists.freedesktop.org \
--cc=kamil.konieczny@linux.intel.com \
--cc=katarzyna.piecielska@intel.com \
--cc=priyanka.dandamudi@intel.com \
--cc=stuart.summers@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox