public inbox for igt-dev@lists.freedesktop.org
 help / color / mirror / Atom feed
From: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
To: Stuart Summers <stuart.summers@intel.com>
Cc: <igt-dev@lists.freedesktop.org>,
	Apoorva Singh <apoorva.singh@intel.com>,
	 Fei Yang <fei.yang@intel.com>,
	Katarzyna Piecielska <katarzyna.piecielska@intel.com>,
	Priyanka Dandamudi <priyanka.dandamudi@intel.com>,
	Daniel Charles <daniel.charles@intel.com>,
	Kamil Konieczny <kamil.konieczny@linux.intel.com>
Subject: Re: [PATCH i-g-t 3/3] tests/intel/xe_exec_reset: Add multi queue subtests
Date: Sun, 12 Apr 2026 19:32:11 -0700	[thread overview]
Message-ID: <adxVq9WK-jdbqZQf@nvishwa1-desk> (raw)
In-Reply-To: <20260409203702.4350-4-stuart.summers@intel.com>

On Thu, Apr 09, 2026 at 08:37:01PM +0000, Stuart Summers wrote:
>From: Apoorva Singh <apoorva.singh@intel.com>
>
>Extend the existing test cases in tests/intel/xe_exec_reset.c
>to include testing of reset flows for both primary queue
>and secondary queues.
>
>Engine resets without CAT faults are triggered via the *-cancel
>cases. These don't include the CANCEL flag as this causes a spinner
>on each queue which isn't adding any extra coverage for multi queue
>over non multi queue.
>
>Since the *-cancel cases are currently implemented only for the
>legacy cases, do the same for multi queue.
>
>New MULTI_QUEUE and SECONDARY_QUEUE flags are added to cover
>the general multi queue cases and the cases where we are triggering
>engine resets and/or cat faults on secondary queues specifically.
>
>Note for multi queue it is interesting to test these secondary
>queue reset scenarios since these are communicated to the driver
>from GuC via the primary queue and after this, the entire queue
>group is torn down. The test cases here are to ensure nothing
>breaks when we hit a scenario like this.
>
>Signed-off-by: Apoorva Singh <apoorva.singh@intel.com>
>Signed-off-by: Fei Yang <fei.yang@intel.com>
>Signed-off-by: Katarzyna Piecielska <katarzyna.piecielska@intel.com>
>Signed-off-by: Priyanka Dandamudi <priyanka.dandamudi@intel.com>
>Signed-off-by: Daniel Charles <daniel.charles@intel.com>
>Signed-off-by: Kamil Konieczny <kamil.konieczny@linux.intel.com>
>Signed-off-by: Stuart Summers <stuart.summers@intel.com>
>---
> lib/xe/xe_legacy.c          |  71 ++++++++++--
> tests/intel/xe_exec_reset.c | 222 ++++++++++++++++++++++++++++++++++--
> 2 files changed, 274 insertions(+), 19 deletions(-)
>
>diff --git a/lib/xe/xe_legacy.c b/lib/xe/xe_legacy.c
>index 3371a91ac..f9bd5bcb6 100644
>--- a/lib/xe/xe_legacy.c
>+++ b/lib/xe/xe_legacy.c
>@@ -13,6 +13,8 @@
>
> /* Batch buffer element count, in number of dwords(u32) */
> #define BATCH_DW_COUNT			16
>+#define SECONDARY_QUEUE			(0x1 << 15)
>+#define MULTI_QUEUE			(0x1 << 14)
> #define COMPRESSION			(0x1 << 13)
> #define SYSTEM				(0x1 << 12)
> #define LONG_SPIN_REUSE_QUEUE		(0x1 << 11)
>@@ -70,10 +72,14 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 			xe_spin_nsec_to_ticks(fd, 0, THREE_SEC) : 0,
> 	};
> 	int i, b;
>+	int hang_position = flags & SECONDARY_QUEUE ? 1 : 0;
> 	int extra_execs = (flags & LONG_SPIN_REUSE_QUEUE) ? n_exec_queues : 0;
>
> 	igt_assert_lte(n_exec_queues, MAX_N_EXECQUEUES);
>
>+	igt_assert_f(!(flags & SECONDARY_QUEUE) || (flags & MULTI_QUEUE),
>+		     "SECONDARY_QUEUE requires MULTI_QUEUE to be set");
>+
> 	if (flags & COMPRESSION)
> 		igt_require(intel_gen(intel_get_drm_devid(fd)) >= 20);
>
>@@ -101,7 +107,20 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 	data = xe_bo_map(fd, bo, bo_size);
>
> 	for (i = 0; i < n_exec_queues; i++) {
>-		exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+		if (flags & MULTI_QUEUE) {
>+			struct drm_xe_ext_set_property multi_queue = {
>+				.base.next_extension = 0,
>+				.base.name = DRM_XE_EXEC_QUEUE_EXTENSION_SET_PROPERTY,
>+				.property = DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP,
>+			};
>+
>+			uint64_t ext = to_user_pointer(&multi_queue);
>+
>+			multi_queue.value = i ? exec_queues[0] : DRM_XE_MULTI_GROUP_CREATE;
>+			exec_queues[i] = xe_exec_queue_create(fd, vm, eci, ext);
>+		} else {
>+			exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+		}
> 		syncobjs[i] = syncobj_create(fd, 0);
> 	}
>
>@@ -123,17 +142,22 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 	}
>
> 	for (i = 0; i < n_execs; i++) {
>-		u64 base_addr = (!use_capture_mode && (flags & CAT_ERROR) && !i)
>-			? (addr + bo_size * 128) : addr;
>+		u64 base_addr = (!use_capture_mode && flags & CAT_ERROR &&
>+				 i == hang_position) ?
>+				(addr + bo_size * 128) : addr;
> 		u64 batch_offset = (char *)&data[i].batch - (char *)data;
> 		u64 batch_addr = base_addr + batch_offset;
> 		u64 spin_offset = (char *)&data[i].spin - (char *)data;
> 		u64 sdi_offset = (char *)&data[i].data - (char *)data;
> 		u64 sdi_addr = base_addr + sdi_offset;
> 		u64 exec_addr;
>-		int e = i % n_exec_queues;
>+		int err, e = i % n_exec_queues;
>
>-		if (!i || flags & CANCEL ||
>+		/*
>+		 * For cat fault on a secondary queue the fault will
>+		 * be on the spinner.
>+		 */
>+		if (i == hang_position || flags & CANCEL ||
> 		    (flags & LONG_SPIN && i < n_exec_queues)) {
> 			spin_opts.addr = base_addr + spin_offset;
> 			xe_spin_init(&data[i].spin, &spin_opts);
>@@ -160,10 +184,17 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 		if (e != i)
> 			syncobj_reset(fd, &syncobjs[e], 1);
>
>-		xe_exec(fd, &exec);
>+		/*
>+		 * Secondary queues are reset when the primary queue
>+		 * is reset. The submission can race here and it is
>+		 * expected for those to fail submission if the primary
>+		 * reset has already happened.
>+		 */
>+		err = __xe_exec(fd, &exec);
>+		igt_assert(!err || ((flags & MULTI_QUEUE) && err == -ECANCELED));
>
>-		if (!i && !(flags & CAT_ERROR) && !use_capture_mode &&
>-		    !(flags & COMPRESSION))
>+		if (i == hang_position && !(flags & CAT_ERROR) &&
>+		    !use_capture_mode && !(flags & COMPRESSION))
> 			xe_spin_wait_started(&data[i].spin);
> 	}
>
>@@ -186,7 +217,21 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 		 * Expectation here is that on reset, submissions will
> 		 * still satisfy the syncobj_wait.
> 		 */
>-		igt_assert(syncobj_wait(fd, &syncobjs[i], 1, INT64_MAX, 0, NULL));
>+		int err = syncobj_wait_err(fd, &syncobjs[i], 1, INT64_MAX, 0);
>+
>+		/*
>+		 * Currently any time GuC resets a queue which is part of a
>+		 * multi queue queue group submitted by the KMD, the KMD
>+		 * will tear down the entire group. This means we don't know
>+		 * whether a particular queue submitted prior to the hanging
>+		 * queue will complete or not. So we have to check all possible
>+		 * return values here.
>+		 *
>+		 * In the event we get an -ECANCELED at the exec above and the
>+		 * syncobj was not installed, we expect this to return -EINVAL
>+		 * here instead.
>+		 */
>+		igt_assert(!err || ((flags & MULTI_QUEUE) && err == -EINVAL));
> 	}
>
> 	igt_assert(syncobj_wait(fd, &sync[0].handle, 1, INT64_MAX, 0, NULL));
>@@ -232,7 +277,13 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 	if (!use_capture_mode && !(flags & (GT_RESET | CANCEL | COMPRESSION))) {
> 		for (i = flags & LONG_SPIN ? n_exec_queues : 0;
> 		     i < n_execs + extra_execs; i++) {
>-			if (!i)
>+			/*
>+			 * For multi-queue there is no guarantee which
>+			 * queue will be scheduled first as they are all
>+			 * submitted at the same priority in this test.
>+			 * So we can't guarantee any data integrity here.
>+			 */
>+			if (i == hang_position || flags & MULTI_QUEUE)
> 				continue;
>
> 			igt_assert_eq(data[i].data, 0xc0ffee);
>diff --git a/tests/intel/xe_exec_reset.c b/tests/intel/xe_exec_reset.c
>index 6b71d9144..ea50567c0 100644
>--- a/tests/intel/xe_exec_reset.c
>+++ b/tests/intel/xe_exec_reset.c
>@@ -112,7 +112,7 @@ static void test_spin(int fd, struct drm_xe_engine_class_instance *eci,
> #define MAX_N_EXECQUEUES	16
> #define GT_RESET			(0x1 << 0)
> #define CLOSE_FD			(0x1 << 1)
>-#define CLOSE_EXEC_QUEUES	(0x1 << 2)
>+#define CLOSE_EXEC_QUEUES		(0x1 << 2)
> #define VIRTUAL				(0x1 << 3)
> #define PARALLEL			(0x1 << 4)
> #define CAT_ERROR			(0x1 << 5)
>@@ -124,6 +124,8 @@ static void test_spin(int fd, struct drm_xe_engine_class_instance *eci,
> #define LONG_SPIN_REUSE_QUEUE		(0x1 << 11)
> #define SYSTEM				(0x1 << 12)
> #define COMPRESSION			(0x1 << 13)
>+#define MULTI_QUEUE			(0x1 << 14)
>+#define SECONDARY_QUEUE			(0x1 << 15)
>
> /**
>  * SUBTEST: %s-cat-error
>@@ -353,6 +355,57 @@ test_balancer(int fd, int gt, int class, int n_exec_queues, int n_execs,
>  *
>  * SUBTEST: cm-close-execqueues-close-fd
>  * Description: Test compute mode close exec_queues close fd
>+ *
>+ * SUBTEST: multi-queue-cat-error
>+ * Sub-category: MultiQ tests
>+ * Description: Test cat error with multi_queue
>+ *
>+ * SUBTEST: multi-queue-cat-error-on-secondary
>+ * Sub-category: MultiQ tests
>+ * Description: Test cat error with multi_queue
>+ *              on a secondary queue
>+ *
>+ * SUBTEST: multi-queue-gt-reset
>+ * Sub-category: MultiQ tests
>+ * Description: Test GT reset with multi_queue
>+ *
>+ * SUBTEST: multi-queue-cancel
>+ * Sub-category: MultiQ tests
>+ * Description: Test engine reset with multi_queue
>+ *
>+ * SUBTEST: multi-queue-cancel-on-secondary
>+ * Sub-category: MultiQ tests
>+ * Description: Test engine reset with multi_queue
>+ *              on a secondary queue
>+ *
>+ * SUBTEST: multi-queue-close-fd
>+ * Sub-category: MultiQ tests
>+ * Description: Test close fd with multi_queue
>+ *
>+ * SUBTEST: multi-queue-close-execqueues
>+ * Sub-category: MultiQ tests
>+ * Description: Test close execqueues with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-cat-error
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode cat error with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-cat-error-on-secondary
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode cat error with multi_queue
>+ *              on a secondary queue
>+ *
>+ * SUBTEST: cm-multi-queue-gt-reset
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode GT reset with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-close-fd
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode close fd with multi_queue
>+ *
>+ * SUBTEST: cm-multi-queue-close-execqueues
>+ * Sub-category: MultiQ tests
>+ * Description: Test compute mode close execqueues with multi_queue

Remove 'Sub-category' tags here as it is not consistent with other
multi-queue tests (Besides, MultiQ is not a consistent naming used
either).

Other than that, patch LGTM
Reviewed-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>

>  */
>
> static void
>@@ -384,9 +437,14 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 	} *data;
> 	struct xe_spin_opts spin_opts = { .preempt = flags & PREEMPT };
> 	int i, b;
>+	int hang_position = flags & SECONDARY_QUEUE ? 1 : 0;
>
> 	igt_assert_lte(n_exec_queues, MAX_N_EXECQUEUES);
>
>+	igt_assert_f(!(flags & SECONDARY_QUEUE) ||
>+		     ((flags & MULTI_QUEUE) && (flags & CAT_ERROR)),
>+		     "SECONDARY_QUEUE requires MULTI_QUEUE and CAT_ERROR to be set");
>+
> 	if (flags & CLOSE_FD)
> 		fd = drm_open_driver(DRIVER_XE);
>
>@@ -401,7 +459,20 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 	memset(data, 0, bo_size);
>
> 	for (i = 0; i < n_exec_queues; i++) {
>-		exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+		if (flags & MULTI_QUEUE) {
>+			struct drm_xe_ext_set_property multi_queue = {
>+				.base.next_extension = 0,
>+				.base.name = DRM_XE_EXEC_QUEUE_EXTENSION_SET_PROPERTY,
>+				.property = DRM_XE_EXEC_QUEUE_SET_PROPERTY_MULTI_GROUP,
>+			};
>+
>+			uint64_t ext = to_user_pointer(&multi_queue);
>+
>+			multi_queue.value = i ? exec_queues[0] : DRM_XE_MULTI_GROUP_CREATE;
>+			exec_queues[i] = xe_exec_queue_create(fd, vm, eci, ext);
>+		} else {
>+			exec_queues[i] = xe_exec_queue_create(fd, vm, eci, 0);
>+		}
> 	};
>
> 	sync[0].addr = to_user_pointer(&data[0].vm_sync);
>@@ -411,17 +482,21 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
> 	data[0].vm_sync = 0;
>
> 	for (i = 0; i < n_execs; i++) {
>-		uint64_t base_addr = flags & CAT_ERROR && !i ?
>-			addr + bo_size * 128 : addr;
>+		uint64_t base_addr = (flags & CAT_ERROR && i == hang_position) ?
>+				     (addr + bo_size * 128) : addr;
> 		uint64_t batch_offset = (char *)&data[i].batch - (char *)data;
> 		uint64_t batch_addr = base_addr + batch_offset;
> 		uint64_t spin_offset = (char *)&data[i].spin - (char *)data;
> 		uint64_t sdi_offset = (char *)&data[i].data - (char *)data;
> 		uint64_t sdi_addr = base_addr + sdi_offset;
> 		uint64_t exec_addr;
>-		int e = i % n_exec_queues;
>+		int err, e = i % n_exec_queues;
>
>-		if (!i || flags & CANCEL) {
>+		/*
>+		 * For cat fault on a secondary queue the fault will
>+		 * be on the spinner.
>+		 */
>+		if (i == hang_position || flags & CANCEL) {
> 			spin_opts.addr = base_addr + spin_offset;
> 			xe_spin_init(&data[i].spin, &spin_opts);
> 			exec_addr = spin_opts.addr;
>@@ -442,7 +517,18 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
>
> 		exec.exec_queue_id = exec_queues[e];
> 		exec.address = exec_addr;
>-		xe_exec(fd, &exec);
>+
>+		/*
>+		 * Secondary queues are reset when the primary queue
>+		 * is reset. The submission can race here and it is
>+		 * expected for those to fail submission if the primary
>+		 * reset has already happened.
>+		 */
>+		err = __xe_exec(fd, &exec);
>+		igt_assert(!err || ((flags & MULTI_QUEUE) && err == -ECANCELED));
>+
>+		if (i == hang_position && !(flags & CAT_ERROR))
>+			xe_spin_wait_started(&data[i].spin);
> 	}
>
> 	if (flags & GT_RESET) {
>@@ -467,8 +553,18 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
>
> 		err = __xe_wait_ufence(fd, &data[i].exec_sync, USER_FENCE_VALUE,
> 				       exec_queues[i % n_exec_queues], &timeout);
>-		if (!i) {
>+		if (i == hang_position) {
> 			igt_assert(err == -ETIME || err == -EIO);
>+		} else if (flags & MULTI_QUEUE) {
>+			/*
>+			 * Currently any time GuC resets a queue submitted
>+			 * by the KMD, the KMD will tear down the entire
>+			 * queue group. This means we don't know whether
>+			 * a particular queue submitted prior to the hanging
>+			 * queue will complete or not. So we have to check
>+			 * all possible return values here.
>+			 */
>+			igt_assert(err == -ETIME || err == -EIO || !err);
> 		} else if (flags & GT_RESET || flags & CAT_ERROR) {
> 			/* exec races with reset: may return -EIO or complete */
> 			igt_assert(err == -EIO || !err);
>@@ -483,7 +579,13 @@ test_compute_mode(int fd, struct drm_xe_engine_class_instance *eci,
>
> 	if (!(flags & (GT_RESET | CANCEL))) {
> 		for (i = 0; i < n_execs; i++) {
>-			if (!i)
>+			/*
>+			 * For multi-queue there is no guarantee which
>+			 * queue will be scheduled first as they are all
>+			 * submitted at the same priority in this test.
>+			 * So we can't guarantee any data integrity here.
>+			 */
>+			if (i == hang_position || flags & MULTI_QUEUE)
> 				continue;
>
> 			igt_assert_eq(data[i].data, 0xc0ffee);
>@@ -986,6 +1088,108 @@ int igt_main()
> 		xe_for_each_gt(fd, gt)
> 			gt_mocs_reset(fd, gt);
>
>+	igt_subtest("multi-queue-cat-error") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(fd, hwe, 16, 16,
>+					    CAT_ERROR | MULTI_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("multi-queue-cat-error-on-secondary") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(fd, hwe, 16, 16,
>+					    CAT_ERROR | MULTI_QUEUE |
>+					    SECONDARY_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("multi-queue-gt-reset") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(fd, hwe, 16, 16,
>+					    GT_RESET | MULTI_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("multi-queue-cancel") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(fd, hwe, 16, 16,
>+					    MULTI_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("multi-queue-cancel-on-secondary") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(fd, hwe, 16, 16,
>+					    MULTI_QUEUE | SECONDARY_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("multi-queue-close-fd") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(-1, hwe, 16, 256,
>+					    CLOSE_FD | MULTI_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("multi-queue-close-execqueues") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			xe_legacy_test_mode(-1, hwe, 16, 256,
>+					    CLOSE_EXEC_QUEUES | CLOSE_FD |
>+					    MULTI_QUEUE,
>+					    LEGACY_MODE_ADDR,
>+					    false);
>+	}
>+
>+	igt_subtest("cm-multi-queue-cat-error") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			test_compute_mode(fd, hwe, 16, 16,
>+					  CAT_ERROR | MULTI_QUEUE);
>+	}
>+
>+	igt_subtest("cm-multi-queue-cat-error-on-secondary") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			test_compute_mode(fd, hwe, 16, 16,
>+					  CAT_ERROR | MULTI_QUEUE |
>+					  SECONDARY_QUEUE);
>+	}
>+
>+	igt_subtest("cm-multi-queue-gt-reset") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			test_compute_mode(fd, hwe, 16, 16,
>+					  GT_RESET | MULTI_QUEUE);
>+	}
>+
>+	igt_subtest("cm-multi-queue-close-fd") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			test_compute_mode(-1, hwe, 16, 256,
>+					  CLOSE_FD | MULTI_QUEUE);
>+	}
>+
>+	igt_subtest("cm-multi-queue-close-execqueues") {
>+		igt_require(intel_graphics_ver(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
>+		xe_for_each_multi_queue_engine(fd, hwe)
>+			test_compute_mode(-1, hwe, 16, 256,
>+					  CLOSE_EXEC_QUEUES | CLOSE_FD |
>+					  MULTI_QUEUE);
>+	}
>+
> 	igt_fixture()
> 		drm_close_driver(fd);
> }
>-- 
>2.43.0
>

  reply	other threads:[~2026-04-13  2:32 UTC|newest]

Thread overview: 16+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-04-09 20:36 [PATCH i-g-t 0/3] Add test coverage for multi queue reset Stuart Summers
2026-04-09 20:36 ` [PATCH i-g-t 1/3] tests/intel/xe_exec_reset: Add a comment about return for syncobj wait Stuart Summers
2026-04-13  2:29   ` Niranjana Vishwanathapura
2026-04-09 20:37 ` [PATCH i-g-t 2/3] tests/intel/xe_exec_reset: Add checks for hanging queue wait_ufence return Stuart Summers
2026-04-13  2:30   ` Niranjana Vishwanathapura
2026-04-09 20:37 ` [PATCH i-g-t 3/3] tests/intel/xe_exec_reset: Add multi queue subtests Stuart Summers
2026-04-13  2:32   ` Niranjana Vishwanathapura [this message]
2026-04-13 21:00     ` Summers, Stuart
2026-04-13  2:34   ` Niranjana Vishwanathapura
2026-04-13 21:01     ` Summers, Stuart
2026-04-13 21:17       ` Summers, Stuart
2026-04-10  4:15 ` ✓ Xe.CI.BAT: success for Add test coverage for multi queue reset Patchwork
2026-04-10  4:29 ` ✓ i915.CI.BAT: " Patchwork
2026-04-10  9:54 ` ✗ Xe.CI.FULL: failure " Patchwork
2026-04-13 21:02   ` Summers, Stuart
2026-04-11  0:11 ` ✗ i915.CI.Full: " Patchwork

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=adxVq9WK-jdbqZQf@nvishwa1-desk \
    --to=niranjana.vishwanathapura@intel.com \
    --cc=apoorva.singh@intel.com \
    --cc=daniel.charles@intel.com \
    --cc=fei.yang@intel.com \
    --cc=igt-dev@lists.freedesktop.org \
    --cc=kamil.konieczny@linux.intel.com \
    --cc=katarzyna.piecielska@intel.com \
    --cc=priyanka.dandamudi@intel.com \
    --cc=stuart.summers@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox