From: "Thomas Hellström" <thomas.hellstrom@linux.intel.com>
To: igt-dev@lists.freedesktop.org
Cc: "Thomas Hellström" <thomas.hellstrom@linux.intel.com>,
"Matthew Brost" <matthew.brost@intel.com>,
"Maarten Lankhorst" <maarten.lankhorst@linux.intel.com>,
"Zbigniew Kempczyński" <zbigniew.kempczynski@intel.com>
Subject: [PATCH i-g-t 1/2] tests/intel/xe_evict: Reduce allocations to maximum working set
Date: Wed, 26 Jun 2024 14:38:32 +0200 [thread overview]
Message-ID: <20240626123833.3164-2-thomas.hellstrom@linux.intel.com> (raw)
In-Reply-To: <20240626123833.3164-1-thomas.hellstrom@linux.intel.com>
Current xe kmd allows for a maximum working set of VRAM plus
half of system memory, or if the working set is allowed only in
VRAM, the working set is limited to VRAM.
Some subtests attempt to exceed that. Detect when that happens
and limit the working set accordingly.
v2:
- The determination for which flags system bos are allowed in the
working set was incorrect. Fix. (Zbigniew Kempczyński)
- Fix a typo.
- Add an assert that vram_size is indeed > 0.
(Zbigniew Kempczyński, Thomas)
- Add asserts and make sure that the bo is bound to the same
vm the exec_queue is using.
- Increase the allowed set size for the multi-vm test.
Cc: Matthew Brost <matthew.brost@intel.com>
Cc: Maarten Lankhorst <maarten.lankhorst@linux.intel.com>
Cc: Zbigniew Kempczyński <zbigniew.kempczynski@intel.com>
Signed-off-by: Thomas Hellström <thomas.hellstrom@linux.intel.com>
---
tests/intel/xe_evict.c | 89 ++++++++++++++++++++++++++++++++++++------
1 file changed, 76 insertions(+), 13 deletions(-)
diff --git a/tests/intel/xe_evict.c b/tests/intel/xe_evict.c
index eebdbc84b..5691ad021 100644
--- a/tests/intel/xe_evict.c
+++ b/tests/intel/xe_evict.c
@@ -97,6 +97,7 @@ test_evict(int fd, struct drm_xe_engine_class_instance *eci,
uint32_t _vm = (flags & EXTERNAL_OBJ) &&
i < n_execs / 8 ? 0 : vm;
+ igt_assert((e & 1) == (i & 1));
if (flags & MULTI_VM) {
__bo = bo[i] = xe_bo_create(fd, 0,
bo_size,
@@ -115,6 +116,7 @@ test_evict(int fd, struct drm_xe_engine_class_instance *eci,
DRM_XE_GEM_CREATE_FLAG_NEEDS_VISIBLE_VRAM);
}
} else {
+ igt_assert((e & 1) == ((i % (n_execs / 2)) & 1));
__bo = bo[i % (n_execs / 2)];
}
if (i)
@@ -273,6 +275,7 @@ test_evict_cm(int fd, struct drm_xe_engine_class_instance *eci,
uint32_t _vm = (flags & EXTERNAL_OBJ) &&
i < n_execs / 8 ? 0 : vm;
+ igt_assert((e & 1) == (i & 1));
if (flags & MULTI_VM) {
__bo = bo[i] = xe_bo_create(fd, 0,
bo_size,
@@ -291,6 +294,7 @@ test_evict_cm(int fd, struct drm_xe_engine_class_instance *eci,
DRM_XE_GEM_CREATE_FLAG_NEEDS_VISIBLE_VRAM);
}
} else {
+ igt_assert((e & 1) == ((i % (n_execs / 2)) & 1));
__bo = bo[i % (n_execs / 2)];
}
if (i)
@@ -458,6 +462,46 @@ static uint64_t calc_bo_size(uint64_t vram_size, int mul, int div)
return (ALIGN(vram_size, SZ_256M) * mul) / div; /* small-bar */
}
+static unsigned int working_set(uint64_t vram_size, uint64_t system_size,
+ uint64_t bo_size, unsigned int num_threads,
+ unsigned int flags)
+{
+ uint64_t set_size;
+ uint64_t total_size;
+
+ igt_assert(vram_size > 0);
+
+ set_size = (vram_size - 1) / bo_size;
+
+ /*
+ * Working set resides also in system?
+ * Currently system graphics memory is limited to 50% of total.
+ */
+ if (!(flags & (THREADED | MULTI_VM)))
+ set_size += (system_size / 2) / bo_size;
+
+ /* Set sizes are per vm. In the multi-vm case we use 2 vms. */
+ if (flags & MULTI_VM)
+ set_size *= 2;
+
+ /* All bos must fit in memory, assuming no swapping */
+ total_size = ((vram_size - 1) / bo_size + system_size / bo_size) /
+ num_threads;
+
+ if (set_size > total_size)
+ set_size = total_size;
+
+ /* bos are only created on half of the execs. */
+ set_size *= 2;
+
+ /*
+ * Align down to ensure the vm the bo is bound to matches the vm
+ * used by the exec_queue, fulfilling the asserts in the
+ * tests.
+ */
+ return ALIGN_DOWN(set_size, 4);
+}
+
/**
* SUBTEST: evict-%s
* Description: %arg[1] evict test.
@@ -748,6 +792,7 @@ igt_main
{ NULL },
};
uint64_t vram_size;
+ uint64_t system_size;
int fd;
igt_fixture {
@@ -755,14 +800,16 @@ igt_main
igt_require(xe_has_vram(fd));
vram_size = xe_visible_vram_size(fd, 0);
igt_assert(vram_size);
+ system_size = igt_get_avail_ram_mb() << 20;
/* Test requires SRAM to about as big as VRAM. For example, small-cm creates
* (448 / 2) BOs with a size (1 / 128) of the total VRAM size. For
* simplicity ensure the SRAM size >= VRAM before running this test.
*/
- igt_skip_on_f(igt_get_avail_ram_mb() < (vram_size >> 20),
- "System memory %lu MiB is less than local memory %lu MiB\n",
- igt_get_avail_ram_mb(), vram_size >> 20);
+ igt_skip_on_f(system_size < vram_size,
+ "System memory %llu MiB is less than local memory %llu MiB\n",
+ (unsigned long long)system_size >> 20,
+ (unsigned long long)vram_size >> 20);
xe_for_each_engine(fd, hwe)
if (hwe->engine_class != DRM_XE_ENGINE_CLASS_COPY)
@@ -770,25 +817,41 @@ igt_main
}
for (const struct section *s = sections; s->name; s++) {
- igt_subtest_f("evict-%s", s->name)
- test_evict(fd, hwe, s->n_exec_queues, s->n_execs,
- calc_bo_size(vram_size, s->mul, s->div),
+ igt_subtest_f("evict-%s", s->name) {
+ uint64_t bo_size = calc_bo_size(vram_size, s->mul, s->div);
+ int ws = working_set(vram_size, system_size, bo_size,
+ 1, s->flags);
+
+ igt_debug("Max working set %d n_execs %d\n", ws, s->n_execs);
+ test_evict(fd, hwe, s->n_exec_queues,
+ min(ws, s->n_execs), bo_size,
s->flags, NULL);
+ }
}
for (const struct section_cm *s = sections_cm; s->name; s++) {
- igt_subtest_f("evict-%s", s->name)
- test_evict_cm(fd, hwe, s->n_exec_queues, s->n_execs,
- calc_bo_size(vram_size, s->mul, s->div),
+ igt_subtest_f("evict-%s", s->name) {
+ uint64_t bo_size = calc_bo_size(vram_size, s->mul, s->div);
+ int ws = working_set(vram_size, system_size, bo_size,
+ 1, s->flags);
+
+ igt_debug("Max working set %d n_execs %d\n", ws, s->n_execs);
+ test_evict_cm(fd, hwe, s->n_exec_queues,
+ min(ws, s->n_execs), bo_size,
s->flags, NULL);
+ }
}
for (const struct section_threads *s = sections_threads; s->name; s++) {
- igt_subtest_f("evict-%s", s->name)
+ igt_subtest_f("evict-%s", s->name) {
+ uint64_t bo_size = calc_bo_size(vram_size, s->mul, s->div);
+ int ws = working_set(vram_size, system_size, bo_size,
+ s->n_threads, s->flags);
+
+ igt_debug("Max working set %d n_execs %d\n", ws, s->n_execs);
threads(fd, hwe, s->n_threads, s->n_exec_queues,
- s->n_execs,
- calc_bo_size(vram_size, s->mul, s->div),
- s->flags);
+ min(ws, s->n_execs), bo_size, s->flags);
+ }
}
igt_fixture
--
2.44.0
next prev parent reply other threads:[~2024-06-26 12:39 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-06-26 12:38 [PATCH i-g-t 0/2] tests/intel/xe_evict: Adapt the working set to memory size Thomas Hellström
2024-06-26 12:38 ` Thomas Hellström [this message]
2024-06-27 6:29 ` [PATCH i-g-t 1/2] tests/intel/xe_evict: Reduce allocations to maximum working set Matthew Brost
2024-06-27 7:28 ` Thomas Hellström
2024-06-26 12:38 ` [PATCH i-g-t 2/2] tests/intel/xe_evict: Reduce the "large" bo size for threaded eviction Thomas Hellström
2024-06-27 6:34 ` Matthew Brost
2024-06-27 7:31 ` Thomas Hellström
2024-06-26 15:20 ` ✗ Fi.CI.BAT: failure for tests/intel/xe_evict: Adapt the working set to memory size Patchwork
2024-06-26 15:23 ` ✓ CI.xeBAT: success " Patchwork
2024-06-26 17:00 ` [PATCH i-g-t 0/2] " Matthew Brost
2024-06-26 21:02 ` ✓ CI.xeFULL: success for " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20240626123833.3164-2-thomas.hellstrom@linux.intel.com \
--to=thomas.hellstrom@linux.intel.com \
--cc=igt-dev@lists.freedesktop.org \
--cc=maarten.lankhorst@linux.intel.com \
--cc=matthew.brost@intel.com \
--cc=zbigniew.kempczynski@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox