From: Xin Wang <x.wang@intel.com>
To: igt-dev@lists.freedesktop.org
Cc: "Xin Wang" <x.wang@intel.com>,
"Kamil Konieczny" <kamil.konieczny@linux.intel.com>,
"Matt Roper" <matthew.d.roper@intel.com>,
"Zbigniew Kempczyński" <zbigniew.kempczynski@intel.com>,
"Ravi Kumar V" <ravi.kumar.vodapalli@intel.com>
Subject: [PATCH v2 3/3] intel/xe: use fd-based graphics/IP version helpers
Date: Thu, 22 Jan 2026 07:15:30 +0000 [thread overview]
Message-ID: <20260122071530.1629046-4-x.wang@intel.com> (raw)
In-Reply-To: <20260122071530.1629046-1-x.wang@intel.com>
Switch Xe‑related libraries and tests to use fd‑based intel_gen() and
intel_graphics_ver() instead of PCI ID lookups, keeping behavior aligned
with Xe IP disaggregation.
Cc: Kamil Konieczny <kamil.konieczny@linux.intel.com>
Cc: Matt Roper <matthew.d.roper@intel.com>
Cc: Zbigniew Kempczyński <zbigniew.kempczynski@intel.com>
Cc: Ravi Kumar V <ravi.kumar.vodapalli@intel.com>
Signed-off-by: Xin Wang <x.wang@intel.com>
---
lib/gpgpu_shader.c | 5 ++-
lib/gpu_cmds.c | 21 ++++++-----
lib/intel_batchbuffer.c | 14 +++-----
lib/intel_blt.c | 21 +++++------
lib/intel_blt.h | 2 +-
lib/intel_bufops.c | 10 +++---
lib/intel_common.c | 2 +-
lib/intel_compute.c | 6 ++--
lib/intel_mocs.c | 48 ++++++++++++++------------
lib/intel_pat.c | 19 +++++-----
lib/rendercopy_gen9.c | 22 ++++++------
lib/xe/xe_legacy.c | 2 +-
lib/xe/xe_spin.c | 4 +--
lib/xe/xe_sriov_provisioning.c | 4 +--
tests/intel/api_intel_allocator.c | 2 +-
tests/intel/kms_ccs.c | 13 +++----
tests/intel/kms_fbcon_fbt.c | 2 +-
tests/intel/kms_frontbuffer_tracking.c | 6 ++--
tests/intel/kms_pipe_stress.c | 4 +--
tests/intel/xe_ccs.c | 16 ++++-----
tests/intel/xe_compute.c | 8 ++---
tests/intel/xe_copy_basic.c | 6 ++--
tests/intel/xe_debugfs.c | 3 +-
tests/intel/xe_eudebug_online.c | 8 ++---
tests/intel/xe_exec_multi_queue.c | 2 +-
tests/intel/xe_exec_store.c | 18 ++++------
tests/intel/xe_fault_injection.c | 8 ++---
tests/intel/xe_intel_bb.c | 7 ++--
tests/intel/xe_multigpu_svm.c | 3 +-
tests/intel/xe_pat.c | 16 ++++-----
tests/intel/xe_query.c | 4 +--
tests/intel/xe_render_copy.c | 2 +-
32 files changed, 135 insertions(+), 173 deletions(-)
diff --git a/lib/gpgpu_shader.c b/lib/gpgpu_shader.c
index 767bddb7b..09a7f5c5e 100644
--- a/lib/gpgpu_shader.c
+++ b/lib/gpgpu_shader.c
@@ -274,11 +274,10 @@ void gpgpu_shader_exec(struct intel_bb *ibb,
struct gpgpu_shader *gpgpu_shader_create(int fd)
{
struct gpgpu_shader *shdr = calloc(1, sizeof(struct gpgpu_shader));
- const struct intel_device_info *info;
+ unsigned ip_ver = intel_graphics_ver(fd);
igt_assert(shdr);
- info = intel_get_device_info(intel_get_drm_devid(fd));
- shdr->gen_ver = 100 * info->graphics_ver + info->graphics_rel;
+ shdr->gen_ver = 100 * (ip_ver >> 8) + (ip_ver & 0xff);
shdr->max_size = 16 * 4;
shdr->code = malloc(4 * shdr->max_size);
shdr->labels = igt_map_create(igt_map_hash_32, igt_map_equal_32);
diff --git a/lib/gpu_cmds.c b/lib/gpu_cmds.c
index ab46fe0de..6842af1ad 100644
--- a/lib/gpu_cmds.c
+++ b/lib/gpu_cmds.c
@@ -313,14 +313,13 @@ fill_binding_table(struct intel_bb *ibb, struct intel_buf *buf)
{
uint32_t binding_table_offset;
uint32_t *binding_table;
- uint32_t devid = intel_get_drm_devid(ibb->fd);
intel_bb_ptr_align(ibb, 64);
binding_table_offset = intel_bb_offset(ibb);
binding_table = intel_bb_ptr(ibb);
intel_bb_ptr_add(ibb, 64);
- if (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0)) {
+ if (intel_graphics_ver(ibb->fd) >= IP_VER(20, 0)) {
/*
* Up until now, SURFACEFORMAT_R8_UNROM was used regardless of the 'bpp' value.
* For bpp 32 this results in a surface that is 4x narrower than expected. However
@@ -342,13 +341,13 @@ fill_binding_table(struct intel_bb *ibb, struct intel_buf *buf)
igt_assert_f(false,
"Surface state for bpp = %u not implemented",
buf->bpp);
- } else if (intel_graphics_ver_legacy(devid) >= IP_VER(12, 50)) {
+ } else if (intel_graphics_ver(ibb->fd) >= IP_VER(12, 50)) {
binding_table[0] = xehp_fill_surface_state(ibb, buf,
SURFACEFORMAT_R8_UNORM, 1);
- } else if (intel_graphics_ver_legacy(devid) >= IP_VER(9, 0)) {
+ } else if (intel_graphics_ver(ibb->fd) >= IP_VER(9, 0)) {
binding_table[0] = gen9_fill_surface_state(ibb, buf,
SURFACEFORMAT_R8_UNORM, 1);
- } else if (intel_graphics_ver_legacy(devid) >= IP_VER(8, 0)) {
+ } else if (intel_graphics_ver(ibb->fd) >= IP_VER(8, 0)) {
binding_table[0] = gen8_fill_surface_state(ibb, buf,
SURFACEFORMAT_R8_UNORM, 1);
} else {
@@ -867,7 +866,7 @@ gen_emit_media_object(struct intel_bb *ibb,
/* inline data (xoffset, yoffset) */
intel_bb_out(ibb, xoffset);
intel_bb_out(ibb, yoffset);
- if (intel_gen_legacy(ibb->devid) >= 8 && !IS_CHERRYVIEW(ibb->devid))
+ if (intel_gen(ibb->fd) >= 8 && !IS_CHERRYVIEW(ibb->devid))
gen8_emit_media_state_flush(ibb);
}
@@ -1011,7 +1010,7 @@ void
xehp_emit_state_compute_mode(struct intel_bb *ibb, bool vrt)
{
- uint32_t dword_length = intel_graphics_ver_legacy(ibb->devid) >= IP_VER(20, 0);
+ uint32_t dword_length = intel_graphics_ver(ibb->fd) >= IP_VER(20, 0);
intel_bb_out(ibb, XEHP_STATE_COMPUTE_MODE | dword_length);
intel_bb_out(ibb, vrt ? (0x10001) << 10 : 0); /* Enable variable number of threads */
@@ -1042,7 +1041,7 @@ xehp_emit_state_base_address(struct intel_bb *ibb)
intel_bb_out(ibb, 0);
/* stateless data port */
- tmp = intel_graphics_ver_legacy(ibb->devid) >= IP_VER(20, 0) ? 0 : BASE_ADDRESS_MODIFY;
+ tmp = intel_graphics_ver(ibb->fd) >= IP_VER(20, 0) ? 0 : BASE_ADDRESS_MODIFY;
intel_bb_out(ibb, 0 | tmp); //dw3
/* surface */
@@ -1068,7 +1067,7 @@ xehp_emit_state_base_address(struct intel_bb *ibb)
/* dynamic state buffer size */
intel_bb_out(ibb, ALIGN(ibb->size, 1 << 12) | 1); //dw13
/* indirect object buffer size */
- if (intel_graphics_ver_legacy(ibb->devid) >= IP_VER(20, 0)) //dw14
+ if (intel_graphics_ver(ibb->fd) >= IP_VER(20, 0)) //dw14
intel_bb_out(ibb, 0);
else
intel_bb_out(ibb, 0xfffff000 | 1);
@@ -1115,7 +1114,7 @@ xehp_emit_compute_walk(struct intel_bb *ibb,
else
mask = (1 << mask) - 1;
- dword_length = intel_graphics_ver_legacy(ibb->devid) >= IP_VER(20, 0) ? 0x26 : 0x25;
+ dword_length = intel_graphics_ver(ibb->fd) >= IP_VER(20, 0) ? 0x26 : 0x25;
intel_bb_out(ibb, XEHP_COMPUTE_WALKER | dword_length);
intel_bb_out(ibb, 0); /* debug object */ //dw1
@@ -1155,7 +1154,7 @@ xehp_emit_compute_walk(struct intel_bb *ibb,
intel_bb_out(ibb, 0); //dw16
intel_bb_out(ibb, 0); //dw17
- if (intel_graphics_ver_legacy(ibb->devid) >= IP_VER(20, 0)) //Xe2:dw18
+ if (intel_graphics_ver(ibb->fd) >= IP_VER(20, 0)) //Xe2:dw18
intel_bb_out(ibb, 0);
/* Interface descriptor data */
for (int i = 0; i < 8; i++) { //dw18-25 (Xe2:dw19-26)
diff --git a/lib/intel_batchbuffer.c b/lib/intel_batchbuffer.c
index f418e7981..4f52e7b6a 100644
--- a/lib/intel_batchbuffer.c
+++ b/lib/intel_batchbuffer.c
@@ -329,11 +329,7 @@ void igt_blitter_copy(int fd,
uint32_t dst_x, uint32_t dst_y,
uint64_t dst_size)
{
- uint32_t devid;
-
- devid = intel_get_drm_devid(fd);
-
- if (intel_graphics_ver_legacy(devid) >= IP_VER(12, 60))
+ if (intel_graphics_ver(fd) >= IP_VER(12, 60))
igt_blitter_fast_copy__raw(fd, ahnd, ctx, NULL,
src_handle, src_delta,
src_stride, src_tiling,
@@ -410,7 +406,7 @@ void igt_blitter_src_copy(int fd,
uint32_t batch_handle;
uint32_t src_pitch, dst_pitch;
uint32_t dst_reloc_offset, src_reloc_offset;
- uint32_t gen = intel_gen_legacy(intel_get_drm_devid(fd));
+ uint32_t gen = intel_gen(fd);
uint64_t batch_offset, src_offset, dst_offset;
const bool has_64b_reloc = gen >= 8;
int i = 0;
@@ -669,7 +665,7 @@ igt_render_copyfunc_t igt_get_render_copyfunc(int fd)
copy = mtl_render_copyfunc;
else if (IS_DG2(devid))
copy = gen12p71_render_copyfunc;
- else if (intel_gen_legacy(devid) >= 20)
+ else if (intel_gen(fd) >= 20)
copy = xe2_render_copyfunc;
else if (IS_GEN12(devid))
copy = gen12_render_copyfunc;
@@ -911,7 +907,7 @@ __intel_bb_create(int fd, uint32_t ctx, uint32_t vm, const intel_ctx_cfg_t *cfg,
igt_assert(ibb);
ibb->devid = intel_get_drm_devid(fd);
- ibb->gen = intel_gen_legacy(ibb->devid);
+ ibb->gen = intel_gen(fd);
ibb->ctx = ctx;
ibb->fd = fd;
@@ -1089,7 +1085,7 @@ struct intel_bb *intel_bb_create_with_allocator(int fd, uint32_t ctx, uint32_t v
static bool aux_needs_softpin(int fd)
{
- return intel_gen_legacy(intel_get_drm_devid(fd)) >= 12;
+ return intel_gen(fd) >= 12;
}
static bool has_ctx_cfg(struct intel_bb *ibb)
diff --git a/lib/intel_blt.c b/lib/intel_blt.c
index 673f204b0..7ae04fccd 100644
--- a/lib/intel_blt.c
+++ b/lib/intel_blt.c
@@ -997,7 +997,7 @@ uint64_t emit_blt_block_copy(int fd,
uint64_t bb_pos,
bool emit_bbe)
{
- unsigned int ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ unsigned int ip_ver = intel_graphics_ver(fd);
struct gen12_block_copy_data data = {};
struct gen12_block_copy_data_ext dext = {};
uint64_t dst_offset, src_offset, bb_offset;
@@ -1285,7 +1285,7 @@ uint64_t emit_blt_ctrl_surf_copy(int fd,
uint64_t bb_pos,
bool emit_bbe)
{
- unsigned int ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ unsigned int ip_ver = intel_graphics_ver(fd);
union ctrl_surf_copy_data data = { };
size_t data_sz;
uint64_t dst_offset, src_offset, bb_offset, alignment;
@@ -1705,7 +1705,7 @@ uint64_t emit_blt_fast_copy(int fd,
uint64_t bb_pos,
bool emit_bbe)
{
- unsigned int ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ unsigned int ip_ver = intel_graphics_ver(fd);
struct gen12_fast_copy_data data = {};
uint64_t dst_offset, src_offset, bb_offset;
uint32_t bbe = MI_BATCH_BUFFER_END;
@@ -1972,11 +1972,10 @@ void blt_mem_copy_init(int fd, struct blt_mem_copy_data *mem,
static void dump_bb_mem_copy_cmd(int fd, struct xe_mem_copy_data *data)
{
uint32_t *cmd = (uint32_t *) data;
- uint32_t devid = intel_get_drm_devid(fd);
igt_info("BB details:\n");
- if (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0)) {
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0)) {
igt_info(" dw00: [%08x] <client: 0x%x, opcode: 0x%x, length: %d> "
"[copy type: %d, mode: %d]\n",
cmd[0], data->dw00.xe2.client, data->dw00.xe2.opcode,
@@ -2006,7 +2005,7 @@ static void dump_bb_mem_copy_cmd(int fd, struct xe_mem_copy_data *data)
cmd[7], data->dw07.dst_address_lo);
igt_info(" dw08: [%08x] dst offset hi (0x%x)\n",
cmd[8], data->dw08.dst_address_hi);
- if (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0)) {
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0)) {
igt_info(" dw09: [%08x] mocs <dst: 0x%x, src: 0x%x>\n",
cmd[9], data->dw09.xe2.dst_mocs,
data->dw09.xe2.src_mocs);
@@ -2025,7 +2024,6 @@ static uint64_t emit_blt_mem_copy(int fd, uint64_t ahnd,
uint64_t dst_offset, src_offset, shift;
uint32_t width, height, width_max, height_max, remain;
uint32_t bbe = MI_BATCH_BUFFER_END;
- uint32_t devid = intel_get_drm_devid(fd);
uint8_t *bb;
if (mem->mode == MODE_BYTE) {
@@ -2049,7 +2047,7 @@ static uint64_t emit_blt_mem_copy(int fd, uint64_t ahnd,
width = mem->src.width;
height = mem->dst.height;
- if (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0)) {
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0)) {
data.dw00.xe2.client = 0x2;
data.dw00.xe2.opcode = 0x5a;
data.dw00.xe2.length = 8;
@@ -2231,7 +2229,6 @@ static void emit_blt_mem_set(int fd, uint64_t ahnd,
int b;
uint32_t *batch;
uint32_t value;
- uint32_t devid = intel_get_drm_devid(fd);
dst_offset = get_offset_pat_index(ahnd, mem->dst.handle, mem->dst.size,
0, mem->dst.pat_index);
@@ -2246,7 +2243,7 @@ static void emit_blt_mem_set(int fd, uint64_t ahnd,
batch[b++] = mem->dst.pitch - 1;
batch[b++] = dst_offset;
batch[b++] = dst_offset << 32;
- if (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0))
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0))
batch[b++] = value | (mem->dst.mocs_index << 3);
else
batch[b++] = value | mem->dst.mocs_index;
@@ -2364,7 +2361,7 @@ blt_create_object(const struct blt_copy_data *blt, uint32_t region,
if (create_mapping && region != system_memory(blt->fd))
flags |= DRM_XE_GEM_CREATE_FLAG_NEEDS_VISIBLE_VRAM;
- if (intel_gen_legacy(intel_get_drm_devid(blt->fd)) >= 20 && compression) {
+ if (intel_gen(blt->fd) >= 20 && compression) {
pat_index = intel_get_pat_idx_uc_comp(blt->fd);
cpu_caching = DRM_XE_GEM_CPU_CACHING_WC;
}
@@ -2590,7 +2587,7 @@ void blt_surface_get_flatccs_data(int fd,
cpu_caching = __xe_default_cpu_caching(fd, sysmem, 0);
ccs_bo_size = ALIGN(ccssize, xe_get_default_alignment(fd));
- if (intel_gen_legacy(intel_get_drm_devid(fd)) >= 20 && obj->compression) {
+ if (intel_gen(fd) >= 20 && obj->compression) {
comp_pat_index = intel_get_pat_idx_uc_comp(fd);
cpu_caching = DRM_XE_GEM_CPU_CACHING_WC;
}
diff --git a/lib/intel_blt.h b/lib/intel_blt.h
index a98a34e95..feba94ebb 100644
--- a/lib/intel_blt.h
+++ b/lib/intel_blt.h
@@ -52,7 +52,7 @@
#include "igt.h"
#include "intel_cmds_info.h"
-#define CCS_RATIO(fd) (intel_gen_legacy(intel_get_drm_devid(fd)) >= 20 ? 512 : 256)
+#define CCS_RATIO(fd) (intel_gen(fd) >= 20 ? 512 : 256)
#define GEN12_MEM_COPY_MOCS_SHIFT 25
#define XE2_MEM_COPY_SRC_MOCS_SHIFT 28
#define XE2_MEM_COPY_DST_MOCS_SHIFT 3
diff --git a/lib/intel_bufops.c b/lib/intel_bufops.c
index ea3742f1e..a2adbf9ef 100644
--- a/lib/intel_bufops.c
+++ b/lib/intel_bufops.c
@@ -1063,7 +1063,7 @@ static void __intel_buf_init(struct buf_ops *bops,
} else {
uint16_t cpu_caching = __xe_default_cpu_caching(bops->fd, region, 0);
- if (intel_gen_legacy(bops->devid) >= 20 && compression)
+ if (intel_gen(bops->fd) >= 20 && compression)
cpu_caching = DRM_XE_GEM_CPU_CACHING_WC;
bo_size = ALIGN(bo_size, xe_get_default_alignment(bops->fd));
@@ -1106,7 +1106,7 @@ void intel_buf_init(struct buf_ops *bops,
uint64_t region;
uint8_t pat_index = DEFAULT_PAT_INDEX;
- if (compression && intel_gen_legacy(bops->devid) >= 20)
+ if (compression && intel_gen(bops->fd) >= 20)
pat_index = intel_get_pat_idx_uc_comp(bops->fd);
region = bops->driver == INTEL_DRIVER_I915 ? I915_SYSTEM_MEMORY :
@@ -1132,7 +1132,7 @@ void intel_buf_init_in_region(struct buf_ops *bops,
{
uint8_t pat_index = DEFAULT_PAT_INDEX;
- if (compression && intel_gen_legacy(bops->devid) >= 20)
+ if (compression && intel_gen(bops->fd) >= 20)
pat_index = intel_get_pat_idx_uc_comp(bops->fd);
__intel_buf_init(bops, 0, buf, width, height, bpp, alignment,
@@ -1203,7 +1203,7 @@ void intel_buf_init_using_handle_and_size(struct buf_ops *bops,
igt_assert(handle);
igt_assert(size);
- if (compression && intel_gen_legacy(bops->devid) >= 20)
+ if (compression && intel_gen(bops->fd) >= 20)
pat_index = intel_get_pat_idx_uc_comp(bops->fd);
__intel_buf_init(bops, handle, buf, width, height, bpp, alignment,
@@ -1758,7 +1758,7 @@ static struct buf_ops *__buf_ops_create(int fd, bool check_idempotency)
igt_assert(bops);
devid = intel_get_drm_devid(fd);
- generation = intel_gen_legacy(devid);
+ generation = intel_gen(fd);
/* Predefined settings: see intel_device_info? */
for (int i = 0; i < ARRAY_SIZE(buf_ops_arr); i++) {
diff --git a/lib/intel_common.c b/lib/intel_common.c
index cd1019bfe..407d53f77 100644
--- a/lib/intel_common.c
+++ b/lib/intel_common.c
@@ -91,7 +91,7 @@ bool is_intel_region_compressible(int fd, uint64_t region)
return true;
/* Integrated Xe2+ supports compression on system memory */
- if (intel_gen_legacy(devid) >= 20 && !is_dgfx && is_intel_system_region(fd, region))
+ if (intel_gen(fd) >= 20 && !is_dgfx && is_intel_system_region(fd, region))
return true;
/* Discrete supports compression on vram */
diff --git a/lib/intel_compute.c b/lib/intel_compute.c
index 1734c1649..66156d194 100644
--- a/lib/intel_compute.c
+++ b/lib/intel_compute.c
@@ -2284,7 +2284,7 @@ static bool __run_intel_compute_kernel(int fd,
struct user_execenv *user,
enum execenv_alloc_prefs alloc_prefs)
{
- unsigned int ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ unsigned int ip_ver = intel_graphics_ver(fd);
int batch;
const struct intel_compute_kernels *kernel_entries = intel_compute_square_kernels, *kernels;
enum intel_driver driver = get_intel_driver(fd);
@@ -2749,7 +2749,7 @@ static bool __run_intel_compute_kernel_preempt(int fd,
bool threadgroup_preemption,
enum execenv_alloc_prefs alloc_prefs)
{
- unsigned int ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ unsigned int ip_ver = ip_ver = intel_graphics_ver(fd);
int batch;
const struct intel_compute_kernels *kernel_entries = intel_compute_square_kernels, *kernels;
enum intel_driver driver = get_intel_driver(fd);
@@ -2803,7 +2803,7 @@ static bool __run_intel_compute_kernel_preempt(int fd,
*/
bool xe_kernel_preempt_check(int fd, enum xe_compute_preempt_type required_preempt)
{
- unsigned int ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ unsigned int ip_ver = ip_ver = intel_graphics_ver(fd);
int batch = find_preempt_batch(ip_ver);
if (batch < 0) {
diff --git a/lib/intel_mocs.c b/lib/intel_mocs.c
index f21c2bf09..b9ea43c7c 100644
--- a/lib/intel_mocs.c
+++ b/lib/intel_mocs.c
@@ -27,8 +27,8 @@ struct drm_intel_mocs_index {
static void get_mocs_index(int fd, struct drm_intel_mocs_index *mocs)
{
- uint16_t devid = intel_get_drm_devid(fd);
- unsigned int ip_ver = intel_graphics_ver_legacy(devid);
+ uint16_t devid;
+ unsigned int ip_ver = intel_graphics_ver(fd);
/*
* Gen >= 12 onwards don't have a setting for PTE,
@@ -42,26 +42,29 @@ static void get_mocs_index(int fd, struct drm_intel_mocs_index *mocs)
mocs->wb_index = 4;
mocs->displayable_index = 1;
mocs->defer_to_pat_index = 0;
- } else if (IS_METEORLAKE(devid)) {
- mocs->uc_index = 5;
- mocs->wb_index = 1;
- mocs->displayable_index = 14;
- } else if (IS_DG2(devid)) {
- mocs->uc_index = 1;
- mocs->wb_index = 3;
- mocs->displayable_index = 3;
- } else if (IS_DG1(devid)) {
- mocs->uc_index = 1;
- mocs->wb_index = 5;
- mocs->displayable_index = 5;
- } else if (ip_ver >= IP_VER(12, 0)) {
- mocs->uc_index = 3;
- mocs->wb_index = 2;
- mocs->displayable_index = 61;
} else {
- mocs->uc_index = I915_MOCS_PTE;
- mocs->wb_index = I915_MOCS_CACHED;
- mocs->displayable_index = I915_MOCS_PTE;
+ devid = intel_get_drm_devid(fd);
+ if (IS_METEORLAKE(devid)) {
+ mocs->uc_index = 5;
+ mocs->wb_index = 1;
+ mocs->displayable_index = 14;
+ } else if (IS_DG2(devid)) {
+ mocs->uc_index = 1;
+ mocs->wb_index = 3;
+ mocs->displayable_index = 3;
+ } else if (IS_DG1(devid)) {
+ mocs->uc_index = 1;
+ mocs->wb_index = 5;
+ mocs->displayable_index = 5;
+ } else if (ip_ver >= IP_VER(12, 0)) {
+ mocs->uc_index = 3;
+ mocs->wb_index = 2;
+ mocs->displayable_index = 61;
+ } else {
+ mocs->uc_index = I915_MOCS_PTE;
+ mocs->wb_index = I915_MOCS_CACHED;
+ mocs->displayable_index = I915_MOCS_PTE;
+ }
}
}
@@ -124,9 +127,8 @@ uint8_t intel_get_displayable_mocs_index(int fd)
uint8_t intel_get_defer_to_pat_mocs_index(int fd)
{
struct drm_intel_mocs_index mocs;
- uint16_t dev_id = intel_get_drm_devid(fd);
- igt_assert(intel_gen_legacy(dev_id) >= 20);
+ igt_assert(intel_gen(fd) >= 20);
get_mocs_index(fd, &mocs);
diff --git a/lib/intel_pat.c b/lib/intel_pat.c
index 9a61c2a45..9bb4800b6 100644
--- a/lib/intel_pat.c
+++ b/lib/intel_pat.c
@@ -96,14 +96,12 @@ int32_t xe_get_pat_sw_config(int drm_fd, struct intel_pat_cache *xe_pat_cache)
static void intel_get_pat_idx(int fd, struct intel_pat_cache *pat)
{
- uint16_t dev_id = intel_get_drm_devid(fd);
-
- if (intel_graphics_ver_legacy(dev_id) == IP_VER(35, 11)) {
+ if (intel_graphics_ver(fd) == IP_VER(35, 11)) {
pat->uc = 3;
pat->wb = 2;
pat->max_index = 31;
- } else if (intel_get_device_info(dev_id)->graphics_ver == 30 ||
- intel_get_device_info(dev_id)->graphics_ver == 20) {
+ } else if (intel_gen(fd) == 30 ||
+ intel_gen(fd) == 20) {
pat->uc = 3;
pat->wt = 15; /* Compressed + WB-transient */
pat->wb = 2;
@@ -111,19 +109,19 @@ static void intel_get_pat_idx(int fd, struct intel_pat_cache *pat)
pat->max_index = 31;
/* Wa_16023588340: CLOS3 entries at end of table are unusable */
- if (intel_graphics_ver_legacy(dev_id) == IP_VER(20, 1))
+ if (intel_graphics_ver(fd) == IP_VER(20, 1))
pat->max_index -= 4;
- } else if (IS_METEORLAKE(dev_id)) {
+ } else if (IS_METEORLAKE(intel_get_drm_devid(fd))) {
pat->uc = 2;
pat->wt = 1;
pat->wb = 3;
pat->max_index = 3;
- } else if (IS_PONTEVECCHIO(dev_id)) {
+ } else if (IS_PONTEVECCHIO(intel_get_drm_devid(fd))) {
pat->uc = 0;
pat->wt = 2;
pat->wb = 3;
pat->max_index = 7;
- } else if (intel_graphics_ver_legacy(dev_id) <= IP_VER(12, 60)) {
+ } else if (intel_graphics_ver(fd) <= IP_VER(12, 60)) {
pat->uc = 3;
pat->wt = 2;
pat->wb = 0;
@@ -152,9 +150,8 @@ uint8_t intel_get_pat_idx_uc(int fd)
uint8_t intel_get_pat_idx_uc_comp(int fd)
{
struct intel_pat_cache pat = {};
- uint16_t dev_id = intel_get_drm_devid(fd);
- igt_assert(intel_gen_legacy(dev_id) >= 20);
+ igt_assert(intel_gen(fd) >= 20);
intel_get_pat_idx(fd, &pat);
return pat.uc_comp;
diff --git a/lib/rendercopy_gen9.c b/lib/rendercopy_gen9.c
index 66415212c..0be557a47 100644
--- a/lib/rendercopy_gen9.c
+++ b/lib/rendercopy_gen9.c
@@ -256,12 +256,12 @@ gen9_bind_buf(struct intel_bb *ibb, const struct intel_buf *buf, int is_dst,
if (buf->compression == I915_COMPRESSION_MEDIA)
ss->ss7.tgl.media_compression = 1;
else if (buf->compression == I915_COMPRESSION_RENDER) {
- if (intel_gen_legacy(ibb->devid) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
ss->ss6.aux_mode = 0x0; /* AUX_NONE, unified compression */
else
ss->ss6.aux_mode = 0x5; /* AUX_CCS_E */
- if (intel_gen_legacy(ibb->devid) < 12 && buf->ccs[0].stride) {
+ if (intel_gen(ibb->fd) < 12 && buf->ccs[0].stride) {
ss->ss6.aux_pitch = (buf->ccs[0].stride / 128) - 1;
address = intel_bb_offset_reloc_with_delta(ibb, buf->handle,
@@ -303,7 +303,7 @@ gen9_bind_buf(struct intel_bb *ibb, const struct intel_buf *buf, int is_dst,
ss->ss7.dg2.disable_support_for_multi_gpu_partial_writes = 1;
ss->ss7.dg2.disable_support_for_multi_gpu_atomics = 1;
- if (intel_gen_legacy(ibb->devid) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
ss->ss12.lnl.compression_format = lnl_compression_format(buf);
else
ss->ss12.dg2.compression_format = dg2_compression_format(buf);
@@ -681,7 +681,7 @@ gen9_emit_state_base_address(struct intel_bb *ibb) {
/* WaBindlessSurfaceStateModifyEnable:skl,bxt */
/* The length has to be one less if we dont modify
bindless state */
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
intel_bb_out(ibb, GEN4_STATE_BASE_ADDRESS | 20);
else
intel_bb_out(ibb, GEN4_STATE_BASE_ADDRESS | (19 - 1 - 2));
@@ -726,7 +726,7 @@ gen9_emit_state_base_address(struct intel_bb *ibb) {
intel_bb_out(ibb, 0);
intel_bb_out(ibb, 0);
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20) {
+ if (intel_gen(ibb->fd) >= 20) {
/* Bindless sampler */
intel_bb_out(ibb, 0);
intel_bb_out(ibb, 0);
@@ -899,7 +899,7 @@ gen9_emit_ds(struct intel_bb *ibb) {
static void
gen8_emit_wm_hz_op(struct intel_bb *ibb) {
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20) {
+ if (intel_gen(ibb->fd) >= 20) {
intel_bb_out(ibb, GEN8_3DSTATE_WM_HZ_OP | (6-2));
intel_bb_out(ibb, 0);
} else {
@@ -989,7 +989,7 @@ gen8_emit_ps(struct intel_bb *ibb, uint32_t kernel, bool fast_clear) {
intel_bb_out(ibb, 0);
intel_bb_out(ibb, GEN7_3DSTATE_PS | (12-2));
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
intel_bb_out(ibb, kernel | 1);
else
intel_bb_out(ibb, kernel);
@@ -1006,7 +1006,7 @@ gen8_emit_ps(struct intel_bb *ibb, uint32_t kernel, bool fast_clear) {
intel_bb_out(ibb, (max_threads - 1) << GEN8_3DSTATE_PS_MAX_THREADS_SHIFT |
GEN6_3DSTATE_WM_16_DISPATCH_ENABLE |
(fast_clear ? GEN8_3DSTATE_FAST_CLEAR_ENABLE : 0));
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
intel_bb_out(ibb, 6 << GEN6_3DSTATE_WM_DISPATCH_START_GRF_0_SHIFT |
GENXE_KERNEL0_POLY_PACK16_FIXED << GENXE_KERNEL0_PACKING_POLICY);
else
@@ -1061,7 +1061,7 @@ gen9_emit_depth(struct intel_bb *ibb)
static void
gen7_emit_clear(struct intel_bb *ibb) {
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
return;
intel_bb_out(ibb, GEN7_3DSTATE_CLEAR_PARAMS | (3-2));
@@ -1072,7 +1072,7 @@ gen7_emit_clear(struct intel_bb *ibb) {
static void
gen6_emit_drawing_rectangle(struct intel_bb *ibb, const struct intel_buf *dst)
{
- if (intel_gen_legacy(intel_get_drm_devid(ibb->fd)) >= 20)
+ if (intel_gen(ibb->fd) >= 20)
intel_bb_out(ibb, GENXE2_3DSTATE_DRAWING_RECTANGLE_FAST | (4 - 2));
else
intel_bb_out(ibb, GEN4_3DSTATE_DRAWING_RECTANGLE | (4 - 2));
@@ -1266,7 +1266,7 @@ void _gen9_render_op(struct intel_bb *ibb,
gen9_emit_state_base_address(ibb);
- if (HAS_4TILE(ibb->devid) || intel_gen_legacy(ibb->devid) > 12) {
+ if (HAS_4TILE(ibb->devid) || intel_gen(ibb->fd) > 12) {
intel_bb_out(ibb, GEN4_3DSTATE_BINDING_TABLE_POOL_ALLOC | 2);
intel_bb_emit_reloc(ibb, ibb->handle,
I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION, 0,
diff --git a/lib/xe/xe_legacy.c b/lib/xe/xe_legacy.c
index 1529ed1cc..c1ce9fa00 100644
--- a/lib/xe/xe_legacy.c
+++ b/lib/xe/xe_legacy.c
@@ -75,7 +75,7 @@ xe_legacy_test_mode(int fd, struct drm_xe_engine_class_instance *eci,
igt_assert_lte(n_exec_queues, MAX_N_EXECQUEUES);
if (flags & COMPRESSION)
- igt_require(intel_gen_legacy(intel_get_drm_devid(fd)) >= 20);
+ igt_require(intel_gen(fd) >= 20);
if (flags & CLOSE_FD)
fd = drm_open_driver(DRIVER_XE);
diff --git a/lib/xe/xe_spin.c b/lib/xe/xe_spin.c
index 36260e3e5..8ca137381 100644
--- a/lib/xe/xe_spin.c
+++ b/lib/xe/xe_spin.c
@@ -54,7 +54,6 @@ void xe_spin_init(struct xe_spin *spin, struct xe_spin_opts *opts)
uint64_t pad_addr = opts->addr + offsetof(struct xe_spin, pad);
uint64_t timestamp_addr = opts->addr + offsetof(struct xe_spin, timestamp);
int b = 0;
- uint32_t devid;
spin->start = 0;
spin->end = 0xffffffff;
@@ -166,8 +165,7 @@ void xe_spin_init(struct xe_spin *spin, struct xe_spin_opts *opts)
spin->batch[b++] = opts->mem_copy->dst_offset;
spin->batch[b++] = opts->mem_copy->dst_offset << 32;
- devid = intel_get_drm_devid(opts->mem_copy->fd);
- if (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0))
+ if (intel_graphics_ver(opts->mem_copy->fd) >= IP_VER(20, 0))
spin->batch[b++] = opts->mem_copy->src->mocs_index << XE2_MEM_COPY_SRC_MOCS_SHIFT |
opts->mem_copy->dst->mocs_index << XE2_MEM_COPY_DST_MOCS_SHIFT;
else
diff --git a/lib/xe/xe_sriov_provisioning.c b/lib/xe/xe_sriov_provisioning.c
index 7b60ccd6c..3d981766c 100644
--- a/lib/xe/xe_sriov_provisioning.c
+++ b/lib/xe/xe_sriov_provisioning.c
@@ -50,9 +50,7 @@ const char *xe_sriov_shared_res_to_string(enum xe_sriov_shared_res res)
static uint64_t get_vfid_mask(int fd)
{
- uint16_t dev_id = intel_get_drm_devid(fd);
-
- return (intel_graphics_ver_legacy(dev_id) >= IP_VER(12, 50)) ?
+ return (intel_graphics_ver(fd) >= IP_VER(12, 50)) ?
GGTT_PTE_VFID_MASK : PRE_1250_IP_VER_GGTT_PTE_VFID_MASK;
}
diff --git a/tests/intel/api_intel_allocator.c b/tests/intel/api_intel_allocator.c
index 869e5e9a0..6b1d17da7 100644
--- a/tests/intel/api_intel_allocator.c
+++ b/tests/intel/api_intel_allocator.c
@@ -625,7 +625,7 @@ static void execbuf_with_allocator(int fd)
uint64_t ahnd, sz = 4096, gtt_size;
unsigned int flags = EXEC_OBJECT_PINNED;
uint32_t *ptr, batch[32], copied;
- int gen = intel_gen_legacy(intel_get_drm_devid(fd));
+ int gen = intel_gen(fd);
int i;
const uint32_t magic = 0x900df00d;
diff --git a/tests/intel/kms_ccs.c b/tests/intel/kms_ccs.c
index 30f2c9465..a0373316a 100644
--- a/tests/intel/kms_ccs.c
+++ b/tests/intel/kms_ccs.c
@@ -565,7 +565,7 @@ static void access_flat_ccs_surface(struct igt_fb *fb, bool verify_compression)
uint16_t cpu_caching = DRM_XE_GEM_CPU_CACHING_WC;
uint8_t uc_mocs = intel_get_uc_mocs_index(fb->fd);
uint8_t comp_pat_index = intel_get_pat_idx_wt(fb->fd);
- uint32_t region = (intel_gen_legacy(intel_get_drm_devid(fb->fd)) >= 20 &&
+ uint32_t region = (intel_gen(fb->fd) >= 20 &&
xe_has_vram(fb->fd)) ? REGION_LMEM(0) : REGION_SMEM;
struct drm_xe_engine_class_instance inst = {
@@ -645,7 +645,7 @@ static void fill_fb_random(int drm_fd, igt_fb_t *fb)
igt_assert_eq(0, gem_munmap(map, fb->size));
/* randomize also ccs surface on Xe2 */
- if (intel_gen_legacy(intel_get_drm_devid(drm_fd)) >= 20)
+ if (intel_gen(drm_fd) >= 20)
access_flat_ccs_surface(fb, false);
}
@@ -1125,11 +1125,6 @@ static bool valid_modifier_test(u64 modifier, const enum test_flags flags)
static void test_output(data_t *data, const int testnum)
{
- uint16_t dev_id;
-
- igt_fixture()
- dev_id = intel_get_drm_devid(data->drm_fd);
-
data->flags = tests[testnum].flags;
for (int i = 0; i < ARRAY_SIZE(ccs_modifiers); i++) {
@@ -1143,10 +1138,10 @@ static void test_output(data_t *data, const int testnum)
igt_subtest_with_dynamic_f("%s-%s", tests[testnum].testname, ccs_modifiers[i].str) {
if (ccs_modifiers[i].modifier == I915_FORMAT_MOD_4_TILED_BMG_CCS ||
ccs_modifiers[i].modifier == I915_FORMAT_MOD_4_TILED_LNL_CCS) {
- igt_require_f(intel_gen_legacy(dev_id) >= 20,
+ igt_require_f(intel_gen(data->drm_fd) >= 20,
"Xe2 platform needed.\n");
} else {
- igt_require_f(intel_gen_legacy(dev_id) < 20,
+ igt_require_f(intel_gen(data->drm_fd) < 20,
"Older than Xe2 platform needed.\n");
}
diff --git a/tests/intel/kms_fbcon_fbt.c b/tests/intel/kms_fbcon_fbt.c
index edf5c0d1b..b28961417 100644
--- a/tests/intel/kms_fbcon_fbt.c
+++ b/tests/intel/kms_fbcon_fbt.c
@@ -179,7 +179,7 @@ static bool fbc_wait_until_update(struct drm_info *drm)
* For older GENs FBC is still expected to be disabled as it still
* relies on a tiled and fenceable framebuffer to track modifications.
*/
- if (intel_gen_legacy(intel_get_drm_devid(drm->fd)) >= 9) {
+ if (intel_gen(drm->fd) >= 9) {
if (!fbc_wait_until_enabled(drm->debugfs_fd))
return false;
/*
diff --git a/tests/intel/kms_frontbuffer_tracking.c b/tests/intel/kms_frontbuffer_tracking.c
index c8c2ce240..5b60587db 100644
--- a/tests/intel/kms_frontbuffer_tracking.c
+++ b/tests/intel/kms_frontbuffer_tracking.c
@@ -3062,13 +3062,13 @@ static bool tiling_is_valid(int feature_flags, enum tiling_type tiling)
switch (tiling) {
case TILING_LINEAR:
- return intel_gen_legacy(drm.devid) >= 9;
+ return intel_gen(drm.fd) >= 9;
case TILING_X:
return (intel_get_device_info(drm.devid)->display_ver > 29) ? false : true;
case TILING_Y:
return true;
case TILING_4:
- return intel_gen_legacy(drm.devid) >= 12;
+ return intel_gen(drm.fd) >= 12;
default:
igt_assert(false);
return false;
@@ -4475,7 +4475,7 @@ int igt_main_args("", long_options, help_str, opt_handler, NULL)
igt_require(igt_draw_supports_method(drm.fd, t.method));
if (t.tiling == TILING_Y) {
- igt_require(intel_gen_legacy(drm.devid) >= 9);
+ igt_require(intel_gen(drm.fd) >= 9);
igt_require(!intel_get_device_info(drm.devid)->has_4tile);
}
diff --git a/tests/intel/kms_pipe_stress.c b/tests/intel/kms_pipe_stress.c
index 1ae32d5fd..f8c994d07 100644
--- a/tests/intel/kms_pipe_stress.c
+++ b/tests/intel/kms_pipe_stress.c
@@ -822,7 +822,7 @@ static void prepare_test(struct data *data)
create_framebuffers(data);
- if (intel_gen_legacy(intel_get_drm_devid(data->drm_fd)) > 9)
+ if (intel_gen(data->drm_fd) > 9)
start_gpu_threads(data);
}
@@ -830,7 +830,7 @@ static void finish_test(struct data *data)
{
int i;
- if (intel_gen_legacy(intel_get_drm_devid(data->drm_fd)) > 9)
+ if (intel_gen(data->drm_fd) > 9)
stop_gpu_threads(data);
/*
diff --git a/tests/intel/xe_ccs.c b/tests/intel/xe_ccs.c
index 914144270..0ba8ae48c 100644
--- a/tests/intel/xe_ccs.c
+++ b/tests/intel/xe_ccs.c
@@ -128,7 +128,7 @@ static void surf_copy(int xe,
int result;
igt_assert(mid->compression);
- if (intel_gen_legacy(devid) >= 20 && mid->compression) {
+ if (intel_gen(xe) >= 20 && mid->compression) {
comp_pat_index = intel_get_pat_idx_uc_comp(xe);
cpu_caching = DRM_XE_GEM_CPU_CACHING_WC;
}
@@ -177,7 +177,7 @@ static void surf_copy(int xe,
if (IS_GEN(devid, 12) && is_intel_dgfx(xe)) {
igt_assert(!strcmp(orig, newsum));
igt_assert(!strcmp(orig2, newsum2));
- } else if (intel_gen_legacy(devid) >= 20) {
+ } else if (intel_gen(xe) >= 20) {
if (is_intel_dgfx(xe)) {
/* buffer object would become
* uncompressed in xe2+ dgfx
@@ -227,7 +227,7 @@ static void surf_copy(int xe,
* uncompressed in xe2+ dgfx, and therefore retrieve the
* ccs by copying 0 to ccsmap
*/
- if (suspend_resume && intel_gen_legacy(devid) >= 20 && is_intel_dgfx(xe))
+ if (suspend_resume && intel_gen(xe) >= 20 && is_intel_dgfx(xe))
memset(ccsmap, 0, ccssize);
else
/* retrieve back ccs */
@@ -353,7 +353,7 @@ static void block_copy(int xe,
uint64_t bb_size = xe_bb_size(xe, SZ_4K);
uint64_t ahnd = intel_allocator_open(xe, ctx->vm, INTEL_ALLOCATOR_RELOC);
uint32_t run_id = mid_tiling;
- uint32_t mid_region = (intel_gen_legacy(intel_get_drm_devid(xe)) >= 20 &&
+ uint32_t mid_region = (intel_gen(xe) >= 20 &&
!xe_has_vram(xe)) ? region1 : region2;
uint32_t bb;
enum blt_compression mid_compression = config->compression;
@@ -441,7 +441,7 @@ static void block_copy(int xe,
if (config->inplace) {
uint8_t pat_index = DEFAULT_PAT_INDEX;
- if (intel_gen_legacy(intel_get_drm_devid(xe)) >= 20 && config->compression)
+ if (intel_gen(xe) >= 20 && config->compression)
pat_index = intel_get_pat_idx_uc_comp(xe);
blt_set_object(&blt.dst, mid->handle, dst->size, mid->region, 0,
@@ -488,7 +488,7 @@ static void block_multicopy(int xe,
uint64_t bb_size = xe_bb_size(xe, SZ_4K);
uint64_t ahnd = intel_allocator_open(xe, ctx->vm, INTEL_ALLOCATOR_RELOC);
uint32_t run_id = mid_tiling;
- uint32_t mid_region = (intel_gen_legacy(intel_get_drm_devid(xe)) >= 20 &&
+ uint32_t mid_region = (intel_gen(xe) >= 20 &&
!xe_has_vram(xe)) ? region1 : region2;
uint32_t bb;
enum blt_compression mid_compression = config->compression;
@@ -530,7 +530,7 @@ static void block_multicopy(int xe,
if (config->inplace) {
uint8_t pat_index = DEFAULT_PAT_INDEX;
- if (intel_gen_legacy(intel_get_drm_devid(xe)) >= 20 && config->compression)
+ if (intel_gen(xe) >= 20 && config->compression)
pat_index = intel_get_pat_idx_uc_comp(xe);
blt_set_object(&blt3.dst, mid->handle, dst->size, mid->region,
@@ -715,7 +715,7 @@ static void block_copy_test(int xe,
int tiling, width, height;
- if (intel_gen_legacy(dev_id) >= 20 && config->compression)
+ if (intel_gen(xe) >= 20 && config->compression)
igt_require(HAS_FLATCCS(dev_id));
if (config->compression && !blt_block_copy_supports_compression(xe))
diff --git a/tests/intel/xe_compute.c b/tests/intel/xe_compute.c
index 7b6c39c77..1cb86920f 100644
--- a/tests/intel/xe_compute.c
+++ b/tests/intel/xe_compute.c
@@ -232,7 +232,7 @@ test_compute_kernel_loop(uint64_t loop_duration)
double elapse_time, lower_bound, upper_bound;
fd = drm_open_driver(DRIVER_XE);
- ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ ip_ver = intel_graphics_ver(fd);
kernels = intel_compute_square_kernels;
while (kernels->kernel) {
@@ -335,7 +335,7 @@ igt_check_supported_pipeline(void)
const struct intel_compute_kernels *kernels;
fd = drm_open_driver(DRIVER_XE);
- ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ ip_ver = intel_graphics_ver(fd);
kernels = intel_compute_square_kernels;
drm_close_driver(fd);
@@ -432,7 +432,7 @@ test_eu_busy(uint64_t duration_sec)
fd = drm_open_driver(DRIVER_XE);
- ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(fd));
+ ip_ver = intel_graphics_ver(fd);
kernels = intel_compute_square_kernels;
while (kernels->kernel) {
if (ip_ver == kernels->ip_ver)
@@ -518,7 +518,7 @@ int igt_main()
igt_fixture() {
xe = drm_open_driver(DRIVER_XE);
sriov_enabled = is_sriov_mode(xe);
- ip_ver = intel_graphics_ver_legacy(intel_get_drm_devid(xe));
+ ip_ver = intel_graphics_ver(xe);
igt_store_ccs_mode(ccs_mode, ARRAY_SIZE(ccs_mode));
}
diff --git a/tests/intel/xe_copy_basic.c b/tests/intel/xe_copy_basic.c
index 55081f938..e37bad746 100644
--- a/tests/intel/xe_copy_basic.c
+++ b/tests/intel/xe_copy_basic.c
@@ -261,7 +261,6 @@ const char *help_str =
int igt_main_args("b", NULL, help_str, opt_handler, NULL)
{
int fd;
- uint16_t dev_id;
struct igt_collection *set, *regions;
uint32_t region;
struct rect linear[] = { { 0, 0xfd, 1, MODE_BYTE },
@@ -275,7 +274,6 @@ int igt_main_args("b", NULL, help_str, opt_handler, NULL)
igt_fixture() {
fd = drm_open_driver(DRIVER_XE);
- dev_id = intel_get_drm_devid(fd);
xe_device_get(fd);
set = xe_get_memory_region_set(fd,
DRM_XE_MEM_REGION_CLASS_SYSMEM,
@@ -295,7 +293,7 @@ int igt_main_args("b", NULL, help_str, opt_handler, NULL)
for (int i = 0; i < ARRAY_SIZE(page); i++) {
igt_subtest_f("mem-page-copy-%u", page[i].width) {
igt_require(blt_has_mem_copy(fd));
- igt_require(intel_get_device_info(dev_id)->graphics_ver >= 20);
+ igt_require(intel_gen(fd) >= 20);
for_each_variation_r(regions, 1, set) {
region = igt_collection_get_value(regions, 0);
copy_test(fd, &page[i], MEM_COPY, region);
@@ -320,7 +318,7 @@ int igt_main_args("b", NULL, help_str, opt_handler, NULL)
* till 0x3FFFF.
*/
if (linear[i].width > 0x3ffff &&
- (intel_get_device_info(dev_id)->graphics_ver < 20))
+ (intel_gen(fd) < 20))
igt_skip("Skipping: width exceeds 18-bit limit on gfx_ver < 20\n");
igt_require(blt_has_mem_set(fd));
for_each_variation_r(regions, 1, set) {
diff --git a/tests/intel/xe_debugfs.c b/tests/intel/xe_debugfs.c
index facb55854..4075b173a 100644
--- a/tests/intel/xe_debugfs.c
+++ b/tests/intel/xe_debugfs.c
@@ -296,7 +296,6 @@ static void test_tile_dir(struct xe_device *xe_dev, uint8_t tile)
*/
static void test_info_read(struct xe_device *xe_dev)
{
- uint16_t devid = intel_get_drm_devid(xe_dev->fd);
struct drm_xe_query_config *config;
const char *name = "info";
bool failed = false;
@@ -329,7 +328,7 @@ static void test_info_read(struct xe_device *xe_dev)
failed = true;
}
- if (intel_gen_legacy(devid) < 20) {
+ if (intel_gen(xe_dev->fd) < 20) {
val = -1;
switch (config->info[DRM_XE_QUERY_CONFIG_VA_BITS]) {
diff --git a/tests/intel/xe_eudebug_online.c b/tests/intel/xe_eudebug_online.c
index f64b12b3f..961cf5afc 100644
--- a/tests/intel/xe_eudebug_online.c
+++ b/tests/intel/xe_eudebug_online.c
@@ -400,9 +400,7 @@ static uint64_t eu_ctl(int debugfd, uint64_t client,
static bool intel_gen_needs_resume_wa(int fd)
{
- const uint32_t id = intel_get_drm_devid(fd);
-
- return intel_gen_legacy(id) == 12 && intel_graphics_ver_legacy(id) < IP_VER(12, 55);
+ return intel_gen(fd) == 12 && intel_graphics_ver(fd) < IP_VER(12, 55);
}
static uint64_t eu_ctl_resume(int fd, int debugfd, uint64_t client,
@@ -1222,8 +1220,6 @@ static void run_online_client(struct xe_eudebug_client *c)
static bool intel_gen_has_lockstep_eus(int fd)
{
- const uint32_t id = intel_get_drm_devid(fd);
-
/*
* Lockstep (or in some parlance, fused) EUs are pair of EUs
* that work in sync, supposedly same clock and same control flow.
@@ -1231,7 +1227,7 @@ static bool intel_gen_has_lockstep_eus(int fd)
* excepted into SIP. In this level, the hardware has only one attention
* thread bit for units. PVC is the first one without lockstepping.
*/
- return !(intel_graphics_ver_legacy(id) == IP_VER(12, 60) || intel_gen_legacy(id) >= 20);
+ return !(intel_graphics_ver(fd) == IP_VER(12, 60) || intel_gen(fd) >= 20);
}
static int query_attention_bitmask_size(int fd, int gt)
diff --git a/tests/intel/xe_exec_multi_queue.c b/tests/intel/xe_exec_multi_queue.c
index 1d416efc9..bf09efcc3 100644
--- a/tests/intel/xe_exec_multi_queue.c
+++ b/tests/intel/xe_exec_multi_queue.c
@@ -1047,7 +1047,7 @@ int igt_main()
igt_fixture() {
fd = drm_open_driver(DRIVER_XE);
- igt_require(intel_graphics_ver_legacy(intel_get_drm_devid(fd)) >= IP_VER(35, 0));
+ igt_require(intel_graphics_ver(fd) >= IP_VER(35, 0));
}
igt_subtest_f("sanity")
diff --git a/tests/intel/xe_exec_store.c b/tests/intel/xe_exec_store.c
index 498ab42b7..9e6a96aa8 100644
--- a/tests/intel/xe_exec_store.c
+++ b/tests/intel/xe_exec_store.c
@@ -55,8 +55,7 @@ static void store_dword_batch(struct data *data, uint64_t addr, int value)
data->addr = batch_addr;
}
-static void cond_batch(struct data *data, uint64_t addr, int value,
- uint16_t dev_id)
+static void cond_batch(int fd, struct data *data, uint64_t addr, int value)
{
int b;
uint64_t batch_offset = (char *)&(data->batch) - (char *)data;
@@ -69,7 +68,7 @@ static void cond_batch(struct data *data, uint64_t addr, int value,
data->batch[b++] = sdi_addr;
data->batch[b++] = sdi_addr >> 32;
- if (intel_graphics_ver_legacy(dev_id) >= IP_VER(20, 0))
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0))
data->batch[b++] = MI_MEM_FENCE | MI_WRITE_FENCE;
data->batch[b++] = MI_CONDITIONAL_BATCH_BUFFER_END | MI_DO_COMPARE | 5 << 12 | 2;
@@ -112,8 +111,7 @@ static void persistance_batch(struct data *data, uint64_t addr)
* SUBTEST: basic-all
* Description: Test to verify store dword on all available engines.
*/
-static void basic_inst(int fd, int inst_type, struct drm_xe_engine_class_instance *eci,
- uint16_t dev_id)
+static void basic_inst(int fd, int inst_type, struct drm_xe_engine_class_instance *eci)
{
struct drm_xe_sync sync[2] = {
{ .type = DRM_XE_SYNC_TYPE_SYNCOBJ, .flags = DRM_XE_SYNC_FLAG_SIGNAL, },
@@ -156,7 +154,7 @@ static void basic_inst(int fd, int inst_type, struct drm_xe_engine_class_instanc
else if (inst_type == COND_BATCH) {
/* A random value where it stops at the below value. */
value = 20 + random() % 10;
- cond_batch(data, addr, value, dev_id);
+ cond_batch(fd, data, addr, value);
}
else
igt_assert_f(inst_type < 2, "Entered wrong inst_type.\n");
@@ -416,23 +414,21 @@ int igt_main()
{
struct drm_xe_engine_class_instance *hwe;
int fd;
- uint16_t dev_id;
struct drm_xe_engine *engine;
igt_fixture() {
fd = drm_open_driver(DRIVER_XE);
xe_device_get(fd);
- dev_id = intel_get_drm_devid(fd);
}
igt_subtest("basic-store") {
engine = xe_engine(fd, 1);
- basic_inst(fd, STORE, &engine->instance, dev_id);
+ basic_inst(fd, STORE, &engine->instance);
}
igt_subtest("basic-cond-batch") {
engine = xe_engine(fd, 1);
- basic_inst(fd, COND_BATCH, &engine->instance, dev_id);
+ basic_inst(fd, COND_BATCH, &engine->instance);
}
igt_subtest_with_dynamic("basic-all") {
@@ -441,7 +437,7 @@ int igt_main()
xe_engine_class_string(hwe->engine_class),
hwe->engine_instance,
hwe->gt_id);
- basic_inst(fd, STORE, hwe, dev_id);
+ basic_inst(fd, STORE, hwe);
}
}
diff --git a/tests/intel/xe_fault_injection.c b/tests/intel/xe_fault_injection.c
index 8adc5c15a..57c5a5579 100644
--- a/tests/intel/xe_fault_injection.c
+++ b/tests/intel/xe_fault_injection.c
@@ -486,12 +486,12 @@ vm_bind_fail(int fd, const char pci_slot[], const char function_name[])
* @xe_oa_alloc_regs: xe_oa_alloc_regs
*/
static void
-oa_add_config_fail(int fd, int sysfs, int devid,
+oa_add_config_fail(int fd, int sysfs,
const char pci_slot[], const char function_name[])
{
char path[512];
uint64_t config_id;
-#define SAMPLE_MUX_REG (intel_graphics_ver_legacy(devid) >= IP_VER(20, 0) ? \
+#define SAMPLE_MUX_REG (intel_graphics_ver(fd) >= IP_VER(20, 0) ? \
0x13000 /* PES* */ : 0x9888 /* NOA_WRITE */)
uint32_t mux_regs[] = { SAMPLE_MUX_REG, 0x0 };
@@ -557,7 +557,6 @@ int igt_main_args("I:", NULL, help_str, opt_handler, NULL)
int fd, sysfs;
struct drm_xe_engine_class_instance *hwe;
struct fault_injection_params fault_params;
- static uint32_t devid;
char pci_slot[NAME_MAX];
bool is_vf_device;
const struct section {
@@ -627,7 +626,6 @@ int igt_main_args("I:", NULL, help_str, opt_handler, NULL)
igt_fixture() {
igt_require(fail_function_injection_enabled());
fd = drm_open_driver(DRIVER_XE);
- devid = intel_get_drm_devid(fd);
sysfs = igt_sysfs_open(fd);
igt_device_get_pci_slot_name(fd, pci_slot);
setup_injection_fault(&default_fault_params);
@@ -659,7 +657,7 @@ int igt_main_args("I:", NULL, help_str, opt_handler, NULL)
for (const struct section *s = oa_add_config_fail_functions; s->name; s++)
igt_subtest_f("oa-add-config-fail-%s", s->name)
- oa_add_config_fail(fd, sysfs, devid, pci_slot, s->name);
+ oa_add_config_fail(fd, sysfs, pci_slot, s->name);
igt_fixture() {
igt_kmod_unbind("xe", pci_slot);
diff --git a/tests/intel/xe_intel_bb.c b/tests/intel/xe_intel_bb.c
index 5c112351f..e37d00d2c 100644
--- a/tests/intel/xe_intel_bb.c
+++ b/tests/intel/xe_intel_bb.c
@@ -710,7 +710,7 @@ static void do_intel_bb_blit(struct buf_ops *bops, int loops, uint32_t tiling)
int i, fails = 0, xe = buf_ops_get_fd(bops);
/* We'll fix it for gen2/3 later. */
- igt_require(intel_gen_legacy(intel_get_drm_devid(xe)) > 3);
+ igt_require(intel_gen(xe) > 3);
for (i = 0; i < loops; i++)
fails += __do_intel_bb_blit(bops, tiling);
@@ -878,10 +878,9 @@ static int render(struct buf_ops *bops, uint32_t tiling,
int xe = buf_ops_get_fd(bops);
uint32_t fails = 0;
char name[128];
- uint32_t devid = intel_get_drm_devid(xe);
igt_render_copyfunc_t render_copy = NULL;
- igt_debug("%s() gen: %d\n", __func__, intel_gen_legacy(devid));
+ igt_debug("%s() gen: %d\n", __func__, intel_gen(xe));
ibb = intel_bb_create(xe, PAGE_SIZE);
@@ -1041,7 +1040,7 @@ int igt_main_args("dpib", NULL, help_str, opt_handler, NULL)
do_intel_bb_blit(bops, 3, I915_TILING_X);
igt_subtest("intel-bb-blit-y") {
- igt_require(intel_gen_legacy(intel_get_drm_devid(xe)) >= 6);
+ igt_require(intel_gen(xe) >= 6);
do_intel_bb_blit(bops, 3, I915_TILING_Y);
}
diff --git a/tests/intel/xe_multigpu_svm.c b/tests/intel/xe_multigpu_svm.c
index ab800476e..2c6f81a10 100644
--- a/tests/intel/xe_multigpu_svm.c
+++ b/tests/intel/xe_multigpu_svm.c
@@ -396,7 +396,6 @@ static void batch_init(int fd, uint32_t vm, uint64_t src_addr,
uint64_t batch_addr;
void *batch;
uint32_t *cmd;
- uint16_t dev_id = intel_get_drm_devid(fd);
uint32_t mocs_index = intel_get_uc_mocs_index(fd);
int i = 0;
@@ -412,7 +411,7 @@ static void batch_init(int fd, uint32_t vm, uint64_t src_addr,
cmd[i++] = upper_32_bits(src_addr);
cmd[i++] = lower_32_bits(dst_addr);
cmd[i++] = upper_32_bits(dst_addr);
- if (intel_graphics_ver_legacy(dev_id) >= IP_VER(20, 0)) {
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0)) {
cmd[i++] = mocs_index << XE2_MEM_COPY_SRC_MOCS_SHIFT | mocs_index;
} else {
cmd[i++] = mocs_index << GEN12_MEM_COPY_MOCS_SHIFT | mocs_index;
diff --git a/tests/intel/xe_pat.c b/tests/intel/xe_pat.c
index 96302ad3a..96d544160 100644
--- a/tests/intel/xe_pat.c
+++ b/tests/intel/xe_pat.c
@@ -119,14 +119,13 @@ static int xe_fetch_pat_sw_config(int fd, struct intel_pat_cache *pat_sw_config)
*/
static void pat_sanity(int fd)
{
- uint16_t dev_id = intel_get_drm_devid(fd);
struct intel_pat_cache pat_sw_config = {};
int32_t parsed;
bool has_uc_comp = false, has_wt = false;
parsed = xe_fetch_pat_sw_config(fd, &pat_sw_config);
- if (intel_graphics_ver_legacy(dev_id) >= IP_VER(20, 0)) {
+ if (intel_graphics_ver(fd) >= IP_VER(20, 0)) {
for (int i = 0; i < parsed; i++) {
uint32_t pat = pat_sw_config.entries[i].pat;
if (pat_sw_config.entries[i].rsvd)
@@ -898,7 +897,6 @@ static void display_vs_wb_transient(int fd)
3, /* UC (baseline) */
6, /* L3:XD (uncompressed) */
};
- uint32_t devid = intel_get_drm_devid(fd);
igt_render_copyfunc_t render_copy = NULL;
igt_crc_t ref_crc = {}, crc = {};
igt_plane_t *primary;
@@ -914,7 +912,7 @@ static void display_vs_wb_transient(int fd)
int bpp = 32;
int i;
- igt_require(intel_get_device_info(devid)->graphics_ver >= 20);
+ igt_require(intel_gen(fd) >= 20);
render_copy = igt_get_render_copyfunc(fd);
igt_require(render_copy);
@@ -1015,10 +1013,8 @@ static uint8_t get_pat_idx_uc(int fd, bool *compressed)
static uint8_t get_pat_idx_wt(int fd, bool *compressed)
{
- uint16_t dev_id = intel_get_drm_devid(fd);
-
if (compressed)
- *compressed = intel_get_device_info(dev_id)->graphics_ver >= 20;
+ *compressed = intel_gen(fd) >= 20;
return intel_get_pat_idx_wt(fd);
}
@@ -1328,7 +1324,7 @@ int igt_main_args("V", NULL, help_str, opt_handler, NULL)
bo_comp_disable_bind(fd);
igt_subtest_with_dynamic("pat-index-xelp") {
- igt_require(intel_graphics_ver_legacy(dev_id) <= IP_VER(12, 55));
+ igt_require(intel_graphics_ver(fd) <= IP_VER(12, 55));
subtest_pat_index_modes_with_regions(fd, xelp_pat_index_modes,
ARRAY_SIZE(xelp_pat_index_modes));
}
@@ -1346,10 +1342,10 @@ int igt_main_args("V", NULL, help_str, opt_handler, NULL)
}
igt_subtest_with_dynamic("pat-index-xe2") {
- igt_require(intel_get_device_info(dev_id)->graphics_ver >= 20);
+ igt_require(intel_gen(fd) >= 20);
igt_assert(HAS_FLATCCS(dev_id));
- if (intel_graphics_ver_legacy(dev_id) == IP_VER(20, 1))
+ if (intel_graphics_ver(fd) == IP_VER(20, 1))
subtest_pat_index_modes_with_regions(fd, bmg_g21_pat_index_modes,
ARRAY_SIZE(bmg_g21_pat_index_modes));
else
diff --git a/tests/intel/xe_query.c b/tests/intel/xe_query.c
index 318a9994a..ae505a5d7 100644
--- a/tests/intel/xe_query.c
+++ b/tests/intel/xe_query.c
@@ -380,7 +380,7 @@ test_query_gt_topology(int fd)
}
/* sanity check EU type */
- if (IS_PONTEVECCHIO(dev_id) || intel_gen_legacy(dev_id) >= 20) {
+ if (IS_PONTEVECCHIO(dev_id) || intel_gen(fd) >= 20) {
igt_assert(topo_types & (1 << DRM_XE_TOPO_SIMD16_EU_PER_DSS));
igt_assert_eq(topo_types & (1 << DRM_XE_TOPO_EU_PER_DSS), 0);
} else {
@@ -428,7 +428,7 @@ test_query_gt_topology_l3_bank_mask(int fd)
}
igt_info(" count: %d\n", count);
- if (intel_get_device_info(dev_id)->graphics_ver < 20) {
+ if (intel_gen(fd) < 20) {
igt_assert_lt(0, count);
}
diff --git a/tests/intel/xe_render_copy.c b/tests/intel/xe_render_copy.c
index 0a6ae9ca2..a3976b5f1 100644
--- a/tests/intel/xe_render_copy.c
+++ b/tests/intel/xe_render_copy.c
@@ -136,7 +136,7 @@ static int compare_bufs(struct intel_buf *buf1, struct intel_buf *buf2,
static bool buf_is_aux_compressed(struct buf_ops *bops, struct intel_buf *buf)
{
int xe = buf_ops_get_fd(bops);
- unsigned int gen = intel_gen_legacy(buf_ops_get_devid(bops));
+ unsigned int gen = intel_gen(buf_ops_get_fd(bops));
uint32_t ccs_size;
uint8_t *ptr;
bool is_compressed = false;
--
2.43.0
next prev parent reply other threads:[~2026-01-22 7:15 UTC|newest]
Thread overview: 12+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-01-22 7:15 [PATCH v2 0/3] lib/intel: switch graphics/IP version queries to fd-based APIs Xin Wang
2026-01-22 7:15 ` [PATCH v2 1/3] lib/intel: suffix PCI ID based gen/graphics_ver with _legacy Xin Wang
2026-02-04 18:30 ` Matt Roper
2026-01-22 7:15 ` [PATCH v2 2/3] lib/intel: add fd-based intel_gen/intel_graphics_ver via Xe query Xin Wang
2026-02-05 9:09 ` Jani Nikula
2026-01-22 7:15 ` Xin Wang [this message]
2026-02-04 18:56 ` [PATCH v2 3/3] intel/xe: use fd-based graphics/IP version helpers Matt Roper
2026-02-25 8:51 ` Wang, X
2026-02-25 23:18 ` Matt Roper
2026-01-22 8:01 ` ✓ i915.CI.BAT: success for lib/intel: switch graphics/IP version queries to fd-based APIs (rev2) Patchwork
2026-01-22 8:04 ` ✓ Xe.CI.BAT: " Patchwork
2026-01-22 18:13 ` ✗ Xe.CI.Full: failure " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260122071530.1629046-4-x.wang@intel.com \
--to=x.wang@intel.com \
--cc=igt-dev@lists.freedesktop.org \
--cc=kamil.konieczny@linux.intel.com \
--cc=matthew.d.roper@intel.com \
--cc=ravi.kumar.vodapalli@intel.com \
--cc=zbigniew.kempczynski@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox