From: Matthew Brost <matthew.brost@intel.com>
To: Matthew Auld <matthew.auld@intel.com>
Cc: <intel-xe@lists.freedesktop.org>
Subject: Re: [PATCH v3 7/7] drm/xe/migrate: skip bounce buffer path on xe2
Date: Wed, 22 Oct 2025 10:59:57 -0700 [thread overview]
Message-ID: <aPkbnWE7pLRODu+i@lstrano-desk.jf.intel.com> (raw)
In-Reply-To: <20251022163836.191405-8-matthew.auld@intel.com>
On Wed, Oct 22, 2025 at 05:38:36PM +0100, Matthew Auld wrote:
> Now that we support MEM_COPY we should be able to use the PAGE_COPY
> mode, otherwise falling back to BYTE_COPY mode when we have odd
> sizing/alignment.
>
> v2:
> - Use info.has_mem_copy_instr
> - Rebase on latest changes.
> v3 (Matt Brost):
> - Allow various pitches including 1byte pitch for MEM_COPY
>
> Signed-off-by: Matthew Auld <matthew.auld@intel.com>
> Cc: Matthew Brost <matthew.brost@intel.com>
Reviewed-by: Matthew Brost <matthew.brost@intel.com>
> ---
> drivers/gpu/drm/xe/xe_migrate.c | 43 ++++++++++++++++++++++++---------
> 1 file changed, 32 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_migrate.c b/drivers/gpu/drm/xe/xe_migrate.c
> index 1bbc7bca33ed..921c9c1ea41f 100644
> --- a/drivers/gpu/drm/xe/xe_migrate.c
> +++ b/drivers/gpu/drm/xe/xe_migrate.c
> @@ -1920,6 +1920,25 @@ enum xe_migrate_copy_dir {
> #define XE_CACHELINE_BYTES 64ull
> #define XE_CACHELINE_MASK (XE_CACHELINE_BYTES - 1)
>
> +static u32 xe_migrate_copy_pitch(struct xe_device *xe, u32 len)
> +{
> + u32 pitch;
> +
> + if (IS_ALIGNED(len, PAGE_SIZE))
> + pitch = PAGE_SIZE;
> + else if (IS_ALIGNED(len, SZ_4K))
> + pitch = SZ_4K;
> + else if (IS_ALIGNED(len, SZ_256))
> + pitch = SZ_256;
> + else if (IS_ALIGNED(len, 4))
> + pitch = 4;
> + else
> + pitch = 1;
> +
> + xe_assert(xe, pitch > 1 || xe->info.has_mem_copy_instr);
> + return pitch;
> +}
> +
> static struct dma_fence *xe_migrate_vram(struct xe_migrate *m,
> unsigned long len,
> unsigned long sram_offset,
> @@ -1937,14 +1956,14 @@ static struct dma_fence *xe_migrate_vram(struct xe_migrate *m,
> struct xe_bb *bb;
> u32 update_idx, pt_slot = 0;
> unsigned long npages = DIV_ROUND_UP(len + sram_offset, PAGE_SIZE);
> - unsigned int pitch = len >= PAGE_SIZE && !(len & ~PAGE_MASK) ?
> - PAGE_SIZE : 4;
> + unsigned int pitch = xe_migrate_copy_pitch(xe, len);
> int err;
> unsigned long i, j;
> bool use_pde = xe_migrate_vram_use_pde(sram_addr, len + sram_offset);
>
> - if (drm_WARN_ON(&xe->drm, (!IS_ALIGNED(len, pitch)) ||
> - (sram_offset | vram_addr) & XE_CACHELINE_MASK))
> + if (!xe->info.has_mem_copy_instr &&
> + drm_WARN_ON(&xe->drm,
> + (!IS_ALIGNED(len, pitch)) || (sram_offset | vram_addr) & XE_CACHELINE_MASK))
> return ERR_PTR(-EOPNOTSUPP);
>
> xe_assert(xe, npages * PAGE_SIZE <= MAX_PREEMPTDISABLE_TRANSFER);
> @@ -2163,9 +2182,10 @@ int xe_migrate_access_memory(struct xe_migrate *m, struct xe_bo *bo,
> xe_bo_assert_held(bo);
>
> /* Use bounce buffer for small access and unaligned access */
> - if (!IS_ALIGNED(len, 4) ||
> - !IS_ALIGNED(page_offset, XE_CACHELINE_BYTES) ||
> - !IS_ALIGNED(offset, XE_CACHELINE_BYTES)) {
> + if (!xe->info.has_mem_copy_instr &&
> + (!IS_ALIGNED(len, 4) ||
> + !IS_ALIGNED(page_offset, XE_CACHELINE_BYTES) ||
> + !IS_ALIGNED(offset, XE_CACHELINE_BYTES))) {
> int buf_offset = 0;
> void *bounce;
> int err;
> @@ -2227,6 +2247,7 @@ int xe_migrate_access_memory(struct xe_migrate *m, struct xe_bo *bo,
> u64 vram_addr = vram_region_gpu_offset(bo->ttm.resource) +
> cursor.start;
> int current_bytes;
> + u32 pitch;
>
> if (cursor.size > MAX_PREEMPTDISABLE_TRANSFER)
> current_bytes = min_t(int, bytes_left,
> @@ -2234,13 +2255,13 @@ int xe_migrate_access_memory(struct xe_migrate *m, struct xe_bo *bo,
> else
> current_bytes = min_t(int, bytes_left, cursor.size);
>
> - if (current_bytes & ~PAGE_MASK) {
> - int pitch = 4;
> -
> + pitch = xe_migrate_copy_pitch(xe, current_bytes);
> + if (xe->info.has_mem_copy_instr)
> + current_bytes = min_t(int, current_bytes, U16_MAX * pitch);
> + else
> current_bytes = min_t(int, current_bytes,
> round_down(S16_MAX * pitch,
> XE_CACHELINE_BYTES));
> - }
>
> __fence = xe_migrate_vram(m, current_bytes,
> (unsigned long)buf & ~PAGE_MASK,
> --
> 2.51.0
>
next prev parent reply other threads:[~2025-10-22 18:00 UTC|newest]
Thread overview: 14+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-10-22 16:38 [PATCH v3 0/7] Some migration fixes/improvements Matthew Auld
2025-10-22 16:38 ` [PATCH v3 1/7] drm/xe/migrate: fix offset and len check Matthew Auld
2025-10-22 17:33 ` Matthew Brost
2025-10-22 16:38 ` [PATCH v3 2/7] drm/xe/migrate: rework size restrictions for sram pte emit Matthew Auld
2025-10-22 16:38 ` [PATCH v3 3/7] drm/xe/migrate: fix chunk handling for 2M page emit Matthew Auld
2025-10-22 16:38 ` [PATCH v3 4/7] drm/xe/migrate: fix batch buffer sizing Matthew Auld
2025-10-22 16:38 ` [PATCH v3 5/7] drm/xe/migrate: trim " Matthew Auld
2025-10-22 16:38 ` [PATCH v3 6/7] drm/xe/migrate: support MEM_COPY instruction Matthew Auld
2025-10-22 17:53 ` Matthew Brost
2025-10-22 16:38 ` [PATCH v3 7/7] drm/xe/migrate: skip bounce buffer path on xe2 Matthew Auld
2025-10-22 17:59 ` Matthew Brost [this message]
2025-10-23 0:16 ` ✓ CI.KUnit: success for Some migration fixes/improvements (rev3) Patchwork
2025-10-23 0:58 ` ✓ Xe.CI.BAT: " Patchwork
2025-10-23 7:31 ` ✗ Xe.CI.Full: failure " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aPkbnWE7pLRODu+i@lstrano-desk.jf.intel.com \
--to=matthew.brost@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=matthew.auld@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox