From: Matthew Brost <matthew.brost@intel.com>
To: Brian Nguyen <brian3.nguyen@intel.com>
Cc: <intel-xe@lists.freedesktop.org>, <tejas.upadhyay@intel.com>,
<shuicheng.lin@intel.com>, <stuart.summers@intel.com>
Subject: Re: [PATCH v2 09/11] drm/xe: Append page reclamation action to tlb inval
Date: Wed, 3 Dec 2025 15:15:25 -0800 [thread overview]
Message-ID: <aTDEjdp0P1CYUBCM@lstrano-desk.jf.intel.com> (raw)
In-Reply-To: <20251126230201.3782788-22-brian3.nguyen@intel.com>
On Thu, Nov 27, 2025 at 07:02:10AM +0800, Brian Nguyen wrote:
> Add page reclamation action to tlb inval backend. The page reclamation
> action is paired with range tlb invalidations so both are issued at the
> same time.
>
> Page reclamation will issue the TLB invalidation with an invalid seqno
> and a H2G page reclamation action with the fence's corresponding seqno
> and handle the fence accordingly on page reclaim action done handler.
>
> If page reclamation fails, tlb timeout handler will be responsible for
> signalling fence and cleaning up.
>
> v2:
> - add send_page_reclaim to patch.
> - Remove flush_cache and use prl_sa pointer to determine PPC flush
> instead of explicit bool. Add NULL as fallback for others. (Matthew B)
>
> Signed-off-by: Brian Nguyen <brian3.nguyen@intel.com>
> Suggested-by: Matthew Brost <matthew.brost@intel.com>
Reviewed-by: Matthew Brost <matthew.brost@intel.com>
> ---
> drivers/gpu/drm/xe/xe_guc_tlb_inval.c | 29 ++++++++++++++++++++-----
> drivers/gpu/drm/xe/xe_tlb_inval.c | 7 +++---
> drivers/gpu/drm/xe/xe_tlb_inval.h | 2 +-
> drivers/gpu/drm/xe/xe_tlb_inval_job.c | 2 +-
> drivers/gpu/drm/xe/xe_tlb_inval_types.h | 4 +++-
> drivers/gpu/drm/xe/xe_vm.c | 4 ++--
> 6 files changed, 35 insertions(+), 13 deletions(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_guc_tlb_inval.c b/drivers/gpu/drm/xe/xe_guc_tlb_inval.c
> index 37ac943cb10f..ffea9c0c5cd0 100644
> --- a/drivers/gpu/drm/xe/xe_guc_tlb_inval.c
> +++ b/drivers/gpu/drm/xe/xe_guc_tlb_inval.c
> @@ -13,6 +13,7 @@
> #include "xe_guc_tlb_inval.h"
> #include "xe_force_wake.h"
> #include "xe_mmio.h"
> +#include "xe_sa.h"
> #include "xe_tlb_inval.h"
>
> #include "regs/xe_guc_regs.h"
> @@ -93,6 +94,20 @@ static int send_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval, u32 seqno)
> return -ECANCELED;
> }
>
> +static int send_page_reclaim(struct xe_guc *guc, u32 seqno,
> + u64 gpu_addr)
> +{
> + u32 action[] = {
> + XE_GUC_ACTION_PAGE_RECLAMATION,
> + seqno,
> + lower_32_bits(gpu_addr),
> + upper_32_bits(gpu_addr),
> + };
> +
> + return xe_guc_ct_send(&guc->ct, action, ARRAY_SIZE(action),
> + G2H_LEN_DW_PAGE_RECLAMATION, 1);
> +}
> +
> /*
> * Ensure that roundup_pow_of_two(length) doesn't overflow.
> * Note that roundup_pow_of_two() operates on unsigned long,
> @@ -101,20 +116,21 @@ static int send_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval, u32 seqno)
> #define MAX_RANGE_TLB_INVALIDATION_LENGTH (rounddown_pow_of_two(ULONG_MAX))
>
> static int send_tlb_inval_ppgtt(struct xe_tlb_inval *tlb_inval, u32 seqno,
> - u64 start, u64 end, u32 asid)
> + u64 start, u64 end, u32 asid,
> + struct drm_suballoc *prl_sa)
> {
> #define MAX_TLB_INVALIDATION_LEN 7
> struct xe_guc *guc = tlb_inval->private;
> struct xe_gt *gt = guc_to_gt(guc);
> u32 action[MAX_TLB_INVALIDATION_LEN];
> u64 length = end - start;
> - int len = 0;
> + int len = 0, err;
>
> if (guc_to_xe(guc)->info.force_execlist)
> return -ECANCELED;
>
> action[len++] = XE_GUC_ACTION_TLB_INVALIDATION;
> - action[len++] = seqno;
> + action[len++] = !prl_sa ? seqno : TLB_INVALIDATION_SEQNO_INVALID;
> if (!gt_to_xe(gt)->info.has_range_tlb_inval ||
> length > MAX_RANGE_TLB_INVALIDATION_LENGTH) {
> action[len++] = MAKE_INVAL_OP(XE_GUC_TLB_INVAL_FULL);
> @@ -155,7 +171,7 @@ static int send_tlb_inval_ppgtt(struct xe_tlb_inval *tlb_inval, u32 seqno,
> ilog2(SZ_2M) + 1)));
> xe_gt_assert(gt, IS_ALIGNED(start, length));
>
> - action[len++] = MAKE_INVAL_OP_FLUSH(XE_GUC_TLB_INVAL_PAGE_SELECTIVE, true);
> + action[len++] = MAKE_INVAL_OP_FLUSH(XE_GUC_TLB_INVAL_PAGE_SELECTIVE, !prl_sa);
> action[len++] = asid;
> action[len++] = lower_32_bits(start);
> action[len++] = upper_32_bits(start);
> @@ -164,7 +180,10 @@ static int send_tlb_inval_ppgtt(struct xe_tlb_inval *tlb_inval, u32 seqno,
>
> xe_gt_assert(gt, len <= MAX_TLB_INVALIDATION_LEN);
>
> - return send_tlb_inval(guc, action, len);
> + err = send_tlb_inval(guc, action, len);
> + if (!err && prl_sa)
> + err = send_page_reclaim(guc, seqno, xe_sa_bo_gpu_addr(prl_sa));
> + return err;
> }
>
> static bool tlb_inval_initialized(struct xe_tlb_inval *tlb_inval)
> diff --git a/drivers/gpu/drm/xe/xe_tlb_inval.c b/drivers/gpu/drm/xe/xe_tlb_inval.c
> index a122fbb9fc4a..dec042248164 100644
> --- a/drivers/gpu/drm/xe/xe_tlb_inval.c
> +++ b/drivers/gpu/drm/xe/xe_tlb_inval.c
> @@ -313,6 +313,7 @@ int xe_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval)
> * @start: start address
> * @end: end address
> * @asid: address space id
> + * @prl_sa: suballocation of page reclaim list if used, NULL indicates PPC flush
> *
> * Issue a range based TLB invalidation if supported, if not fallback to a full
> * TLB invalidation. Completion of TLB is asynchronous and caller can use
> @@ -322,10 +323,10 @@ int xe_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval)
> */
> int xe_tlb_inval_range(struct xe_tlb_inval *tlb_inval,
> struct xe_tlb_inval_fence *fence, u64 start, u64 end,
> - u32 asid)
> + u32 asid, struct drm_suballoc *prl_sa)
> {
> return xe_tlb_inval_issue(tlb_inval, fence, tlb_inval->ops->ppgtt,
> - start, end, asid);
> + start, end, asid, prl_sa);
> }
>
> /**
> @@ -341,7 +342,7 @@ void xe_tlb_inval_vm(struct xe_tlb_inval *tlb_inval, struct xe_vm *vm)
> u64 range = 1ull << vm->xe->info.va_bits;
>
> xe_tlb_inval_fence_init(tlb_inval, &fence, true);
> - xe_tlb_inval_range(tlb_inval, &fence, 0, range, vm->usm.asid);
> + xe_tlb_inval_range(tlb_inval, &fence, 0, range, vm->usm.asid, NULL);
> xe_tlb_inval_fence_wait(&fence);
> }
>
> diff --git a/drivers/gpu/drm/xe/xe_tlb_inval.h b/drivers/gpu/drm/xe/xe_tlb_inval.h
> index 05614915463a..858d0690f995 100644
> --- a/drivers/gpu/drm/xe/xe_tlb_inval.h
> +++ b/drivers/gpu/drm/xe/xe_tlb_inval.h
> @@ -23,7 +23,7 @@ int xe_tlb_inval_ggtt(struct xe_tlb_inval *tlb_inval);
> void xe_tlb_inval_vm(struct xe_tlb_inval *tlb_inval, struct xe_vm *vm);
> int xe_tlb_inval_range(struct xe_tlb_inval *tlb_inval,
> struct xe_tlb_inval_fence *fence,
> - u64 start, u64 end, u32 asid);
> + u64 start, u64 end, u32 asid, struct drm_suballoc *prl_sa);
>
> void xe_tlb_inval_fence_init(struct xe_tlb_inval *tlb_inval,
> struct xe_tlb_inval_fence *fence,
> diff --git a/drivers/gpu/drm/xe/xe_tlb_inval_job.c b/drivers/gpu/drm/xe/xe_tlb_inval_job.c
> index 2185f42b9644..b59e322e499d 100644
> --- a/drivers/gpu/drm/xe/xe_tlb_inval_job.c
> +++ b/drivers/gpu/drm/xe/xe_tlb_inval_job.c
> @@ -60,7 +60,7 @@ static struct dma_fence *xe_tlb_inval_job_run(struct xe_dep_job *dep_job)
> }
>
> xe_tlb_inval_range(job->tlb_inval, ifence, job->start,
> - job->end, job->vm->usm.asid);
> + job->end, job->vm->usm.asid, prl_sa);
>
> return job->fence;
> }
> diff --git a/drivers/gpu/drm/xe/xe_tlb_inval_types.h b/drivers/gpu/drm/xe/xe_tlb_inval_types.h
> index 7a6967ce3b76..48d1503e8460 100644
> --- a/drivers/gpu/drm/xe/xe_tlb_inval_types.h
> +++ b/drivers/gpu/drm/xe/xe_tlb_inval_types.h
> @@ -9,6 +9,7 @@
> #include <linux/workqueue.h>
> #include <linux/dma-fence.h>
>
> +struct drm_suballoc;
> struct xe_tlb_inval;
>
> /** struct xe_tlb_inval_ops - TLB invalidation ops (backend) */
> @@ -40,12 +41,13 @@ struct xe_tlb_inval_ops {
> * @start: Start address
> * @end: End address
> * @asid: Address space ID
> + * @prl_sa: Suballocation for page reclaim list
> *
> * Return 0 on success, -ECANCELED if backend is mid-reset, error on
> * failure
> */
> int (*ppgtt)(struct xe_tlb_inval *tlb_inval, u32 seqno, u64 start,
> - u64 end, u32 asid);
> + u64 end, u32 asid, struct drm_suballoc *prl_sa);
>
> /**
> * @initialized: Backend is initialized
> diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> index 8ab726289583..fc7fc8243326 100644
> --- a/drivers/gpu/drm/xe/xe_vm.c
> +++ b/drivers/gpu/drm/xe/xe_vm.c
> @@ -3924,7 +3924,7 @@ int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
>
> err = xe_tlb_inval_range(&tile->primary_gt->tlb_inval,
> &fence[fence_id], start, end,
> - vm->usm.asid);
> + vm->usm.asid, NULL);
> if (err)
> goto wait;
> ++fence_id;
> @@ -3937,7 +3937,7 @@ int xe_vm_range_tilemask_tlb_inval(struct xe_vm *vm, u64 start,
>
> err = xe_tlb_inval_range(&tile->media_gt->tlb_inval,
> &fence[fence_id], start, end,
> - vm->usm.asid);
> + vm->usm.asid, NULL);
> if (err)
> goto wait;
> ++fence_id;
> --
> 2.52.0
>
next prev parent reply other threads:[~2025-12-03 23:15 UTC|newest]
Thread overview: 38+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-26 23:02 [PATCH v2 00/11] Page Reclamation Support for Xe3p Platforms Brian Nguyen
2025-11-26 23:02 ` [PATCH v2 01/11] [DO, NOT, REVIEW] drm/xe: Do not forward invalid TLB invalidation seqnos to upper layers Brian Nguyen
2025-11-26 23:02 ` [PATCH v2 02/11] drm/xe: Reset tlb fence timeout on invalid seqno received Brian Nguyen
2025-12-02 22:24 ` Matthew Brost
2025-11-26 23:02 ` [PATCH v2 03/11] drm/xe/xe_tlb_inval: Modify fence interface to support PPC flush Brian Nguyen
2025-12-02 22:18 ` Matthew Brost
2025-11-26 23:02 ` [PATCH v2 04/11] drm/xe: Add page reclamation info to device info Brian Nguyen
2025-11-26 23:02 ` [PATCH v2 05/11] drm/xe/guc: Add page reclamation interface to GuC Brian Nguyen
2025-12-02 22:21 ` Matthew Brost
2025-12-03 0:17 ` Lin, Shuicheng
2025-11-26 23:02 ` [PATCH v2 06/11] drm/xe: Create page reclaim list on unbind Brian Nguyen
2025-12-01 21:45 ` Nguyen, Brian3
2025-12-03 22:56 ` Matthew Brost
2025-12-04 0:19 ` Nguyen, Brian3
2025-11-26 23:02 ` [PATCH v2 07/11] drm/xe: Suballocate BO for page reclaim Brian Nguyen
2025-12-03 23:06 ` Matthew Brost
2025-11-26 23:02 ` [PATCH v2 08/11] drm/xe: Prep page reclaim in tlb inval job Brian Nguyen
2025-12-03 23:13 ` Matthew Brost
2025-12-04 0:22 ` Nguyen, Brian3
2025-12-04 1:20 ` Matthew Brost
2025-12-04 5:42 ` Nguyen, Brian3
2025-12-04 18:05 ` Matthew Brost
2025-12-04 20:02 ` Nguyen, Brian3
2025-12-09 5:57 ` Upadhyay, Tejas
2025-12-09 6:14 ` Nguyen, Brian3
2025-11-26 23:02 ` [PATCH v2 09/11] drm/xe: Append page reclamation action to tlb inval Brian Nguyen
2025-12-03 23:15 ` Matthew Brost [this message]
2025-11-26 23:02 ` [PATCH v2 10/11] drm/xe: Optimize flushing of L2$ by skipping unnecessary page reclaim Brian Nguyen
2025-12-09 14:23 ` Upadhyay, Tejas
2025-11-26 23:02 ` [PATCH v2 11/11] drm/xe: Add debugfs support for page reclamation Brian Nguyen
2025-12-02 22:28 ` Matthew Brost
2025-12-02 22:51 ` Nguyen, Brian3
2025-12-02 22:59 ` Matthew Brost
2025-12-05 18:02 ` Lin, Shuicheng
2025-11-26 23:52 ` ✗ CI.checkpatch: warning for Page Reclamation Support for Xe3p Platforms (rev2) Patchwork
2025-11-26 23:54 ` ✓ CI.KUnit: success " Patchwork
2025-11-27 0:54 ` ✗ Xe.CI.BAT: failure " Patchwork
2025-11-27 1:42 ` ✗ Xe.CI.Full: " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aTDEjdp0P1CYUBCM@lstrano-desk.jf.intel.com \
--to=matthew.brost@intel.com \
--cc=brian3.nguyen@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=shuicheng.lin@intel.com \
--cc=stuart.summers@intel.com \
--cc=tejas.upadhyay@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox