From: Maciej Patelczyk <maciej.patelczyk@intel.com>
To: Matthew Brost <matthew.brost@intel.com>,
<intel-xe@lists.freedesktop.org>
Cc: <stuart.summers@intel.com>, <arvind.yadav@intel.com>,
<himal.prasad.ghimiray@intel.com>,
<thomas.hellstrom@linux.intel.com>, <francois.dugast@intel.com>
Subject: Re: [PATCH v4 12/12] drm/xe: Track parallel page fault activity in GT stats
Date: Thu, 7 May 2026 15:56:37 +0200 [thread overview]
Message-ID: <400d55b6-652b-4918-a5a1-1180dc42b13b@intel.com> (raw)
In-Reply-To: <20260226042834.2963245-13-matthew.brost@intel.com>
On 26/02/2026 05:28, Matthew Brost wrote:
> Add a new GT statistic, PARALLEL_PAGEFAULT_COUNT, to record when
> multiple page fault workers are active concurrently.
>
> When a worker dequeues a fault, scan peer workers for an active
> cache entry and increment the counter if another fault is already
> in flight. This provides basic visibility into parallel fault
> handling behavior for performance analysis and tuning.
>
> Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> ---
> drivers/gpu/drm/xe/xe_gt_stats.c | 1 +
> drivers/gpu/drm/xe/xe_gt_stats_types.h | 1 +
> drivers/gpu/drm/xe/xe_pagefault.c | 18 +++++++++++++++++-
> 3 files changed, 19 insertions(+), 1 deletion(-)
>
> diff --git a/drivers/gpu/drm/xe/xe_gt_stats.c b/drivers/gpu/drm/xe/xe_gt_stats.c
> index cdd467dfb46d..621d1a2df067 100644
> --- a/drivers/gpu/drm/xe/xe_gt_stats.c
> +++ b/drivers/gpu/drm/xe/xe_gt_stats.c
> @@ -58,6 +58,7 @@ static const char *const stat_description[__XE_GT_STATS_NUM_IDS] = {
> DEF_STAT_STR(CHAIN_IRQ_PAGEFAULT_COUNT, "chain_irq_pagefault_count"),
> DEF_STAT_STR(CHAIN_DRAIN_IRQ_PAGEFAULT_COUNT, "chain_drain_irq_pagefault_count"),
> DEF_STAT_STR(CHAIN_MISMATCH_PAGEFAULT_COUNT, "chain_mismatch_pagefault_count"),
> + DEF_STAT_STR(PARALLEL_PAGEFAULT_COUNT, "parallel_pagefault_count"),
> DEF_STAT_STR(LAST_PAGEFAULT_COUNT, "last_pagefault_count"),
> DEF_STAT_STR(SVM_PAGEFAULT_COUNT, "svm_pagefault_count"),
> DEF_STAT_STR(TLB_INVAL, "tlb_inval_count"),
> diff --git a/drivers/gpu/drm/xe/xe_gt_stats_types.h b/drivers/gpu/drm/xe/xe_gt_stats_types.h
> index 591e614e1cfc..075a12152ae2 100644
> --- a/drivers/gpu/drm/xe/xe_gt_stats_types.h
> +++ b/drivers/gpu/drm/xe/xe_gt_stats_types.h
> @@ -13,6 +13,7 @@ enum xe_gt_stats_id {
> XE_GT_STATS_ID_CHAIN_IRQ_PAGEFAULT_COUNT,
> XE_GT_STATS_ID_CHAIN_DRAIN_IRQ_PAGEFAULT_COUNT,
> XE_GT_STATS_ID_CHAIN_MISMATCH_PAGEFAULT_COUNT,
> + XE_GT_STATS_ID_PARALLEL_PAGEFAULT_COUNT,
> XE_GT_STATS_ID_LAST_PAGEFAULT_COUNT,
> XE_GT_STATS_ID_SVM_PAGEFAULT_COUNT,
> XE_GT_STATS_ID_TLB_INVAL,
> diff --git a/drivers/gpu/drm/xe/xe_pagefault.c b/drivers/gpu/drm/xe/xe_pagefault.c
> index d252a8c9d88c..2a37a4c97aad 100644
> --- a/drivers/gpu/drm/xe/xe_pagefault.c
> +++ b/drivers/gpu/drm/xe/xe_pagefault.c
> @@ -458,9 +458,10 @@ static bool xe_pagefault_queue_pop(struct xe_pagefault_queue *pf_queue,
> {
> struct xe_device *xe = container_of(pf_queue, typeof(*xe),
> usm.pf_queue);
> - struct xe_pagefault_work *pf_work;
> + struct xe_pagefault_work *pf_work, *__pf_work;
> struct xe_pagefault *lpf;
> size_t align = SZ_2M;
> + int i;
>
> guard(spinlock_irq)(&pf_queue->lock);
>
> @@ -497,6 +498,21 @@ static bool xe_pagefault_queue_pop(struct xe_pagefault_queue *pf_queue,
> pf_work->cache.pf = lpf;
> lpf->consumer.alloc_state = XE_PAGEFAULT_ALLOC_STATE_ACTIVE;
>
> + for (i = 0, __pf_work = xe->usm.pf_workers;
> + i < xe->info.num_pf_work; ++i, ++__pf_work) {
> + u64 cache_start = __pf_work->cache.start;
> +
> + if (__pf_work == pf_work)
> + continue;
> +
> + if (cache_start != XE_PAGEFAULT_CACHE_START_INVALID) {
> + xe_gt_stats_incr(xe_root_mmio_gt(xe),
> + XE_GT_STATS_ID_PARALLEL_PAGEFAULT_COUNT,
> + 1);
In the xe_pagefault_queue_work() right after first ACTIVE pagefault is
ACK'ed the worke's cache is invalidated.
Worker may still run releasing chained entries but this stat won't catch it.
Maybe it doesn't matter. Maybe the work is done (the
xe_pagefault_service() being the core work) at that time.
Just noticing.
If it doesn't matter then
Reviewed-by: Maciej Patelczyk <maciej.patelczyk@intel.com>
> + break;
> + }
> + }
> +
> /* Drain queue until empty or new fault found */
> while (1) {
> if (pf_queue->tail == pf_queue->head)
next prev parent reply other threads:[~2026-05-07 13:56 UTC|newest]
Thread overview: 33+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-02-26 4:28 [PATCH v4 00/12] Fine grained fault locking, threaded prefetch, storm cache Matthew Brost
2026-02-26 4:28 ` [PATCH v4 01/12] drm/xe: Fine grained page fault locking Matthew Brost
2026-02-26 4:28 ` [PATCH v4 02/12] drm/xe: Allow prefetch-only VM bind IOCTLs to use VM read lock Matthew Brost
2026-02-26 4:28 ` [PATCH v4 03/12] drm/xe: Thread prefetch of SVM ranges Matthew Brost
2026-02-26 4:28 ` [PATCH v4 04/12] drm/xe: Use a single page-fault queue with multiple workers Matthew Brost
2026-05-06 15:46 ` Maciej Patelczyk
2026-05-06 19:42 ` Matthew Brost
2026-05-07 12:41 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 05/12] drm/xe: Add num_pf_work modparam Matthew Brost
2026-05-06 15:59 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 06/12] drm/xe: Engine class and instance into a u8 Matthew Brost
2026-05-06 16:04 ` Maciej Patelczyk
2026-05-07 16:20 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 07/12] drm/xe: Track pagefault worker runtime Matthew Brost
2026-05-07 12:51 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 08/12] drm/xe: Chain page faults via queue-resident cache to avoid fault storms Matthew Brost
2026-05-08 12:03 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 09/12] drm/xe: Add pagefault chaining stats Matthew Brost
2026-05-07 13:15 ` Maciej Patelczyk
2026-05-07 13:52 ` Francois Dugast
2026-02-26 4:28 ` [PATCH v4 10/12] drm/xe: Add debugfs pagefault_info Matthew Brost
2026-05-07 10:07 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 11/12] drm/xe: batch CT pagefault acks with periodic flush Matthew Brost
2026-05-08 9:24 ` Maciej Patelczyk
2026-02-26 4:28 ` [PATCH v4 12/12] drm/xe: Track parallel page fault activity in GT stats Matthew Brost
2026-05-07 13:56 ` Maciej Patelczyk [this message]
2026-05-07 14:23 ` Francois Dugast
2026-02-26 4:35 ` ✗ CI.checkpatch: warning for Fine grained fault locking, threaded prefetch, storm cache (rev4) Patchwork
2026-02-26 4:36 ` ✓ CI.KUnit: success " Patchwork
2026-02-26 5:26 ` ✗ Xe.CI.BAT: failure " Patchwork
2026-02-26 8:59 ` ✗ Xe.CI.FULL: " Patchwork
2026-02-26 13:43 ` [PATCH v4 00/12] Fine grained fault locking, threaded prefetch, storm cache Thomas Hellström
2026-02-26 19:36 ` Matthew Brost
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=400d55b6-652b-4918-a5a1-1180dc42b13b@intel.com \
--to=maciej.patelczyk@intel.com \
--cc=arvind.yadav@intel.com \
--cc=francois.dugast@intel.com \
--cc=himal.prasad.ghimiray@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=matthew.brost@intel.com \
--cc=stuart.summers@intel.com \
--cc=thomas.hellstrom@linux.intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.