From: Matthew Brost <matthew.brost@intel.com>
To: "Summers, Stuart" <stuart.summers@intel.com>
Cc: "intel-xe@lists.freedesktop.org" <intel-xe@lists.freedesktop.org>
Subject: Re: [PATCH 7/7] drm/xe: Check for GuC responses on disabling scheduling
Date: Fri, 3 Oct 2025 12:49:29 -0700 [thread overview]
Message-ID: <aOAoyZGIpkEWXILt@lstrano-desk.jf.intel.com> (raw)
In-Reply-To: <0e4a67db333a000b722e114ec52b8c8000db32e2.camel@intel.com>
On Fri, Oct 03, 2025 at 01:42:26PM -0600, Summers, Stuart wrote:
> On Fri, 2025-10-03 at 12:38 -0700, Matthew Brost wrote:
> > On Fri, Oct 03, 2025 at 12:58:37PM -0600, Summers, Stuart wrote:
> > > On Fri, 2025-10-03 at 11:54 -0700, Matthew Brost wrote:
> > > > On Thu, Oct 02, 2025 at 11:04:44PM +0000, Stuart Summers wrote:
> > > > > In the event the GuC becomes unresponsive during a scheduling
> > > > > disable event, we still want the driver to be able to recover.
> > > > > This patch follows the same methodology we already have in
> > > > > place
> > > > > for TLB invalidation requests, where we send a request to GuC
> > > > > and wait for that invalidation done response. If the response
> > > > > doesn't come back in time we then at least print a message
> > > > > indicating the invalidation failed for some reason.
> > > > >
> > > > > In this case, we send the schedule disable and the expectation
> > > > > is that GuC will respond with a schedule done response. The KMD
> > > > > then catches that response and in turn sends a context
> > > > > deregistration
> > > > > response. So in the event GuC becomes unresponsive after we
> > > > > send
> > > > > the schedule disable, we actually have two g2h responses that
> > > > > have been reserved but never received.
> > > > >
> > > > > To handle this, make sure the pending disable event in the
> > > > > exec queue gets cleared (i.e. we received that response from
> > > > > GuC). If it doesn't in a reasonable amount of time, assume
> > > > > GuC is dead: ban the exec queue, queue up a GT reset, and
> > > > > manually call the schedule done handler. Then in the schedule
> > > > > done handler, in turn, check whether the context had been
> > > > > banned. If so, manually call the deregistration done handler
> > > > > to ensure all resources related to that exec queue get
> > > > > cleaned up properly. Without this, if the device becomes
> > > > > wedged after an exec queue has been created, the attached
> > > > > resources like the LRC will not get feed properly resulting
> > > > > in a memory leak.
> > > > >
> > > > > Signed-off-by: Stuart Summers <stuart.summers@intel.com>
> > > > > ---
> > > > > drivers/gpu/drm/xe/xe_guc_submit.c | 23
> > > > > ++++++++++++++++++++++-
> > > > > 1 file changed, 22 insertions(+), 1 deletion(-)
> > > > >
> > > > > diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c
> > > > > b/drivers/gpu/drm/xe/xe_guc_submit.c
> > > > > index 45b72bebfc63..a177d87c8524 100644
> > > > > --- a/drivers/gpu/drm/xe/xe_guc_submit.c
> > > > > +++ b/drivers/gpu/drm/xe/xe_guc_submit.c
> > > > > @@ -939,6 +939,9 @@ int xe_guc_read_stopped(struct xe_guc *guc)
> > > > > GUC_CONTEXT_##enable_disable,
> > > > >
> > > > > \
> > > > > }
> > > > >
> > > > > +static void handle_sched_done(struct xe_guc *guc, struct
> > > > > xe_exec_queue *q,
> > > > > + u32 runnable_state);
> > > > > +
> > > > > static void disable_scheduling_deregister(struct xe_guc *guc,
> > > > > struct xe_exec_queue
> > > > > *q)
> > > > > {
> > > > > @@ -974,6 +977,17 @@ static void
> > > > > disable_scheduling_deregister(struct xe_guc *guc,
> > > > > xe_guc_ct_send(&guc->ct, action, ARRAY_SIZE(action),
> > > > > G2H_LEN_DW_SCHED_CONTEXT_MODE_SET +
> > > > > G2H_LEN_DW_DEREGISTER_CONTEXT, 2);
> > > > > +
> > > > > + ret = wait_event_timeout(guc->ct.wq,
> > > > > + !exec_queue_pending_disable(q)
> > > > > ||
> > > > > + xe_guc_read_stopped(guc),
> > > > > + HZ * 5);
> > > >
> > > > This doesn't look right. Deregister is designed to be fully
> > > > async. If
> > > > this flow stops working for whatever reason the GuC is dead and
> > > > eventually somewhere in driver will detect this and trigger a GT
> > > > reset
> > > > which is cleanup all lost H2G.
> > > >
> > > > > + if (!ret || xe_guc_read_stopped(guc)) {
> > > > > + xe_gt_warn(guc_to_gt(guc), "Schedule disable
> > > > > failed
> > > > > to respond");
> > > > > + set_exec_queue_banned(q);
> > > > > + handle_sched_done(guc, q, 0);
> > > > > + xe_gt_reset_async(q->gt);
> > > > > + }
> > > > > }
> > > > >
> > > > > static void xe_guc_exec_queue_trigger_cleanup(struct
> > > > > xe_exec_queue
> > > > > *q)
> > > > > @@ -2117,6 +2131,8 @@ g2h_exec_queue_lookup(struct xe_guc *guc,
> > > > > u32
> > > > > guc_id)
> > > > > return q;
> > > > > }
> > > > >
> > > > > +static void handle_deregister_done(struct xe_guc *guc, struct
> > > > > xe_exec_queue *q);
> > > > > +
> > > > > static void deregister_exec_queue(struct xe_guc *guc, struct
> > > > > xe_exec_queue *q)
> > > > > {
> > > > > u32 action[] = {
> > > > > @@ -2131,7 +2147,12 @@ static void deregister_exec_queue(struct
> > > > > xe_guc *guc, struct xe_exec_queue *q)
> > > > >
> > > > > trace_xe_exec_queue_deregister(q);
> > > > >
> > > > > - xe_guc_ct_send_g2h_handler(&guc->ct, action,
> > > > > ARRAY_SIZE(action));
> > > > > + if (exec_queue_banned(q)) {
> > > > > + handle_deregister_done(guc, q);
> > > >
> > > > This would leave the GuC with reference to guc_id and subsequent
> > > > reuse
> > > > of the guc_id (i.e., next register) will fall.
> > >
> > > But again, in this case the GuC is dead and we should be getting
> > > that
> > > reset event you had mentioned above. The issue I'm having is
> >
> > Banned is a per thing queue and more than likely we wont be doing a
> > GT
> > reset, thus we still need remove references to the queue from the
> > GuC.
>
> Yeah this makes sense to me. My use of "banned" here was probably not
> ideal.
>
> >
> > > specifically around wedge events. Without a GT wedge, we will
> > > normally
> > > go through the GT reset flow and recover like you mentioned. But in
> > > the
> >
> > No. See above.
> >
> > > case of a wedge, we don't redo the software part of the reset (i.e.
> > > we
> > > don't reset contexts, etc) per gt_reset():
> > > static int gt_reset(struct xe_gt *gt)
> > > {
> > > unsigned int fw_ref;
> > > int err;
> > >
> > > if (xe_device_wedged(gt_to_xe(gt)))
> > > return -ECANCELED;
> > >
> > > Maybe instead of banned I can check for banned and wedged here? Or
> > > maybe we should rethink the software reset flow in the event of a
> > > wedge?
> >
> > The idea with wedged is we leave all hardward state, including the
> > GuC,
> > intacted for inspection. So I think a xe_device_wedged checked here
> > makes sense. This would cover the case where start a queue teardown
> > via
> > a CLEANUP message and mid-flow we wedge the device.
>
> But hardware state doesn't mean software state. Are you saying when the
> device is wedged we want the memory to all be intact as well? And how
> do we determine when that gets freed? On unbind?
When we wedge a device we take a reference to all queues which perserves
the software queue state and also prevents any queue cleanup flows
regardless of what a user application does. On driver unbind after wedge
we drop wedge reference which should free all memory in the KMD.
Matt
>
> Thanks,
> Stuart
>
> >
> > Matt
> >
> > >
> > > Thanks,
> > > Stuart
> > >
> > > >
> > > > Matt
> > > >
> > > > > + } else {
> > > > > + xe_guc_ct_send_g2h_handler(&guc->ct, action,
> > > > > + ARRAY_SIZE(action));
> > > > > + }
> > > > > }
> > > > >
> > > > > static void handle_sched_done(struct xe_guc *guc, struct
> > > > > xe_exec_queue *q,
> > > > > --
> > > > > 2.34.1
> > > > >
> > >
>
next prev parent reply other threads:[~2025-10-03 19:49 UTC|newest]
Thread overview: 28+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-10-02 23:04 [PATCH 0/7] Fix a couple of wedge corner-case memory leaks Stuart Summers
2025-10-02 23:04 ` [PATCH 1/7] drm/xe: Add additional trace points for LRCs Stuart Summers
2025-10-02 23:04 ` [PATCH 2/7] drm/xe: Add a trace point for VM close Stuart Summers
2025-10-02 23:04 ` [PATCH 3/7] drm/xe: Add the BO pointer info to the BO trace Stuart Summers
2025-10-02 23:04 ` [PATCH 4/7] drm/xe: Add new exec queue trace points Stuart Summers
2025-10-02 23:04 ` [PATCH 5/7] drm/xe: Handle missing migration VM on VM creation Stuart Summers
2025-10-02 23:34 ` Lin, Shuicheng
2025-10-03 6:56 ` Matthew Brost
2025-10-03 14:33 ` Summers, Stuart
2025-10-02 23:04 ` [PATCH 6/7] drm/xe: Don't send a CLEANUP message on sched pause Stuart Summers
2025-10-03 18:50 ` Matthew Brost
2025-10-03 18:53 ` Summers, Stuart
2025-10-02 23:04 ` [PATCH 7/7] drm/xe: Check for GuC responses on disabling scheduling Stuart Summers
2025-10-03 18:54 ` Matthew Brost
2025-10-03 18:58 ` Summers, Stuart
2025-10-03 19:38 ` Matthew Brost
2025-10-03 19:42 ` Summers, Stuart
2025-10-03 19:49 ` Matthew Brost [this message]
2025-10-03 19:53 ` Summers, Stuart
2025-10-02 23:11 ` ✗ CI.checkpatch: warning for Fix a couple of wedge corner-case memory leaks Patchwork
2025-10-02 23:12 ` ✓ CI.KUnit: success " Patchwork
2025-10-02 23:58 ` ✗ Xe.CI.BAT: failure " Patchwork
2025-10-03 2:16 ` ✗ Xe.CI.Full: " Patchwork
2025-10-03 14:38 ` Summers, Stuart
-- strict thread matches above, loose matches on Subject: below --
2025-10-13 16:24 [PATCH 0/7] " Stuart Summers
2025-10-13 16:25 ` [PATCH 7/7] drm/xe: Check for GuC responses on disabling scheduling Stuart Summers
2025-10-13 22:31 [PATCH 0/7] Fix a couple of wedge corner-case memory leaks Stuart Summers
2025-10-13 22:31 ` [PATCH 7/7] drm/xe: Check for GuC responses on disabling scheduling Stuart Summers
2025-10-14 2:09 ` Matthew Brost
2025-10-14 3:10 ` Summers, Stuart
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aOAoyZGIpkEWXILt@lstrano-desk.jf.intel.com \
--to=matthew.brost@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=stuart.summers@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox