Intel-XE Archive on lore.kernel.org
 help / color / mirror / Atom feed
From: Matthew Brost <matthew.brost@intel.com>
To: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>
Cc: <intel-xe@lists.freedesktop.org>, <lucas.demarchi@intel.com>,
	<stuart.summers@intel.com>
Subject: Re: [PATCH] drm/xe: Separate out sched/deregister_done handling
Date: Tue, 19 Mar 2024 23:54:22 +0000	[thread overview]
Message-ID: <ZfolrobherEHBYb9@DUT025-TGLU.fm.intel.com> (raw)
In-Reply-To: <20240319184153.16667-1-niranjana.vishwanathapura@intel.com>

On Tue, Mar 19, 2024 at 11:41:53AM -0700, Niranjana Vishwanathapura wrote:
> Abstract out the core part of sched_done and deregister_done handlers
> to separate functions to decouple them from any protocol error handling
> part and make them more readable.
> 
> Signed-off-by: Niranjana Vishwanathapura <niranjana.vishwanathapura@intel.com>

Good clean up.
Reviewed-by: Matthew Brost <matthew.brost@intel.com>

> ---
>  drivers/gpu/drm/xe/xe_guc_submit.c | 64 +++++++++++++++++-------------
>  1 file changed, 37 insertions(+), 27 deletions(-)
> 
> diff --git a/drivers/gpu/drm/xe/xe_guc_submit.c b/drivers/gpu/drm/xe/xe_guc_submit.c
> index 82c955a2a15c..4a2b8e6b81b8 100644
> --- a/drivers/gpu/drm/xe/xe_guc_submit.c
> +++ b/drivers/gpu/drm/xe/xe_guc_submit.c
> @@ -1574,28 +1574,8 @@ static void deregister_exec_queue(struct xe_guc *guc, struct xe_exec_queue *q)
>  	xe_guc_ct_send_g2h_handler(&guc->ct, action, ARRAY_SIZE(action));
>  }
>  
> -int xe_guc_sched_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
> +static void handle_sched_done(struct xe_guc *guc, struct xe_exec_queue *q)
>  {
> -	struct xe_device *xe = guc_to_xe(guc);
> -	struct xe_exec_queue *q;
> -	u32 guc_id = msg[0];
> -
> -	if (unlikely(len < 2)) {
> -		drm_err(&xe->drm, "Invalid length %u", len);
> -		return -EPROTO;
> -	}
> -
> -	q = g2h_exec_queue_lookup(guc, guc_id);
> -	if (unlikely(!q))
> -		return -EPROTO;
> -
> -	if (unlikely(!exec_queue_pending_enable(q) &&
> -		     !exec_queue_pending_disable(q))) {
> -		drm_err(&xe->drm, "Unexpected engine state 0x%04x",
> -			atomic_read(&q->guc->state));
> -		return -EPROTO;
> -	}
> -
>  	trace_xe_exec_queue_scheduling_done(q);
>  
>  	if (exec_queue_pending_enable(q)) {
> @@ -1615,17 +1595,15 @@ int xe_guc_sched_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
>  			deregister_exec_queue(guc, q);
>  		}
>  	}
> -
> -	return 0;
>  }
>  
> -int xe_guc_deregister_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
> +int xe_guc_sched_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
>  {
>  	struct xe_device *xe = guc_to_xe(guc);
>  	struct xe_exec_queue *q;
>  	u32 guc_id = msg[0];
>  
> -	if (unlikely(len < 1)) {
> +	if (unlikely(len < 2)) {
>  		drm_err(&xe->drm, "Invalid length %u", len);
>  		return -EPROTO;
>  	}
> @@ -1634,13 +1612,20 @@ int xe_guc_deregister_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
>  	if (unlikely(!q))
>  		return -EPROTO;
>  
> -	if (!exec_queue_destroyed(q) || exec_queue_pending_disable(q) ||
> -	    exec_queue_pending_enable(q) || exec_queue_enabled(q)) {
> +	if (unlikely(!exec_queue_pending_enable(q) &&
> +		     !exec_queue_pending_disable(q))) {
>  		drm_err(&xe->drm, "Unexpected engine state 0x%04x",
>  			atomic_read(&q->guc->state));
>  		return -EPROTO;
>  	}
>  
> +	handle_sched_done(guc, q);
> +
> +	return 0;
> +}
> +
> +static void handle_deregister_done(struct xe_guc *guc, struct xe_exec_queue *q)
> +{
>  	trace_xe_exec_queue_deregister_done(q);
>  
>  	clear_exec_queue_registered(q);
> @@ -1649,6 +1634,31 @@ int xe_guc_deregister_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
>  		xe_exec_queue_put(q);
>  	else
>  		__guc_exec_queue_fini(guc, q);
> +}
> +
> +int xe_guc_deregister_done_handler(struct xe_guc *guc, u32 *msg, u32 len)
> +{
> +	struct xe_device *xe = guc_to_xe(guc);
> +	struct xe_exec_queue *q;
> +	u32 guc_id = msg[0];
> +
> +	if (unlikely(len < 1)) {
> +		drm_err(&xe->drm, "Invalid length %u", len);
> +		return -EPROTO;
> +	}
> +
> +	q = g2h_exec_queue_lookup(guc, guc_id);
> +	if (unlikely(!q))
> +		return -EPROTO;
> +
> +	if (!exec_queue_destroyed(q) || exec_queue_pending_disable(q) ||
> +	    exec_queue_pending_enable(q) || exec_queue_enabled(q)) {
> +		drm_err(&xe->drm, "Unexpected engine state 0x%04x",
> +			atomic_read(&q->guc->state));
> +		return -EPROTO;
> +	}
> +
> +	handle_deregister_done(guc, q);
>  
>  	return 0;
>  }
> -- 
> 2.43.0
> 

      parent reply	other threads:[~2024-03-19 23:56 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-03-19 18:41 [PATCH] drm/xe: Separate out sched/deregister_done handling Niranjana Vishwanathapura
2024-03-19 21:02 ` ✓ CI.Patch_applied: success for " Patchwork
2024-03-19 21:02 ` ✓ CI.checkpatch: " Patchwork
2024-03-19 21:03 ` ✓ CI.KUnit: " Patchwork
2024-03-19 21:14 ` ✓ CI.Build: " Patchwork
2024-03-19 21:16 ` ✓ CI.Hooks: " Patchwork
2024-03-19 21:17 ` ✓ CI.checksparse: " Patchwork
2024-03-19 21:39 ` ✓ CI.BAT: " Patchwork
2024-03-19 23:54 ` Matthew Brost [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=ZfolrobherEHBYb9@DUT025-TGLU.fm.intel.com \
    --to=matthew.brost@intel.com \
    --cc=intel-xe@lists.freedesktop.org \
    --cc=lucas.demarchi@intel.com \
    --cc=niranjana.vishwanathapura@intel.com \
    --cc=stuart.summers@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox