Intel-XE Archive on lore.kernel.org
 help / color / mirror / Atom feed
From: "Belgaumkar, Vinay" <vinay.belgaumkar@intel.com>
To: "Nilawar, Badal" <badal.nilawar@intel.com>,
	<intel-xe@lists.freedesktop.org>
Subject: Re: [PATCH 2/2] drm/xe: Add forcewake status to powergate_info
Date: Mon, 2 Feb 2026 09:12:09 -0800	[thread overview]
Message-ID: <e40bb59b-d40e-46a1-9792-93019db06221@intel.com> (raw)
In-Reply-To: <7a7b4693-8a28-423c-bbbb-8b16b4ee2de7@intel.com>


On 2/1/2026 10:38 PM, Nilawar, Badal wrote:
>
> On 30-01-2026 23:04, Belgaumkar, Vinay wrote:
>>
>> On 1/30/2026 7:20 AM, Nilawar, Badal wrote:
>>>
>>> On 16-01-2026 04:10, Vinay Belgaumkar wrote:
>>>> Dump forcewake status and ref counts for all domains as part
>>>> of this debugfs. This is the sample output from gt1-
>>>>
>>>> $ cat /sys/kernel/debug/dri//0/gt1/powergate_info
>>>> Media Power Gating Enabled: yes
>>>> Media Slice0 Power Gate Status: down
>>>> GSC Power Gate Status: down
>>>> GT.ref_count=0, GT.forcewake=0x10000
>>>> VDBox0.ref_count=0, VDBox0.forcewake=0x10000
>>>> VEBox0.ref_count=0, VEBox0.forcewake=0x10000
>>>> GSC.ref_count=0, GSC.forcewake=0x10000
>>>>
>>>> Signed-off-by: Vinay Belgaumkar<vinay.belgaumkar@intel.com>
>>>> ---
>>>>   drivers/gpu/drm/xe/xe_force_wake.c | 46 
>>>> ++++++++++++++++++++++++++----
>>>>   drivers/gpu/drm/xe/xe_force_wake.h | 11 +++++++
>>>>   drivers/gpu/drm/xe/xe_gt_idle.c    | 20 +++++++++++++
>>>>   3 files changed, 71 insertions(+), 6 deletions(-)
>>>>
>>>> diff --git a/drivers/gpu/drm/xe/xe_force_wake.c 
>>>> b/drivers/gpu/drm/xe/xe_force_wake.c
>>>> index 76e054f314ee..197e2197bd0a 100644
>>>> --- a/drivers/gpu/drm/xe/xe_force_wake.c
>>>> +++ b/drivers/gpu/drm/xe/xe_force_wake.c
>>>> @@ -148,12 +148,6 @@ static int domain_sleep_wait(struct xe_gt *gt,
>>>>       return __domain_wait(gt, domain, false);
>>>>   }
>>>>   -#define for_each_fw_domain_masked(domain__, mask__, fw__, tmp__) \
>>>> -    for (tmp__ = (mask__); tmp__; tmp__ &= ~BIT(ffs(tmp__) - 1)) \
>>>> -        for_each_if((domain__ = ((fw__)->domains + \
>>>> -                     (ffs(tmp__) - 1))) && \
>>>> -                     domain__->reg_ctl.addr)
>>>> -
>>>>   /**
>>>>    * xe_force_wake_get() : Increase the domain refcount
>>>>    * @fw: struct xe_force_wake
>>>> @@ -266,3 +260,43 @@ void xe_force_wake_put(struct xe_force_wake 
>>>> *fw, unsigned int fw_ref)
>>>>       xe_gt_WARN(gt, ack_fail, "Forcewake domain%s %#x failed to 
>>>> acknowledge sleep request\n",
>>>>              str_plural(hweight_long(ack_fail)), ack_fail);
>>>>   }
>>>> +
>>>> +const char *xe_force_wake_domain_to_str(enum 
>>>> xe_force_wake_domain_id id)
>>>> +{
>>>> +    switch (id) {
>>>> +    case XE_FW_DOMAIN_ID_GT:
>>>> +        return "GT";
>>>> +    case XE_FW_DOMAIN_ID_RENDER:
>>>> +        return "Render";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA:
>>>> +        return "Media";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX0:
>>>> +        return "VDBox0";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX1:
>>>> +        return "VDBox1";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX2:
>>>> +        return "VDBox2";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX3:
>>>> +        return "VDBox3";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX4:
>>>> +        return "VDBox4";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX5:
>>>> +        return "VDBox5";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX6:
>>>> +        return "VDBox6";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VDBOX7:
>>>> +        return "VDBox7";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VEBOX0:
>>>> +        return "VEBox0";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VEBOX1:
>>>> +        return "VEBox1";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VEBOX2:
>>>> +        return "VEBox2";
>>>> +    case XE_FW_DOMAIN_ID_MEDIA_VEBOX3:
>>>> +        return "VEBox3";
>>>> +    case XE_FW_DOMAIN_ID_GSC:
>>>> +        return "GSC";
>>>
>>> How about creating static look up table.
>>>
>>>     static const char * const domain_names[] = {
>>>         [XE_FW_DOMAIN_ID_GT] = "GT",
>>>         [XE_FW_DOMAIN_ID_RENDER] = "Render",
>>>         [XE_FW_DOMAIN_ID_MEDIA] = "Media",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX0] = "VDBox0",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX1] = "VDBox1",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX2] = "VDBox2",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX3] = "VDBox3",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX4] = "VDBox4",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX5] = "VDBox5",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX6] = "VDBox6",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VDBOX7] = "VDBox7",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VEBOX0] = "VEBox0",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VEBOX1] = "VEBox1",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VEBOX2] = "VEBox2",
>>>         [XE_FW_DOMAIN_ID_MEDIA_VEBOX3] = "VEBox3",
>>>         [XE_FW_DOMAIN_ID_GSC] = "GSC",
>>>     };
>>>
>>>     if (id < ARRAY_SIZE(domain_names) && domain_names[id])
>>>         return domain_names[id];
>>
>> I was trying to make it a little more dynamic where, if something 
>> changes in the FW table, we don't need to update 2 locations.
>
> Ok, but even with a switch-case statement, you’d still need to update 
> it whenever a new enum value is added.
> So, updates in two places can’t be completely avoided.

True. Similar thing was needed in sriov and guc code, and switch/case 
was used there. So, just following the same method to keep it uniform 
might be better? I believe i915 used the array definition method.

Thanks,

Vinay.

>
>>
>> Thanks,
>>
>> Vinay.
>>
>>>
>>> Thanks,
>>> Badal
>>>
>>>> +    default:
>>>> +        return "Unknown";
>>>> +    }
>>>> +}
>>>> diff --git a/drivers/gpu/drm/xe/xe_force_wake.h 
>>>> b/drivers/gpu/drm/xe/xe_force_wake.h
>>>> index 1e2198f6a007..f7690cb34ef7 100644
>>>> --- a/drivers/gpu/drm/xe/xe_force_wake.h
>>>> +++ b/drivers/gpu/drm/xe/xe_force_wake.h
>>>> @@ -19,6 +19,17 @@ unsigned int __must_check 
>>>> xe_force_wake_get(struct xe_force_wake *fw,
>>>>                           enum xe_force_wake_domains domains);
>>>>   void xe_force_wake_put(struct xe_force_wake *fw, unsigned int 
>>>> fw_ref);
>>>>   +const char *xe_force_wake_domain_to_str(enum 
>>>> xe_force_wake_domain_id id);
>>>> +
>>>> +#define for_each_fw_domain_masked(domain__, mask__, fw__, tmp__) \
>>>> +    for (tmp__ = (mask__); tmp__; tmp__ &= ~BIT(ffs(tmp__) - 1)) \
>>>> +        for_each_if((domain__ = ((fw__)->domains + \
>>>> +                     (ffs(tmp__) - 1))) && \
>>>> +                     domain__->reg_ctl.addr)
>>>> +
>>>> +#define for_each_fw_domain(domain__, fw__, tmp__) \
>>>> +    for_each_fw_domain_masked(domain__, fw__->initialized_domains, 
>>>> fw__, tmp__)
>>>> +
>>>>   static inline int
>>>>   xe_force_wake_ref(struct xe_force_wake *fw,
>>>>             enum xe_force_wake_domains domain)
>>>> diff --git a/drivers/gpu/drm/xe/xe_gt_idle.c 
>>>> b/drivers/gpu/drm/xe/xe_gt_idle.c
>>>> index 52436dcb6381..8e36202f1a4f 100644
>>>> --- a/drivers/gpu/drm/xe/xe_gt_idle.c
>>>> +++ b/drivers/gpu/drm/xe/xe_gt_idle.c
>>>> @@ -169,6 +169,24 @@ void xe_gt_idle_disable_pg(struct xe_gt *gt)
>>>>       xe_mmio_write32(&gt->mmio, POWERGATE_ENABLE, 
>>>> gtidle->powergate_enable);
>>>>   }
>>>>   +static void force_wake_domains_show(struct xe_gt *gt, struct 
>>>> drm_printer *p)
>>>> +{
>>>> +    struct xe_force_wake_domain *domain;
>>>> +    struct xe_force_wake *fw = gt_to_fw(gt);
>>>> +    unsigned int tmp;
>>>> +    unsigned long flags;
>>>> +
>>>> +    spin_lock_irqsave(&fw->lock, flags);
>>>> +    for_each_fw_domain(domain, fw, tmp) {
>>>> +        drm_printf(p, "%s.ref_count=%u, %s.fwake=0x%x\n",
>>>> + xe_force_wake_domain_to_str(domain->id),
>>>> +                   READ_ONCE(domain->ref),
>>>> + xe_force_wake_domain_to_str(domain->id),
>>>> +                   xe_mmio_read32(&gt->mmio, domain->reg_ctl));
>>>> +    }
>>>> +    spin_unlock_irqrestore(&fw->lock, flags);
>>>> +}
>>>> +
>>>>   /**
>>>>    * xe_gt_idle_pg_print - Xe powergating info
>>>>    * @gt: GT object
>>>> @@ -260,6 +278,8 @@ int xe_gt_idle_pg_print(struct xe_gt *gt, 
>>>> struct drm_printer *p)
>>>>                  str_up_down(pg_status & GSC_AWAKE_STATUS));
>>>>       }
>>>>   +    force_wake_domains_show(gt, p);
>>>> +
>>>>       return 0;
>>>>   }

  reply	other threads:[~2026-02-02 17:12 UTC|newest]

Thread overview: 14+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-01-15 22:40 [PATCH 0/2] drm/xe: Add more info to powergate_info debugfs Vinay Belgaumkar
2026-01-15 22:40 ` [PATCH 1/2] drm/xe: Add GSC to powergate_info Vinay Belgaumkar
2026-01-30 15:23   ` Nilawar, Badal
2026-01-15 22:40 ` [PATCH 2/2] drm/xe: Add forcewake status " Vinay Belgaumkar
2026-01-30 15:20   ` Nilawar, Badal
2026-01-30 17:34     ` Belgaumkar, Vinay
2026-02-02  6:38       ` Nilawar, Badal
2026-02-02 17:12         ` Belgaumkar, Vinay [this message]
2026-02-02 17:20           ` Nilawar, Badal
2026-02-03  6:07             ` Nilawar, Badal
2026-01-15 22:48 ` ✗ CI.checkpatch: warning for drm/xe: Add more info to powergate_info debugfs Patchwork
2026-01-15 22:50 ` ✓ CI.KUnit: success " Patchwork
2026-01-15 23:36 ` ✗ Xe.CI.BAT: failure " Patchwork
2026-01-16  4:34 ` ✗ Xe.CI.Full: " Patchwork

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=e40bb59b-d40e-46a1-9792-93019db06221@intel.com \
    --to=vinay.belgaumkar@intel.com \
    --cc=badal.nilawar@intel.com \
    --cc=intel-xe@lists.freedesktop.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox