From: Greg KH <gregkh@linuxfoundation.org>
To: peter.wang@mediatek.com
Cc: stanley.chu@mediatek.com, linux-scsi@vger.kernel.org,
martin.petersen@oracle.com, avri.altman@wdc.com,
alim.akhtar@samsung.com, jejb@linux.ibm.com,
wsd_upstream@mediatek.com, linux-mediatek@lists.infradead.org,
chun-hung.wu@mediatek.com, alice.chao@mediatek.com,
cc.chou@mediatek.com, chaotian.jing@mediatek.com,
jiajie.hao@mediatek.com, powen.kao@mediatek.com,
qilin.tan@mediatek.com, lin.gui@mediatek.com,
stable@vger.kernel.org
Subject: Re: [PATCH v3] ufs: core: fix lockdep warning of clk_scaling_lock
Date: Tue, 26 Jul 2022 17:05:24 +0200 [thread overview]
Message-ID: <YuACtFnW6QJ7QFyq@kroah.com> (raw)
In-Reply-To: <20220726091433.22755-1-peter.wang@mediatek.com>
On Tue, Jul 26, 2022 at 05:14:33PM +0800, peter.wang@mediatek.com wrote:
> From: Peter Wang <peter.wang@mediatek.com>
>
> There have a lockdep warning like below in current flow.
> kworker/u16:0: Possible unsafe locking scenario:
>
> kworker/u16:0: CPU0 CPU1
> kworker/u16:0: ---- ----
> kworker/u16:0: lock(&hba->clk_scaling_lock);
> kworker/u16:0: lock(&hba->dev_cmd.lock);
> kworker/u16:0: lock(&hba->clk_scaling_lock);
> kworker/u16:0: lock(&hba->dev_cmd.lock);
> kworker/u16:0:
>
> Before this patch clk_scaling_lock was held in reader mode during the ufshcd_wb_toggle() call.
> With this patch applied clk_scaling_lock is not held while ufshcd_wb_toggle() is called.
>
> This is safe because ufshcd_wb_toggle will held clk_scaling_lock in reader mode "again" in flow
> ufshcd_wb_toggle -> __ufshcd_wb_toggle -> ufshcd_query_flag_retry -> ufshcd_query_flag ->
> ufshcd_exec_dev_cmd -> down_read(&hba->clk_scaling_lock);
> The protect should enough and make sure clock is not change while send command.
>
> ufshcd_wb_toggle can protected by hba->clk_scaling.is_allowed to make sure
> ufshcd_devfreq_scale function not run concurrently.
>
> Fixes: 0e9d4ca43ba8 ("scsi: ufs: Protect some contexts from unexpected clock scaling")
> Signed-off-by: Peter Wang <peter.wang@mediatek.com>
> ---
> drivers/ufs/core/ufshcd.c | 27 ++++++++++++++++-----------
> 1 file changed, 16 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/ufs/core/ufshcd.c b/drivers/ufs/core/ufshcd.c
> index c7b337480e3e..aa57126fdb49 100644
> --- a/drivers/ufs/core/ufshcd.c
> +++ b/drivers/ufs/core/ufshcd.c
> @@ -272,6 +272,7 @@ static void ufshcd_wb_toggle_flush_during_h8(struct ufs_hba *hba, bool set);
> static inline void ufshcd_wb_toggle_flush(struct ufs_hba *hba, bool enable);
> static void ufshcd_hba_vreg_set_lpm(struct ufs_hba *hba);
> static void ufshcd_hba_vreg_set_hpm(struct ufs_hba *hba);
> +static void ufshcd_clk_scaling_allow(struct ufs_hba *hba, bool allow);
>
> static inline void ufshcd_enable_irq(struct ufs_hba *hba)
> {
> @@ -1249,12 +1250,10 @@ static int ufshcd_clock_scaling_prepare(struct ufs_hba *hba)
> return ret;
> }
>
> -static void ufshcd_clock_scaling_unprepare(struct ufs_hba *hba, bool writelock)
> +static void ufshcd_clock_scaling_unprepare(struct ufs_hba *hba)
> {
> - if (writelock)
> - up_write(&hba->clk_scaling_lock);
> - else
> - up_read(&hba->clk_scaling_lock);
> + up_write(&hba->clk_scaling_lock);
> +
> ufshcd_scsi_unblock_requests(hba);
> ufshcd_release(hba);
> }
> @@ -1271,7 +1270,7 @@ static void ufshcd_clock_scaling_unprepare(struct ufs_hba *hba, bool writelock)
> static int ufshcd_devfreq_scale(struct ufs_hba *hba, bool scale_up)
> {
> int ret = 0;
> - bool is_writelock = true;
> + bool wb_toggle = false;
>
> ret = ufshcd_clock_scaling_prepare(hba);
> if (ret)
> @@ -1300,13 +1299,19 @@ static int ufshcd_devfreq_scale(struct ufs_hba *hba, bool scale_up)
> }
> }
>
> - /* Enable Write Booster if we have scaled up else disable it */
> - downgrade_write(&hba->clk_scaling_lock);
> - is_writelock = false;
> - ufshcd_wb_toggle(hba, scale_up);
> + /* Disable clk_scaling until ufshcd_wb_toggle finish */
> + hba->clk_scaling.is_allowed = false;
> + wb_toggle = true;
>
> out_unprepare:
> - ufshcd_clock_scaling_unprepare(hba, is_writelock);
> + ufshcd_clock_scaling_unprepare(hba);
> +
> + /* Enable Write Booster if we have scaled up else disable it */
> + if (wb_toggle) {
> + ufshcd_wb_toggle(hba, scale_up);
> + ufshcd_clk_scaling_allow(hba, true);
> + }
> +
> return ret;
> }
>
> --
> 2.18.0
>
<formletter>
This is not the correct way to submit patches for inclusion in the
stable kernel tree. Please read:
https://www.kernel.org/doc/html/latest/process/stable-kernel-rules.html
for how to do this properly.
</formletter>
next prev parent reply other threads:[~2022-07-26 15:05 UTC|newest]
Thread overview: 3+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-07-26 9:14 [PATCH v3] ufs: core: fix lockdep warning of clk_scaling_lock peter.wang
2022-07-26 15:05 ` Greg KH [this message]
2022-07-27 3:08 ` Peter Wang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=YuACtFnW6QJ7QFyq@kroah.com \
--to=gregkh@linuxfoundation.org \
--cc=alice.chao@mediatek.com \
--cc=alim.akhtar@samsung.com \
--cc=avri.altman@wdc.com \
--cc=cc.chou@mediatek.com \
--cc=chaotian.jing@mediatek.com \
--cc=chun-hung.wu@mediatek.com \
--cc=jejb@linux.ibm.com \
--cc=jiajie.hao@mediatek.com \
--cc=lin.gui@mediatek.com \
--cc=linux-mediatek@lists.infradead.org \
--cc=linux-scsi@vger.kernel.org \
--cc=martin.petersen@oracle.com \
--cc=peter.wang@mediatek.com \
--cc=powen.kao@mediatek.com \
--cc=qilin.tan@mediatek.com \
--cc=stable@vger.kernel.org \
--cc=stanley.chu@mediatek.com \
--cc=wsd_upstream@mediatek.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox