linux-rtc.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Alexandre Belloni <alexandre.belloni@bootlin.com>
To: Tetsuo Handa <penguin-kernel@i-love.sakura.ne.jp>
Cc: Alessandro Zummo <a.zummo@towertech.it>, linux-rtc@vger.kernel.org
Subject: Re: [PATCH v3] rtc: Avoid flush_scheduled_work() usage
Date: Fri, 20 May 2022 18:19:44 +0200	[thread overview]
Message-ID: <Yoe/oPW8MFZ02fEn@mail.local> (raw)
In-Reply-To: <ef8d46a0-bef8-f747-8f13-7ad4ec514cbf@I-love.SAKURA.ne.jp>

Hello,

On 20/05/2022 23:33:47+0900, Tetsuo Handa wrote:
> Flushing system-wide workqueues is dangerous and will be forbidden.
> Replace system_wq with local rtc_wq.
> 
> While we are at it, remove unused rtc_dev_exit().
> 
> Link: https://lkml.kernel.org/r/49925af7-78a8-a3dd-bce6-cfc02e1a9236@I-love.SAKURA.ne.jp
> Signed-off-by: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
> ---
> Changes in v3:
>   Fix build error.
>   Forgot to replace flush_scheduled_work().
> 
> Changes in v2:
>   Remove unused rtc_dev_exit().
> 
> Since rtc_dev_init() is built into vmlinux, there is no point with
> recovery.
> 
> This patch blindly converts schedule_work() into queue_work() within
> drivers/rtc/dev.c, based on an assumption that none of work items
> outside of drivers/rtc/dev.c needs to be handled by rtc_wq.
> Did I convert correctly?
> 

Yes and no, this could be a bit more clever and create the workqueue
only for the devices that actually need it. I worked on something after
seeing your first email a while ago but I needed more time to test it.

>  drivers/rtc/dev.c      | 18 ++++++++----------
>  drivers/rtc/rtc-core.h |  5 -----
>  2 files changed, 8 insertions(+), 15 deletions(-)
> 
> diff --git a/drivers/rtc/dev.c b/drivers/rtc/dev.c
> index 69325aeede1a..a7346f03a5b2 100644
> --- a/drivers/rtc/dev.c
> +++ b/drivers/rtc/dev.c
> @@ -17,6 +17,7 @@
>  #include "rtc-core.h"
>  
>  static dev_t rtc_devt;
> +static struct workqueue_struct *rtc_wq;
>  
>  #define RTC_DEV_MAX 16 /* 16 RTCs should be enough for everyone... */
>  
> @@ -62,7 +63,7 @@ static void rtc_uie_task(struct work_struct *work)
>  		rtc->uie_timer_active = 1;
>  		rtc->uie_task_active = 0;
>  		add_timer(&rtc->uie_timer);
> -	} else if (schedule_work(&rtc->uie_task) == 0) {
> +	} else if (queue_work(rtc_wq, &rtc->uie_task) == 0) {
>  		rtc->uie_task_active = 0;
>  	}
>  	spin_unlock_irq(&rtc->irq_lock);
> @@ -78,7 +79,7 @@ static void rtc_uie_timer(struct timer_list *t)
>  	spin_lock_irqsave(&rtc->irq_lock, flags);
>  	rtc->uie_timer_active = 0;
>  	rtc->uie_task_active = 1;
> -	if ((schedule_work(&rtc->uie_task) == 0))
> +	if (queue_work(rtc_wq, &rtc->uie_task) == 0)
>  		rtc->uie_task_active = 0;
>  	spin_unlock_irqrestore(&rtc->irq_lock, flags);
>  }
> @@ -96,7 +97,7 @@ static int clear_uie(struct rtc_device *rtc)
>  		}
>  		if (rtc->uie_task_active) {
>  			spin_unlock_irq(&rtc->irq_lock);
> -			flush_scheduled_work();
> +			flush_workqueue(rtc_wq);
>  			spin_lock_irq(&rtc->irq_lock);
>  		}
>  		rtc->uie_irq_active = 0;
> @@ -119,7 +120,7 @@ static int set_uie(struct rtc_device *rtc)
>  		rtc->stop_uie_polling = 0;
>  		rtc->oldsecs = tm.tm_sec;
>  		rtc->uie_task_active = 1;
> -		if (schedule_work(&rtc->uie_task) == 0)
> +		if (queue_work(rtc_wq, &rtc->uie_task) == 0)
>  			rtc->uie_task_active = 0;
>  	}
>  	rtc->irq_data = 0;
> @@ -562,13 +563,10 @@ void __init rtc_dev_init(void)
>  {
>  	int err;
>  
> +	rtc_wq = alloc_workqueue("rtc_wq", 0, 0);
> +	BUG_ON(!rtc_wq);
> +
>  	err = alloc_chrdev_region(&rtc_devt, 0, RTC_DEV_MAX, "rtc");
>  	if (err < 0)
>  		pr_err("failed to allocate char dev region\n");
>  }
> -
> -void __exit rtc_dev_exit(void)
> -{
> -	if (rtc_devt)
> -		unregister_chrdev_region(rtc_devt, RTC_DEV_MAX);

This is very unrelated and should be in a different patch.

> -}
> diff --git a/drivers/rtc/rtc-core.h b/drivers/rtc/rtc-core.h
> index 0abf98983e13..4b10a1b8f370 100644
> --- a/drivers/rtc/rtc-core.h
> +++ b/drivers/rtc/rtc-core.h
> @@ -2,7 +2,6 @@
>  #ifdef CONFIG_RTC_INTF_DEV
>  
>  extern void __init rtc_dev_init(void);
> -extern void __exit rtc_dev_exit(void);
>  extern void rtc_dev_prepare(struct rtc_device *rtc);
>  
>  #else
> @@ -11,10 +10,6 @@ static inline void rtc_dev_init(void)
>  {
>  }
>  
> -static inline void rtc_dev_exit(void)
> -{
> -}
> -
>  static inline void rtc_dev_prepare(struct rtc_device *rtc)
>  {
>  }
> -- 
> 2.34.1
> 
> 

-- 
Alexandre Belloni, co-owner and COO, Bootlin
Embedded Linux and Kernel engineering
https://bootlin.com

  reply	other threads:[~2022-05-20 16:19 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-04-19 14:27 [PATCH] rtc: Avoid flush_scheduled_work() usage Tetsuo Handa
2022-04-30 10:25 ` [PATCH v2] " Tetsuo Handa
2022-05-17  4:14   ` Tetsuo Handa
2022-05-20 14:33   ` [PATCH v3] " Tetsuo Handa
2022-05-20 16:19     ` Alexandre Belloni [this message]
2022-06-10 10:48       ` [PATCH v4] rtc: Replace flush_scheduled_work() with flush_work() Tetsuo Handa
2022-06-24 17:14         ` Alexandre Belloni

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=Yoe/oPW8MFZ02fEn@mail.local \
    --to=alexandre.belloni@bootlin.com \
    --cc=a.zummo@towertech.it \
    --cc=linux-rtc@vger.kernel.org \
    --cc=penguin-kernel@i-love.sakura.ne.jp \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).