From: Frederic Weisbecker <fweisbec@gmail.com>
To: Lai Jiangshan <laijs@cn.fujitsu.com>
Cc: Oleg Nesterov <oleg@tv-sign.ru>, Ingo Molnar <mingo@elte.hu>,
Andrew Morton <akpm@linux-foundation.org>,
Peter Zijlstra <peterz@infradead.org>,
Eric Dumazet <dada1@cosmosbay.com>,
Linux Kernel Mailing List <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH 2/3] workqueue: not allow recursion run_workqueue
Date: Thu, 22 Jan 2009 10:39:38 +0100 [thread overview]
Message-ID: <20090122093937.GA6009@nowhere> (raw)
In-Reply-To: <20090122093046.GC5891@nowhere>
On Thu, Jan 22, 2009 at 10:30:46AM +0100, Frederic Weisbecker wrote:
> On Thu, Jan 22, 2009 at 05:14:24PM +0800, Lai Jiangshan wrote:
> > 1) lockdep will complain when recursion run_workqueue
> > 2) works is not run orderly when recursion run_workqueue
> >
> > 3) BUG!
> > We use recursion run_workqueue to hidden deadlock when
> > keventd trying to flush its own queue.
> >
> > It's bug. When flush_workqueue()(nested in a work callback)returns,
> > the workqueue is not really flushed, the sequence statement of
> > this work callback will do some thing bad.
> >
> > So we should not allow workqueue trying to flush its own queue.
> >
> > Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
> > ---
> > diff --git a/kernel/workqueue.c b/kernel/workqueue.c
> > index 2f44583..1129cde 100644
> > --- a/kernel/workqueue.c
> > +++ b/kernel/workqueue.c
> > @@ -48,8 +48,6 @@ struct cpu_workqueue_struct {
> >
> > struct workqueue_struct *wq;
> > struct task_struct *thread;
> > -
> > - int run_depth; /* Detect run_workqueue() recursion depth */
> > } ____cacheline_aligned;
> >
> > /*
> > @@ -262,13 +260,6 @@ EXPORT_SYMBOL_GPL(queue_delayed_work_on);
> > static void run_workqueue(struct cpu_workqueue_struct *cwq)
> > {
> > spin_lock_irq(&cwq->lock);
> > - cwq->run_depth++;
> > - if (cwq->run_depth > 3) {
> > - /* morton gets to eat his hat */
> > - printk("%s: recursion depth exceeded: %d\n",
> > - __func__, cwq->run_depth);
> > - dump_stack();
> > - }
> > while (!list_empty(&cwq->worklist)) {
> > struct work_struct *work = list_entry(cwq->worklist.next,
> > struct work_struct, entry);
> > @@ -311,7 +302,6 @@ static void run_workqueue(struct cpu_workqueue_struct *cwq)
> > spin_lock_irq(&cwq->lock);
> > cwq->current_work = NULL;
> > }
> > - cwq->run_depth--;
> > spin_unlock_irq(&cwq->lock);
> > }
> >
> > @@ -368,29 +358,20 @@ static void insert_wq_barrier(struct cpu_workqueue_struct *cwq,
> >
> > static int flush_cpu_workqueue(struct cpu_workqueue_struct *cwq)
> > {
> > - int active;
> > + int active = 0;
> > + struct wq_barrier barr;
> >
> > - if (cwq->thread == current) {
> > - /*
> > - * Probably keventd trying to flush its own queue. So simply run
> > - * it by hand rather than deadlocking.
> > - */
> > - run_workqueue(cwq);
> > - active = 1;
> > - } else {
> > - struct wq_barrier barr;
> > + BUG_ON(cwq->thread == current);
>
> Hi Lai,
>
> BUG_ON seems perhaps a bit too much for such case. The system
> will run in an endless loop because of a mistake that will not have
> necessarily a fatal end.
> WARN_ON should be enough (plus the warn that lockdep will raise
> too in this case).
>
> Thanks.
> Frederic.
And perhaps add a comment for the developers who will encounter such a warn,
and then fall down in this call site while searching which warned.
To easily find the reason of the WARN. cwq->thread == current is perhaps not
verbose enough to help the developer finding the source of the problem.
They could solve the issue and say "Doh!" more quickly if they find
in a one shot sight: /* Never flush a workqueue from a work */
:-)
>
> > - active = 0;
> > - spin_lock_irq(&cwq->lock);
> > - if (!list_empty(&cwq->worklist) || cwq->current_work != NULL) {
> > - insert_wq_barrier(cwq, &barr, &cwq->worklist);
> > - active = 1;
> > - }
> > - spin_unlock_irq(&cwq->lock);
> > -
> > - if (active)
> > - wait_for_completion(&barr.done);
> > + spin_lock_irq(&cwq->lock);
> > + if (!list_empty(&cwq->worklist) || cwq->current_work != NULL) {
> > + insert_wq_barrier(cwq, &barr, &cwq->worklist);
> > + active = 1;
> > }
> > + spin_unlock_irq(&cwq->lock);
> > +
> > + if (active)
> > + wait_for_completion(&barr.done);
> >
> > return active;
> > }
> >
next prev parent reply other threads:[~2009-01-22 9:39 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-01-22 9:14 [PATCH 2/3] workqueue: not allow recursion run_workqueue Lai Jiangshan
2009-01-22 9:30 ` Frederic Weisbecker
2009-01-22 9:36 ` Ingo Molnar
2009-01-22 11:06 ` Frédéric Weisbecker
2009-01-22 11:10 ` Peter Zijlstra
2009-02-05 8:18 ` Lai Jiangshan
2009-02-05 13:47 ` Frederic Weisbecker
2009-02-05 17:01 ` Oleg Nesterov
2009-02-05 17:24 ` Frederic Weisbecker
2009-02-05 18:00 ` Oleg Nesterov
2009-02-06 1:20 ` Lai Jiangshan
2009-02-06 16:46 ` Oleg Nesterov
2009-02-09 7:20 ` Lai Jiangshan
2009-02-06 1:46 ` Lai Jiangshan
2009-02-09 19:14 ` Oleg Nesterov
2009-02-10 20:53 ` Andrew Morton
2009-01-22 9:39 ` Frederic Weisbecker [this message]
2009-01-22 17:23 ` Oleg Nesterov
2009-01-22 17:47 ` Frédéric Weisbecker
2009-01-22 18:22 ` Oleg Nesterov
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20090122093937.GA6009@nowhere \
--to=fweisbec@gmail.com \
--cc=akpm@linux-foundation.org \
--cc=dada1@cosmosbay.com \
--cc=laijs@cn.fujitsu.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=oleg@tv-sign.ru \
--cc=peterz@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox