public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Alan Huang <mmpgouride@gmail.com>
To: paulmck@kernel.org
Cc: Joel Fernandes <joel@joelfernandes.org>,
	Gao Xiang <hsiangkao@linux.alibaba.com>,
	Sandeep Dhavale <dhavale@google.com>,
	Frederic Weisbecker <frederic@kernel.org>,
	Neeraj Upadhyay <quic_neeraju@quicinc.com>,
	Josh Triplett <josh@joshtriplett.org>,
	Boqun Feng <boqun.feng@gmail.com>,
	Steven Rostedt <rostedt@goodmis.org>,
	Mathieu Desnoyers <mathieu.desnoyers@efficios.com>,
	Lai Jiangshan <jiangshanlai@gmail.com>,
	Zqiang <qiang.zhang1211@gmail.com>,
	Matthias Brugger <matthias.bgg@gmail.com>,
	AngeloGioacchino Del Regno 
	<angelogioacchino.delregno@collabora.com>,
	linux-erofs@lists.ozlabs.org, xiang@kernel.org,
	Will Shiu <Will.Shiu@mediatek.com>,
	kernel-team@android.com, rcu@vger.kernel.org,
	linux-kernel@vger.kernel.org,
	linux-arm-kernel@lists.infradead.org,
	linux-mediatek@lists.infradead.org
Subject: Re: [PATCH v1] rcu: Fix and improve RCU read lock checks when !CONFIG_DEBUG_LOCK_ALLOC
Date: Sat, 15 Jul 2023 02:40:08 +0800	[thread overview]
Message-ID: <D042B1CB-2ED4-4DF9-8CF5-5E455E7EAB73@gmail.com> (raw)
In-Reply-To: <e8ee7006-c1d0-4c04-bd25-0f518fb6534b@paulmck-laptop>


> 2023年7月15日 01:02,Paul E. McKenney <paulmck@kernel.org> 写道:
> 
> On Fri, Jul 14, 2023 at 11:54:47PM +0800, Alan Huang wrote:
>> 
>>> 2023年7月14日 23:35,Alan Huang <mmpgouride@gmail.com> 写道:
>>> 
>>>> 
>>>> 2023年7月14日 10:16,Paul E. McKenney <paulmck@kernel.org> 写道:
>>>> 
>>>> On Thu, Jul 13, 2023 at 09:33:35AM -0700, Paul E. McKenney wrote:
>>>>> On Thu, Jul 13, 2023 at 11:33:24AM -0400, Joel Fernandes wrote:
>>>>>> On Thu, Jul 13, 2023 at 10:34 AM Gao Xiang <hsiangkao@linux.alibaba.com> wrote:
>>>>>>> On 2023/7/13 22:07, Joel Fernandes wrote:
>>>>>>>> On Thu, Jul 13, 2023 at 12:59 AM Gao Xiang <hsiangkao@linux.alibaba.com> wrote:
>>>>>>>>> On 2023/7/13 12:52, Paul E. McKenney wrote:
>>>>>>>>>> On Thu, Jul 13, 2023 at 12:41:09PM +0800, Gao Xiang wrote:
>>>>>>>>> 
>>>>>>>>> ...
>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> There are lots of performance issues here and even a plumber
>>>>>>>>>>> topic last year to show that, see:
>>>>>>>>>>> 
>>>>>>>>>>> [1] https://lore.kernel.org/r/20230519001709.2563-1-tj@kernel.org
>>>>>>>>>>> [2] https://lore.kernel.org/r/CAHk-=wgE9kORADrDJ4nEsHHLirqPCZ1tGaEPAZejHdZ03qCOGg@mail.gmail.com
>>>>>>>>>>> [3] https://lore.kernel.org/r/CAB=BE-SBtO6vcoyLNA9F-9VaN5R0t3o_Zn+FW8GbO6wyUqFneQ@mail.gmail.com
>>>>>>>>>>> [4] https://lpc.events/event/16/contributions/1338/
>>>>>>>>>>> and more.
>>>>>>>>>>> 
>>>>>>>>>>> I'm not sure if it's necessary to look info all of that,
>>>>>>>>>>> andSandeep knows more than I am (the scheduling issue
>>>>>>>>>>> becomes vital on some aarch64 platform.)
>>>>>>>>>> 
>>>>>>>>>> Hmmm...  Please let me try again.
>>>>>>>>>> 
>>>>>>>>>> Assuming that this approach turns out to make sense, the resulting
>>>>>>>>>> patch will need to clearly state the performance benefits directly in
>>>>>>>>>> the commit log.
>>>>>>>>>> 
>>>>>>>>>> And of course, for the approach to make sense, it must avoid breaking
>>>>>>>>>> the existing lockdep-RCU debugging code.
>>>>>>>>>> 
>>>>>>>>>> Is that more clear?
>>>>>>>>> 
>>>>>>>>> Personally I'm not working on Android platform any more so I don't
>>>>>>>>> have a way to reproduce, hopefully Sandeep could give actually
>>>>>>>>> number _again_ if dm-verity is enabled and trigger another
>>>>>>>>> workqueue here and make a comparsion why the scheduling latency of
>>>>>>>>> the extra work becomes unacceptable.
>>>>>>>>> 
>>>>>>>> 
>>>>>>>> Question from my side, are we talking about only performance issues or
>>>>>>>> also a crash? It appears z_erofs_decompress_pcluster() takes
>>>>>>>> mutex_lock(&pcl->lock);
>>>>>>>> 
>>>>>>>> So if it is either in an RCU read-side critical section or in an
>>>>>>>> atomic section, like the softirq path, then it may
>>>>>>>> schedule-while-atomic or trigger RCU warnings.
>>>>>>>> 
>>>>>>>> z_erofs_decompressqueue_endio
>>>>>>>> -> z_erofs_decompress_kickoff
>>>>>>>> ->z_erofs_decompressqueue_work
>>>>>>>> ->z_erofs_decompress_queue
>>>>>>>>  -> z_erofs_decompress_pcluster
>>>>>>>>   -> mutex_lock
>>>>>>>> 
>>>>>>> 
>>>>>>> Why does the softirq path not trigger a workqueue instead?
>>>>>> 
>>>>>> I said "if it is". I was giving a scenario. mutex_lock() is not
>>>>>> allowed in softirq context or in an RCU-reader.
>>>>>> 
>>>>>>>> Per Sandeep in [1], this stack happens under RCU read-lock in:
>>>>>>>> 
>>>>>>>> #define __blk_mq_run_dispatch_ops(q, check_sleep, dispatch_ops) \
>>>>>>>> [...]
>>>>>>>>               rcu_read_lock();
>>>>>>>>               (dispatch_ops);
>>>>>>>>               rcu_read_unlock();
>>>>>>>> [...]
>>>>>>>> 
>>>>>>>> Coming from:
>>>>>>>> blk_mq_flush_plug_list ->
>>>>>>>>                          blk_mq_run_dispatch_ops(q,
>>>>>>>>                               __blk_mq_flush_plug_list(q, plug));
>>>>>>>> 
>>>>>>>> and __blk_mq_flush_plug_list does this:
>>>>>>>>         q->mq_ops->queue_rqs(&plug->mq_list);
>>>>>>>> 
>>>>>>>> This somehow ends up calling the bio_endio and the
>>>>>>>> z_erofs_decompressqueue_endio which grabs the mutex.
>>>>>>>> 
>>>>>>>> So... I have a question, it looks like one of the paths in
>>>>>>>> __blk_mq_run_dispatch_ops() uses SRCU.  Where are as the alternate
>>>>>>>> path uses RCU. Why does this alternate want to block even if it is not
>>>>>>>> supposed to? Is the real issue here that the BLK_MQ_F_BLOCKING should
>>>>>>>> be set? It sounds like you want to block in the "else" path even
>>>>>>>> though BLK_MQ_F_BLOCKING is not set:
>>>>>>> 
>>>>>>> BLK_MQ_F_BLOCKING is not a flag that a filesystem can do anything with.
>>>>>>> That is block layer and mq device driver stuffs. filesystems cannot set
>>>>>>> this value.
>>>>>>> 
>>>>>>> As I said, as far as I understand, previously,
>>>>>>> .end_io() can only be called without RCU context, so it will be fine,
>>>>>>> but I don't know when .end_io() can be called under some RCU context
>>>>>>> now.
>>>>>> 
>>>>>>> From what Sandeep described, the code path is in an RCU reader. My
>>>>>> question is more, why doesn't it use SRCU instead since it clearly
>>>>>> does so if BLK_MQ_F_BLOCKING. What are the tradeoffs? IMHO, a deeper
>>>>>> dive needs to be made into that before concluding that the fix is to
>>>>>> use rcu_read_lock_any_held().
>>>>> 
>>>>> How can this be solved?
>>>>> 
>>>>> 1. Always use a workqueue.  Simple, but is said to have performance
>>>>> issues.
>>>>> 
>>>>> 2. Pass a flag in that indicates whether or not the caller is in an
>>>>> RCU read-side critical section.  Conceptually simple, but might
>>>>> or might not be reasonable to actually implement in the code as
>>>>> it exists now. (You tell me!)
>>>>> 
>>>>> 3. Create a function in z_erofs that gives you a decent
>>>>> approximation, maybe something like the following.
>>>>> 
>>>>> 4. Other ideas here.
>>>> 
>>>> 5. #3 plus make the corresponding Kconfig option select
>>>> PREEMPT_COUNT, assuming that any users needing compression in
>>>> non-preemptible kernels are OK with PREEMPT_COUNT being set.
>>>> (Some users of non-preemptible kernels object strenuously
>>>> to the added overhead from CONFIG_PREEMPT_COUNT=y.)
>>> 
>>> 6. Set one bit in bio->bi_private, check the bit and flip it in rcu_read_lock() path,
>>> then in z_erofs_decompressqueue_endio, check if the bit has changed.
>> 
>> Seems bad, read and modify bi_private is a bad idea.
> 
> Is there some other field that would work?

Maybe bio->bi_opf, btrfs uses some bits of it.

> 
> Thanx, Paul
> 
>>> Not sure if this is feasible or acceptable. :)
>>> 
>>>> 
>>>> Thanx, Paul
>>>> 
>>>>> The following is untested, and is probably quite buggy, but it should
>>>>> provide you with a starting point.
>>>>> 
>>>>> static bool z_erofs_wq_needed(void)
>>>>> {
>>>>> if (IS_ENABLED(CONFIG_PREEMPTION) && rcu_preempt_depth())
>>>>> return true;  // RCU reader
>>>>> if (IS_ENABLED(CONFIG_PREEMPT_COUNT) && !preemptible())
>>>>> return true;  // non-preemptible
>>>>> if (!IS_ENABLED(CONFIG_PREEMPT_COUNT))
>>>>> return true;  // non-preeemptible kernel, so play it safe
>>>>> return false;
>>>>> }
>>>>> 
>>>>> You break it, you buy it!  ;-)
>>>>> 
>>>>> Thanx, Paul



  reply	other threads:[~2023-07-14 18:40 UTC|newest]

Thread overview: 39+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-07-11 23:38 [PATCH v1] rcu: Fix and improve RCU read lock checks when !CONFIG_DEBUG_LOCK_ALLOC Sandeep Dhavale
2023-07-12 17:02 ` Joel Fernandes
2023-07-12 21:20   ` Sandeep Dhavale
2023-07-13  0:32     ` Joel Fernandes
2023-07-13  2:02       ` Gao Xiang
2023-07-13  2:10         ` Gao Xiang
2023-07-13  2:16         ` Joel Fernandes
2023-07-13  4:27         ` Paul E. McKenney
2023-07-13  4:41           ` Gao Xiang
2023-07-13  4:52             ` Paul E. McKenney
2023-07-13  4:59               ` Gao Xiang
2023-07-13 14:07                 ` Joel Fernandes
2023-07-13 14:34                   ` Gao Xiang
2023-07-13 15:33                     ` Joel Fernandes
2023-07-13 16:09                       ` Alan Huang
2023-07-13 18:14                         ` Paul E. McKenney
2023-07-13 19:00                           ` Gao Xiang
2023-07-13 22:27                             ` Paul E. McKenney
2023-07-13 16:33                       ` Paul E. McKenney
2023-07-13 17:05                         ` Sandeep Dhavale
2023-07-13 17:35                           ` Paul E. McKenney
2023-07-13 18:51                             ` Sandeep Dhavale
2023-07-13 22:49                               ` Paul E. McKenney
2023-07-13 23:08                                 ` Sandeep Dhavale
2023-07-13 23:28                                   ` Paul E. McKenney
2023-07-14  2:16                         ` Paul E. McKenney
2023-07-14  3:16                           ` Gao Xiang
2023-07-14 13:42                             ` Joel Fernandes
2023-07-14 13:51                               ` Gao Xiang
2023-07-14 14:56                                 ` Steven Rostedt
2023-07-14 15:13                                   ` Paul E. McKenney
2023-07-14 15:35                           ` Alan Huang
2023-07-14 15:54                             ` Alan Huang
2023-07-14 17:02                               ` Paul E. McKenney
2023-07-14 18:40                                 ` Alan Huang [this message]
2023-07-14 18:44                                   ` Paul E. McKenney
2023-07-14 19:15                                     ` Sandeep Dhavale
2023-07-14 19:36                                       ` Paul E. McKenney
2023-07-13  4:51           ` Gao Xiang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=D042B1CB-2ED4-4DF9-8CF5-5E455E7EAB73@gmail.com \
    --to=mmpgouride@gmail.com \
    --cc=Will.Shiu@mediatek.com \
    --cc=angelogioacchino.delregno@collabora.com \
    --cc=boqun.feng@gmail.com \
    --cc=dhavale@google.com \
    --cc=frederic@kernel.org \
    --cc=hsiangkao@linux.alibaba.com \
    --cc=jiangshanlai@gmail.com \
    --cc=joel@joelfernandes.org \
    --cc=josh@joshtriplett.org \
    --cc=kernel-team@android.com \
    --cc=linux-arm-kernel@lists.infradead.org \
    --cc=linux-erofs@lists.ozlabs.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mediatek@lists.infradead.org \
    --cc=mathieu.desnoyers@efficios.com \
    --cc=matthias.bgg@gmail.com \
    --cc=paulmck@kernel.org \
    --cc=qiang.zhang1211@gmail.com \
    --cc=quic_neeraju@quicinc.com \
    --cc=rcu@vger.kernel.org \
    --cc=rostedt@goodmis.org \
    --cc=xiang@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox