From: Alan Huang <mmpgouride@gmail.com>
To: Joel Fernandes <joel@joelfernandes.org>
Cc: Gao Xiang <hsiangkao@linux.alibaba.com>,
paulmck@kernel.org, Sandeep Dhavale <dhavale@google.com>,
Frederic Weisbecker <frederic@kernel.org>,
Neeraj Upadhyay <quic_neeraju@quicinc.com>,
Josh Triplett <josh@joshtriplett.org>,
Boqun Feng <boqun.feng@gmail.com>,
Steven Rostedt <rostedt@goodmis.org>,
Mathieu Desnoyers <mathieu.desnoyers@efficios.com>,
Lai Jiangshan <jiangshanlai@gmail.com>,
Zqiang <qiang.zhang1211@gmail.com>,
Matthias Brugger <matthias.bgg@gmail.com>,
AngeloGioacchino Del Regno
<angelogioacchino.delregno@collabora.com>,
linux-erofs@lists.ozlabs.org, xiang@kernel.org,
Will Shiu <Will.Shiu@mediatek.com>,
kernel-team@android.com, rcu@vger.kernel.org,
linux-kernel@vger.kernel.org,
linux-arm-kernel@lists.infradead.org,
linux-mediatek@lists.infradead.org
Subject: Re: [PATCH v1] rcu: Fix and improve RCU read lock checks when !CONFIG_DEBUG_LOCK_ALLOC
Date: Fri, 14 Jul 2023 00:09:27 +0800 [thread overview]
Message-ID: <F7D5032D-908E-4227-8A38-AF740AC86CDC@gmail.com> (raw)
In-Reply-To: <CAEXW_YSM1yik4yWTgZoxCS9RM6TbsL26VCVCH=41+uMA8chfAQ@mail.gmail.com>
> 2023年7月13日 23:33,Joel Fernandes <joel@joelfernandes.org> 写道:
>
> On Thu, Jul 13, 2023 at 10:34 AM Gao Xiang <hsiangkao@linux.alibaba.com> wrote:
>>
>>
>>
>> On 2023/7/13 22:07, Joel Fernandes wrote:
>>> On Thu, Jul 13, 2023 at 12:59 AM Gao Xiang <hsiangkao@linux.alibaba.com> wrote:
>>>> On 2023/7/13 12:52, Paul E. McKenney wrote:
>>>>> On Thu, Jul 13, 2023 at 12:41:09PM +0800, Gao Xiang wrote:
>>>>>>
>>>>>>
>>>>
>>>> ...
>>>>
>>>>>>
>>>>>> There are lots of performance issues here and even a plumber
>>>>>> topic last year to show that, see:
>>>>>>
>>>>>> [1] https://lore.kernel.org/r/20230519001709.2563-1-tj@kernel.org
>>>>>> [2] https://lore.kernel.org/r/CAHk-=wgE9kORADrDJ4nEsHHLirqPCZ1tGaEPAZejHdZ03qCOGg@mail.gmail.com
>>>>>> [3] https://lore.kernel.org/r/CAB=BE-SBtO6vcoyLNA9F-9VaN5R0t3o_Zn+FW8GbO6wyUqFneQ@mail.gmail.com
>>>>>> [4] https://lpc.events/event/16/contributions/1338/
>>>>>> and more.
>>>>>>
>>>>>> I'm not sure if it's necessary to look info all of that,
>>>>>> andSandeep knows more than I am (the scheduling issue
>>>>>> becomes vital on some aarch64 platform.)
>>>>>
>>>>> Hmmm... Please let me try again.
>>>>>
>>>>> Assuming that this approach turns out to make sense, the resulting
>>>>> patch will need to clearly state the performance benefits directly in
>>>>> the commit log.
>>>>>
>>>>> And of course, for the approach to make sense, it must avoid breaking
>>>>> the existing lockdep-RCU debugging code.
>>>>>
>>>>> Is that more clear?
>>>>
>>>> Personally I'm not working on Android platform any more so I don't
>>>> have a way to reproduce, hopefully Sandeep could give actually
>>>> number _again_ if dm-verity is enabled and trigger another
>>>> workqueue here and make a comparsion why the scheduling latency of
>>>> the extra work becomes unacceptable.
>>>>
>>>
>>> Question from my side, are we talking about only performance issues or
>>> also a crash? It appears z_erofs_decompress_pcluster() takes
>>> mutex_lock(&pcl->lock);
>>>
>>> So if it is either in an RCU read-side critical section or in an
>>> atomic section, like the softirq path, then it may
>>> schedule-while-atomic or trigger RCU warnings.
>>>
>>> z_erofs_decompressqueue_endio
>>> -> z_erofs_decompress_kickoff
>>> ->z_erofs_decompressqueue_work
>>> ->z_erofs_decompress_queue
>>> -> z_erofs_decompress_pcluster
>>> -> mutex_lock
>>>
>>
>> Why does the softirq path not trigger a workqueue instead?
>
> I said "if it is". I was giving a scenario. mutex_lock() is not
> allowed in softirq context or in an RCU-reader.
>
>>> Per Sandeep in [1], this stack happens under RCU read-lock in:
>>>
>>> #define __blk_mq_run_dispatch_ops(q, check_sleep, dispatch_ops) \
>>> [...]
>>> rcu_read_lock();
>>> (dispatch_ops);
>>> rcu_read_unlock();
>>> [...]
>>>
>>> Coming from:
>>> blk_mq_flush_plug_list ->
>>> blk_mq_run_dispatch_ops(q,
>>> __blk_mq_flush_plug_list(q, plug));
>>>
>>> and __blk_mq_flush_plug_list does this:
>>> q->mq_ops->queue_rqs(&plug->mq_list);
>>>
>>> This somehow ends up calling the bio_endio and the
>>> z_erofs_decompressqueue_endio which grabs the mutex.
>>>
>>> So... I have a question, it looks like one of the paths in
>>> __blk_mq_run_dispatch_ops() uses SRCU. Where are as the alternate
>>> path uses RCU. Why does this alternate want to block even if it is not
>>> supposed to? Is the real issue here that the BLK_MQ_F_BLOCKING should
>>> be set? It sounds like you want to block in the "else" path even
>>> though BLK_MQ_F_BLOCKING is not set:
>>
>> BLK_MQ_F_BLOCKING is not a flag that a filesystem can do anything with.
>> That is block layer and mq device driver stuffs. filesystems cannot set
>> this value.
>>
>> As I said, as far as I understand, previously,
>> .end_io() can only be called without RCU context, so it will be fine,
>> but I don't know when .end_io() can be called under some RCU context
>> now.
>
> From what Sandeep described, the code path is in an RCU reader. My
> question is more, why doesn't it use SRCU instead since it clearly
> does so if BLK_MQ_F_BLOCKING. What are the tradeoffs? IMHO, a deeper
> dive needs to be made into that before concluding that the fix is to
> use rcu_read_lock_any_held().
Copied from [1]:
"Background: Historically erofs would always schedule a kworker for
decompression which would incur the scheduling cost regardless of
the context. But z_erofs_decompressqueue_endio() may not always
be in atomic context and we could actually benefit from doing the
decompression in z_erofs_decompressqueue_endio() if we are in
thread context, for example when running with dm-verity.
This optimization was later added in patch [2] which has shown
improvement in performance benchmarks.”
I’m not sure if it is a design issue.
[1] https://lore.kernel.org/all/20230621220848.3379029-1-dhavale@google.com/
>
> - Joel
_______________________________________________
linux-arm-kernel mailing list
linux-arm-kernel@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
next prev parent reply other threads:[~2023-07-13 16:10 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-07-11 23:38 [PATCH v1] rcu: Fix and improve RCU read lock checks when !CONFIG_DEBUG_LOCK_ALLOC Sandeep Dhavale
2023-07-12 17:02 ` Joel Fernandes
2023-07-12 21:20 ` Sandeep Dhavale
2023-07-13 0:32 ` Joel Fernandes
2023-07-13 2:02 ` Gao Xiang
2023-07-13 2:10 ` Gao Xiang
2023-07-13 2:16 ` Joel Fernandes
2023-07-13 4:27 ` Paul E. McKenney
2023-07-13 4:41 ` Gao Xiang
2023-07-13 4:52 ` Paul E. McKenney
2023-07-13 4:59 ` Gao Xiang
2023-07-13 14:07 ` Joel Fernandes
2023-07-13 14:34 ` Gao Xiang
2023-07-13 15:33 ` Joel Fernandes
2023-07-13 16:09 ` Alan Huang [this message]
2023-07-13 18:14 ` Paul E. McKenney
2023-07-13 19:00 ` Gao Xiang
2023-07-13 22:27 ` Paul E. McKenney
2023-07-13 16:33 ` Paul E. McKenney
2023-07-13 17:05 ` Sandeep Dhavale
2023-07-13 17:35 ` Paul E. McKenney
2023-07-13 18:51 ` Sandeep Dhavale
2023-07-13 22:49 ` Paul E. McKenney
2023-07-13 23:08 ` Sandeep Dhavale
2023-07-13 23:28 ` Paul E. McKenney
2023-07-14 2:16 ` Paul E. McKenney
2023-07-14 3:16 ` Gao Xiang
2023-07-14 13:42 ` Joel Fernandes
2023-07-14 13:51 ` Gao Xiang
2023-07-14 14:56 ` Steven Rostedt
2023-07-14 15:13 ` Paul E. McKenney
2023-07-14 15:35 ` Alan Huang
2023-07-14 15:54 ` Alan Huang
2023-07-14 17:02 ` Paul E. McKenney
2023-07-14 18:40 ` Alan Huang
2023-07-14 18:44 ` Paul E. McKenney
2023-07-14 19:15 ` Sandeep Dhavale
2023-07-14 19:36 ` Paul E. McKenney
2023-07-13 4:51 ` Gao Xiang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=F7D5032D-908E-4227-8A38-AF740AC86CDC@gmail.com \
--to=mmpgouride@gmail.com \
--cc=Will.Shiu@mediatek.com \
--cc=angelogioacchino.delregno@collabora.com \
--cc=boqun.feng@gmail.com \
--cc=dhavale@google.com \
--cc=frederic@kernel.org \
--cc=hsiangkao@linux.alibaba.com \
--cc=jiangshanlai@gmail.com \
--cc=joel@joelfernandes.org \
--cc=josh@joshtriplett.org \
--cc=kernel-team@android.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-erofs@lists.ozlabs.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mediatek@lists.infradead.org \
--cc=mathieu.desnoyers@efficios.com \
--cc=matthias.bgg@gmail.com \
--cc=paulmck@kernel.org \
--cc=qiang.zhang1211@gmail.com \
--cc=quic_neeraju@quicinc.com \
--cc=rcu@vger.kernel.org \
--cc=rostedt@goodmis.org \
--cc=xiang@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).