From: Trond Myklebust <trondmy@hammerspace.com>
To: "david@fromorbit.com" <david@fromorbit.com>
Cc: "bfoster@redhat.com" <bfoster@redhat.com>,
"linux-fsdevel@vger.kernel.org" <linux-fsdevel@vger.kernel.org>,
"trondmy@kernel.org" <trondmy@kernel.org>,
"hch@infradead.org" <hch@infradead.org>,
"axboe@kernel.dk" <axboe@kernel.dk>,
"djwong@kernel.org" <djwong@kernel.org>,
"linux-xfs@vger.kernel.org" <linux-xfs@vger.kernel.org>,
"willy@infradead.org" <willy@infradead.org>
Subject: Re: [PATCH] iomap: Address soft lockup in iomap_finish_ioend()
Date: Tue, 4 Jan 2022 03:01:36 +0000 [thread overview]
Message-ID: <4072d4a2990d4739b08273ec0efda9fc0c5f1d55.camel@hammerspace.com> (raw)
In-Reply-To: <20220104012215.GH945095@dread.disaster.area>
On Tue, 2022-01-04 at 12:22 +1100, Dave Chinner wrote:
> On Tue, Jan 04, 2022 at 12:04:23AM +0000, Trond Myklebust wrote:
> > On Tue, 2022-01-04 at 09:03 +1100, Dave Chinner wrote:
> > > On Sat, Jan 01, 2022 at 05:39:45PM +0000, Trond Myklebust wrote:
> > > > On Sat, 2022-01-01 at 14:55 +1100, Dave Chinner wrote:
> > > > > As it is, if you are getting soft lockups in this location,
> > > > > that's
> > > > > an indication that the ioend chain that is being built by XFS
> > > > > is
> > > > > way, way too long. IOWs, the completion latency problem is
> > > > > caused
> > > > > by
> > > > > a lack of submit side ioend chain length bounding in
> > > > > combination
> > > > > with unbound completion side merging in xfs_end_bio - it's
> > > > > not a
> > > > > problem with the generic iomap code....
> > > > >
> > > > > Let's try to address this in the XFS code, rather than hack
> > > > > unnecessary band-aids over the problem in the generic code...
> > > > >
> > > > > Cheers,
> > > > >
> > > > > Dave.
> > > >
> > > > Fair enough. As long as someone is working on a solution, then
> > > > I'm
> > > > happy. Just a couple of things:
> > > >
> > > > Firstly, we've verified that the cond_resched() in the bio loop
> > > > does
> > > > suffice to resolve the issue with XFS, which would tend to
> > > > confirm
> > > > what
> > > > you're saying above about the underlying issue being the ioend
> > > > chain
> > > > length.
> > > >
> > > > Secondly, note that we've tested this issue with a variety of
> > > > older
> > > > kernels, including 4.18.x, 5.1.x and 5.15.x, so please bear in
> > > > mind
> > > > that it would be useful for any fix to be backward portable
> > > > through
> > > > the
> > > > stable mechanism.
> > >
> > > The infrastructure hasn't changed that much, so whatever the
> > > result
> > > is it should be backportable.
> > >
> > > As it is, is there a specific workload that triggers this issue?
> > > Or
> > > a specific machine config (e.g. large memory, slow storage). Are
> > > there large fragmented files in use (e.g. randomly written VM
> > > image
> > > files)? There are a few factors that can exacerbate the ioend
> > > chain
> > > lengths, so it would be handy to have some idea of what is
> > > actually
> > > triggering this behaviour...
> > >
> > > Cheers,
> > >
> > > Dave.
> >
> > We have different reproducers. The common feature appears to be the
> > need for a decently fast box with fairly large memory (128GB in one
> > case, 400GB in the other). It has been reproduced with HDs, SSDs
> > and
> > NVME systems.
> >
> > On the 128GB box, we had it set up with 10+ disks in a JBOD
> > configuration and were running the AJA system tests.
> >
> > On the 400GB box, we were just serially creating large (> 6GB)
> > files
> > using fio and that was occasionally triggering the issue. However
> > doing
> > an strace of that workload to disk reproduced the problem faster :-
> > ).
>
> Ok, that matches up with the "lots of logically sequential dirty
> data on a single inode in cache" vector that is required to create
> really long bio chains on individual ioends.
>
> Can you try the patch below and see if addresses the issue?
>
> Cheers,
>
> Dave.
Thanks Dave!
I'm building a new kernel for testing now and should have results ready
tomorrow at the latest.
--
Trond Myklebust
Linux NFS client maintainer, Hammerspace
trond.myklebust@hammerspace.com
next prev parent reply other threads:[~2022-01-04 3:01 UTC|newest]
Thread overview: 51+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-12-30 19:35 [PATCH] iomap: Address soft lockup in iomap_finish_ioend() trondmy
2021-12-30 21:24 ` Jens Axboe
2021-12-30 22:25 ` Trond Myklebust
2021-12-30 22:27 ` Jens Axboe
2021-12-30 22:55 ` Trond Myklebust
2021-12-31 1:42 ` Matthew Wilcox
2021-12-31 6:16 ` Trond Myklebust
2022-01-01 3:55 ` Dave Chinner
2022-01-01 17:39 ` Trond Myklebust
2022-01-03 22:03 ` Dave Chinner
2022-01-04 0:04 ` Trond Myklebust
2022-01-04 1:22 ` Dave Chinner
2022-01-04 3:01 ` Trond Myklebust [this message]
2022-01-04 7:08 ` hch
2022-01-04 18:08 ` Matthew Wilcox
2022-01-04 18:14 ` hch
2022-01-04 19:22 ` Darrick J. Wong
2022-01-04 21:52 ` Dave Chinner
2022-01-04 23:12 ` Darrick J. Wong
2022-01-05 2:10 ` Dave Chinner
2022-01-05 13:56 ` Brian Foster
2022-01-05 22:04 ` Dave Chinner
2022-01-06 16:44 ` Brian Foster
2022-01-10 8:18 ` Dave Chinner
2022-01-10 17:45 ` Brian Foster
2022-01-10 18:11 ` hch
2022-01-11 14:33 ` Trond Myklebust
2022-01-05 13:42 ` hch
2022-01-04 21:16 ` Dave Chinner
2022-01-05 13:43 ` hch
2022-01-05 22:34 ` Dave Chinner
2022-01-05 2:09 ` Trond Myklebust
2022-01-05 20:45 ` Trond Myklebust
2022-01-05 22:48 ` Dave Chinner
2022-01-05 23:29 ` Trond Myklebust
2022-01-06 0:01 ` Darrick J. Wong
2022-01-09 23:09 ` Dave Chinner
2022-01-06 18:36 ` Trond Myklebust
2022-01-06 18:38 ` Trond Myklebust
2022-01-06 20:07 ` Brian Foster
2022-01-07 3:08 ` Trond Myklebust
2022-01-07 15:15 ` Brian Foster
2022-01-09 23:34 ` Dave Chinner
2022-01-10 23:37 ` Dave Chinner
2022-01-11 0:08 ` Dave Chinner
2022-01-13 17:01 ` Trond Myklebust
2022-01-17 17:24 ` Trond Myklebust
2022-01-17 17:36 ` Darrick J. Wong
2022-01-04 13:36 ` Brian Foster
2022-01-04 19:23 ` Darrick J. Wong
2022-01-05 2:31 ` [iomap] f5934dda54: BUG:sleeping_function_called_from_invalid_context_at_fs/iomap/buffered-io.c kernel test robot
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4072d4a2990d4739b08273ec0efda9fc0c5f1d55.camel@hammerspace.com \
--to=trondmy@hammerspace.com \
--cc=axboe@kernel.dk \
--cc=bfoster@redhat.com \
--cc=david@fromorbit.com \
--cc=djwong@kernel.org \
--cc=hch@infradead.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-xfs@vger.kernel.org \
--cc=trondmy@kernel.org \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox