From: Carlos Maiolino <cem@kernel.org>
To: "Darrick J. Wong" <djwong@kernel.org>
Cc: linux-xfs@vger.kernel.org
Subject: Re: [ANNOUNCE] xfsprogs for-next updated
Date: Thu, 1 Sep 2022 12:55:08 +0200 [thread overview]
Message-ID: <20220901105508.kg5nu7xet6ylujcz@andromeda> (raw)
In-Reply-To: <Yw95W6ZAvBqQe7qf@magnolia>
>
> Oh, the usual problems of adding a new interface...
>
> 1. Who is responsible for setting m_sb_bp?
>
> Should libxfs_mount attach m_sb_bp? Should individual programs decide
> to do that if they require the functionality? Should we instead have a
> xfs_getsb function that returns m_sb_bp if set, or libxfs_getbufr's a
> new buffer and tries to cmpxchg it with the pointer?
>
> What about mkfs, which needs to libxfs_mount before it's even written
> anything to disk?
>
> 2. Should it be a cached buffer so that any other program (e.g. xfs_db)
> doing open-coded accesses of the superblock will get the same cached
> buffer, or should it be uncached like the kernel?
>
> If we decide on uncached, this will necessitate a full audit of xfsprogs
> to catch open-coded calls to libxfs_getbuf for the primary super, or
> else coherency problems will result.
>
> If we decide on using a cached buffer and setting it in libxfs_mount,
> then the part of xfs_repair that tears down the buffer cache and
> reinitializes it with a different hash size will also have to learn to
> brelse m_sb_bp before destroying the cache and re-assign it afterwards.
> Alternately, I suppose it could learn to rehash itself.
>
> This is a /lot/ to think about to solve one problem in one program.
Yeah, I do agree. Maybe it's better to go with your initial idea for 6.0, and we
postpone the SB buffer pinning to a later release avoiding people to hit this
issue ASAP?!
>
> > I didn't have time to try to reproduce those deadlocks yet though.
>
> If you modify cache_node_get like this to make reclaim more aggressive:
>
> diff --git a/libxfs/cache.c b/libxfs/cache.c
> index 139c7c1b..b5e1bcf8 100644
> --- a/libxfs/cache.c
> +++ b/libxfs/cache.c
> @@ -448,10 +448,10 @@ cache_node_get(
> /*
> * not found, allocate a new entry
> */
> + priority = cache_shake(cache, priority, false);
> node = cache_node_allocate(cache, key);
> if (node)
> break;
> - priority = cache_shake(cache, priority, false);
> /*
> * We start at 0; if we free CACHE_SHAKE_COUNT we get
> * back the same priority, if not we get back priority+1.
>
> It's trivially reproducible with xfs_repair (do not specify -n).
Thanks, I'm gonna try it :)
>
> --D
>
> > >
> > > [1] https://lore.kernel.org/linux-xfs/166007921743.3294543.7334567013352169774.stgit@magnolia/
> > > [2] https://lore.kernel.org/linux-xfs/20220811221541.GQ3600936@dread.disaster.area/
> >
> > --
> > Carlos Maiolino
--
Carlos Maiolino
next prev parent reply other threads:[~2022-09-01 10:56 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-08-30 11:52 [ANNOUNCE] xfsprogs for-next updated Carlos Maiolino
2022-08-30 15:12 ` Darrick J. Wong
2022-08-31 9:43 ` Carlos Maiolino
2022-08-31 15:08 ` Darrick J. Wong
2022-09-01 10:55 ` Carlos Maiolino [this message]
-- strict thread matches above, loose matches on Subject: below --
2022-11-16 9:58 Carlos Maiolino
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20220901105508.kg5nu7xet6ylujcz@andromeda \
--to=cem@kernel.org \
--cc=djwong@kernel.org \
--cc=linux-xfs@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).