From: Nick Piggin <npiggin@kernel.dk>
To: Peter Zijlstra <peterz@infradead.org>
Cc: Christoph Hellwig <hch@infradead.org>,
Ingo Molnar <mingo@redhat.com>, Nick Piggin <npiggin@kernel.dk>,
xfs@oss.sgi.com
Subject: Re: XFS reclaim lock order bug
Date: Thu, 25 Nov 2010 18:08:17 +1100 [thread overview]
Message-ID: <20101125070817.GA4120@amd> (raw)
In-Reply-To: <1290666325.2072.535.camel@laptop>
On Thu, Nov 25, 2010 at 07:25:25AM +0100, Peter Zijlstra wrote:
> On Thu, 2010-11-25 at 14:48 +1100, Nick Piggin wrote:
> > On Wed, Nov 24, 2010 at 03:03:41PM -0500, Christoph Hellwig wrote:
> > > On Wed, Nov 24, 2010 at 08:12:58AM +1100, Dave Chinner wrote:
> > > > It is supposed to be handled by the re-initialisation of the
> > > > ip->i_iolock in ->evict_inode (xfs_fs_evict_inode). An inode found
> > > > in the reclaim state must have passed through this reinitialisation,
> > > > so from a lockdep perspective the iolock in the vfs path is a
> > > > different context to the iolock in the reclaim path. That fixed all
> > > > the non-reclaim state related lockdep false positives, so Perhaps
> > > > there is an issue with the lockdep reclaim state checking that does
> > > > not interact well with re-initialised lock contexts?
> > >
> > > I've been looking through this again, and I think it's indeed not
> > > enough. We don't just need to re-initialize it, but also set a
> > > different lockclass for it.
> >
> > Doesn't init_rwsem give it a new class?
>
> Per call-site, yes it should.
>
> > Guys, can you take a quick look at the code Dave is referring to
> > (xfs_fs_evict_inode), and check that it actually does what he
> > intends?
>
> Right, so this is trying to set a different class from the regular init
> site, which (/me applies grep) lives in xfs_inode_alloc(), right?
>
> Ought to work.. assuming the inode will be fully destroyed and new
> inodes are always obtained through xfs_inode_alloc() and not reused.
>
> > We're getting what seems to be false positives in reclaim inversion
> > of lockings. Backtraces here
> > http://oss.sgi.com/pipermail/xfs/2010-November/048092.html
>
> Right, so there its holding the inode in the read path while taking a
> page-fault which does an allocation.
>
> vs
>
> acquiring the inode in the xfs_reclaim_node_shrink() path.
>
>
> Presumably the whole xfs_fs_evict_inode() stuff will happen _after_ its
> possible to end up in that read path?
I think that's the idea.
> Something like the below would give the lock-class an explicit name,
> because both sites now use the exact same init thing they're called:
>
> (&(&ip->i_iolock)->mr_lock)
> (&(&ip->i_iolock)->mr_lock#2)
>
> Which is hard to tell apart, but I suspect #2 is the dead one, since
> they get numbered in order of appearance and its hard to have a dead
> inode before having a life one ;-)
>
> In that case though, it would suggest the inode got re-used instead of
> destroyed and re-created using xfs_alloc_inode(), is that at all
> possible?
Ah, I see what you mean. An inode that has been through the evict_inode
path is now found to be locked for read(2). The rwsem is in the
init_always path of the allocator, so it seems like it's getting reused
after evict_inode. Dave?
>
> ---
> fs/xfs/linux-2.6/xfs_super.c | 4 ++++
> 1 files changed, 4 insertions(+), 0 deletions(-)
>
> diff --git a/fs/xfs/linux-2.6/xfs_super.c b/fs/xfs/linux-2.6/xfs_super.c
> index 064f964..721c1c5 100644
> --- a/fs/xfs/linux-2.6/xfs_super.c
> +++ b/fs/xfs/linux-2.6/xfs_super.c
> @@ -1091,6 +1091,8 @@ xfs_fs_write_inode(
> return -error;
> }
>
> +static struct lock_class_key xfs_dead_inode;
> +
> STATIC void
> xfs_fs_evict_inode(
> struct inode *inode)
> @@ -1118,6 +1120,8 @@ xfs_fs_evict_inode(
> */
> ASSERT(!rwsem_is_locked(&ip->i_iolock.mr_lock));
> mrlock_init(&ip->i_iolock, MRLOCK_BARRIER, "xfsio", ip->i_ino);
> + lockdep_set_class_and_name(&ip->i_iolock->mr_lock, &xfs_dead_inode,
> + "xfd_dead_inode");
>
> xfs_inactive(ip);
> }
With this change, I assume the mrlock_init can go? (it would be nice
to have a wrapper to allocate the class by itself)
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2010-11-25 7:06 UTC|newest]
Thread overview: 14+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-11-23 12:18 XFS reclaim lock order bug Nick Piggin
2010-11-23 21:12 ` Dave Chinner
2010-11-24 0:58 ` Nick Piggin
2010-11-24 2:26 ` Dave Chinner
2010-11-24 20:03 ` Christoph Hellwig
2010-11-25 3:48 ` Nick Piggin
2010-11-25 6:25 ` Peter Zijlstra
2010-11-25 7:08 ` Nick Piggin [this message]
2010-11-25 7:28 ` Peter Zijlstra
2010-11-25 10:32 ` Dave Chinner
2010-11-25 10:29 ` Dave Chinner
2010-11-25 10:36 ` Peter Zijlstra
2010-11-25 11:25 ` Christoph Hellwig
2010-11-25 11:37 ` Peter Zijlstra
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20101125070817.GA4120@amd \
--to=npiggin@kernel.dk \
--cc=hch@infradead.org \
--cc=mingo@redhat.com \
--cc=peterz@infradead.org \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox