public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Dave Chinner <david@fromorbit.com>
To: Matthew Wilcox <willy@infradead.org>
Cc: linux-xfs@vger.kernel.org, bfoster@redhat.com
Subject: Re: fs-next-20251103 reclaim lockdep splat
Date: Thu, 6 Nov 2025 08:05:11 +1100	[thread overview]
Message-ID: <aQu8B63pEAzGRAkj@dread.disaster.area> (raw)
In-Reply-To: <aQux3yPwLFU42qof@casper.infradead.org>

On Wed, Nov 05, 2025 at 08:21:51PM +0000, Matthew Wilcox wrote:
> In trying to bisect the earlier reported transaction assertion failure,
> I hit this:
> 
> generic/476       run fstests generic/476 at 2025-11-05 20:16:46
> XFS (vdb): Mounting V5 Filesystem 7f483353-a0f6-4710-9adc-4b72f25598f8
> XFS (vdb): Ending clean mount
> XFS (vdc): Mounting V5 Filesystem 47fa2f49-e8e1-4622-a62c-53ea07b3d714
> XFS (vdc): Ending clean mount
> 
> ======================================================
> WARNING: possible circular locking dependency detected
> 6.18.0-rc4-ktest-00382-ga1e94de7fbd5 #116 Tainted: G        W
> ------------------------------------------------------
> kswapd0/111 is trying to acquire lock:
> ffff888102462418 (&xfs_nondir_ilock_class){++++}-{4:4}, at: xfs_ilock+0x14b/0x2b0
> 
> but task is already holding lock:
> ffffffff82710f00 (fs_reclaim){+.+.}-{0:0}, at: balance_pgdat+0x3e0/0x780
> 
> which lock already depends on the new lock.
> 
> 
> the existing dependency chain (in reverse order) is:
> 
> -> #1 (fs_reclaim){+.+.}-{0:0}:
>        fs_reclaim_acquire+0x67/0xa0
>        __kmalloc_cache_noprof+0x4d/0x500
>        iomap_fill_dirty_folios+0x6b/0xe0
>        xfs_buffered_write_iomap_begin+0xaee/0x1060
>        iomap_iter+0x1a1/0x4a0
>        iomap_zero_range+0xb0/0x420
>        xfs_zero_range+0x54/0x70
>        xfs_file_write_checks+0x21d/0x320
>        xfs_file_dio_write_unaligned+0x140/0x2b0
>        xfs_file_write_iter+0x22a/0x270
>        vfs_write+0x23f/0x540
>        ksys_write+0x6d/0x100
>        __x64_sys_write+0x1d/0x30
>        x64_sys_call+0x7d/0x1da0
>        do_syscall_64+0x6a/0x2e0
>        entry_SYSCALL_64_after_hwframe+0x76/0x7e
> 
> -> #0 (&xfs_nondir_ilock_class){++++}-{4:4}:
>        __lock_acquire+0x15be/0x27d0
>        lock_acquire+0xb2/0x290
>        down_write_nested+0x2a/0xb0
>        xfs_ilock+0x14b/0x2b0
>        xfs_icwalk_ag+0x517/0xaf0
>        xfs_icwalk+0x46/0x80
>        xfs_reclaim_inodes_nr+0x8c/0xb0
>        xfs_fs_free_cached_objects+0x1d/0x30
>        super_cache_scan+0x178/0x1d0
>        do_shrink_slab+0x16d/0x6a0
>        shrink_slab+0x4cf/0x8c0
>        shrink_node+0x334/0x870
>        balance_pgdat+0x35f/0x780

As I said on #xfs: false positive on the inode lock.

Reclaim is running in GFP_KERNEL context, so it's allowed to lock
unreferenced inodes.

The inodes that the allocation context holds locked are referenced
inodes, so it cannot self-deadlock on the inode locks it holds
because reclaim does not access or lock referenced inodes.

That being said, looking at this patch:

https://lore.kernel.org/linux-xfs/20251003134642.604736-4-bfoster@redhat.com/

I think the allocation that iomap_fill_dirty_folios() should
probably be using mapping_gfp_constraint(mapping, GFP_KERNEL) rather
than a hard coded GFP_KERNEL allocation. This is deep in the
buffered write path and the xfs ILOCK is held when
iomap_fill_dirty_folios() and it does folio lookups in that
context.

Huh - that kinda feels like a lock order violation. ILOCK is not
supposed to be held when we do page cache operations as the lock
order defined by writback operations is folio lookup -> folio lock
-> ILOCK.

So maybe this is a problem here, but not the one lockdep flagged...

Brian?

-Dave.
-- 
Dave Chinner
david@fromorbit.com

  reply	other threads:[~2025-11-05 21:05 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-11-05 20:21 fs-next-20251103 reclaim lockdep splat Matthew Wilcox
2025-11-05 21:05 ` Dave Chinner [this message]
2025-11-06 12:20   ` Brian Foster
2025-11-10 22:02     ` Dave Chinner
2025-11-11 14:07       ` Brian Foster

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=aQu8B63pEAzGRAkj@dread.disaster.area \
    --to=david@fromorbit.com \
    --cc=bfoster@redhat.com \
    --cc=linux-xfs@vger.kernel.org \
    --cc=willy@infradead.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox