linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* Processes hung in "D" state in ext4, mm, md and dmcrypt
@ 2023-07-26 10:02 David Howells
  2023-07-26 15:29 ` [dm-devel] " Ming Lei
  2023-07-27  2:38 ` Yu Kuai
  0 siblings, 2 replies; 5+ messages in thread
From: David Howells @ 2023-07-26 10:02 UTC (permalink / raw)
  To: Theodore Ts'o, Andrew Morton, Song Liu, Christoph Hellwig,
	Alasdair Kergon
  Cc: dhowells, linux-ext4, linux-mm, linux-raid, dm-devel,
	linux-kernel

Hi,

With 6.5-rc2 (6.5.0-0.rc2.20230721gitf7e3a1bafdea.20.fc39.x86_64), I'm seeing
a bunch of processes getting stuck in the D state on my desktop after a few
hours of reading email and compiling stuff.  It's happened every day this week
so far and I managed to grab stack traces of the stuck processes this morning
(see attached).

There are two blockdevs involved below, /dev/md2 and /dev/md3.  md3 is a raid1
array with two partitions with an ext4 partition on it.  md2 is similar but
it's dm-crypted and ext4 is on top of that.

David
---

   1015 ?        D      0:04 [md2_raid1]
   1074 ?        D      0:00 [jbd2/sda6-8]
   1138 ?        D      0:00 [jbd2/md3-8]
   1167 ?        D      0:10 [dmcrypt_write/253:0]
   1202 ?        D      0:03 [jbd2/dm-0-8]
 117547 ?        D      5:12 [kworker/u16:8+flush-9:3]
 121540 ?        D      0:46 [kworker/u16:10+flush-253:0]
 125431 pts/2    Dl+    0:00 emacs .stgit-edit.txt
 125469 ?        D      0:00 /usr/libexec/nmh/rcvstore +kernel

===1015===
    PID TTY      STAT   TIME COMMAND
   1015 ?        D      0:04 [md2_raid1]
[<0>] md_super_wait+0xa2/0xe0
[<0>] md_bitmap_daemon_work+0x183/0x3b0
[<0>] md_check_recovery+0x42/0x5a0
[<0>] raid1d+0x87/0x16f0 [raid1]
[<0>] md_thread+0xab/0x190
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===1074===
    PID TTY      STAT   TIME COMMAND
   1074 ?        D      0:00 [jbd2/sda6-8]
[<0>] jbd2_journal_commit_transaction+0x11a6/0x1a20
[<0>] kjournald2+0xad/0x280
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===1138===
    PID TTY      STAT   TIME COMMAND
   1138 ?        D      0:00 [jbd2/md3-8]
[<0>] jbd2_journal_commit_transaction+0x162d/0x1a20
[<0>] kjournald2+0xad/0x280
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===1167===
    PID TTY      STAT   TIME COMMAND
   1167 ?        D      0:10 [dmcrypt_write/253:0]
[<0>] md_super_wait+0xa2/0xe0
[<0>] md_bitmap_unplug+0xad/0x120
[<0>] flush_bio_list+0xf3/0x100 [raid1]
[<0>] raid1_unplug+0x3b/0xb0 [raid1]
[<0>] __blk_flush_plug+0xd8/0x160
[<0>] blk_finish_plug+0x29/0x40
[<0>] dmcrypt_write+0x132/0x140 [dm_crypt]
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===1202===
    PID TTY      STAT   TIME COMMAND
   1202 ?        D      0:03 [jbd2/dm-0-8]
[<0>] jbd2_journal_commit_transaction+0x162d/0x1a20
[<0>] kjournald2+0xad/0x280
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===117547===
    PID TTY      STAT   TIME COMMAND
 117547 ?        D      5:12 [kworker/u16:8+flush-9:3]
[<0>] blk_mq_get_tag+0x11e/0x2b0
[<0>] __blk_mq_alloc_requests+0x1bc/0x350
[<0>] blk_mq_submit_bio+0x2c7/0x680
[<0>] __submit_bio+0x8b/0x170
[<0>] submit_bio_noacct_nocheck+0x159/0x370
[<0>] __block_write_full_folio+0x1e1/0x400
[<0>] writepage_cb+0x1a/0x70
[<0>] write_cache_pages+0x144/0x3b0
[<0>] do_writepages+0x164/0x1e0
[<0>] __writeback_single_inode+0x3d/0x360
[<0>] writeback_sb_inodes+0x1ed/0x4b0
[<0>] __writeback_inodes_wb+0x4c/0xf0
[<0>] wb_writeback+0x298/0x310
[<0>] wb_workfn+0x35b/0x510
[<0>] process_one_work+0x1de/0x3f0
[<0>] worker_thread+0x51/0x390
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===121540===
    PID TTY      STAT   TIME COMMAND
 121540 ?        D      0:46 [kworker/u16:10+flush-253:0]
[<0>] folio_wait_bit_common+0x13d/0x350
[<0>] mpage_prepare_extent_to_map+0x309/0x4d0
[<0>] ext4_do_writepages+0x25d/0xc90
[<0>] ext4_writepages+0xad/0x180
[<0>] do_writepages+0xcf/0x1e0
[<0>] __writeback_single_inode+0x3d/0x360
[<0>] writeback_sb_inodes+0x1ed/0x4b0
[<0>] __writeback_inodes_wb+0x4c/0xf0
[<0>] wb_writeback+0x298/0x310
[<0>] wb_workfn+0x35b/0x510
[<0>] process_one_work+0x1de/0x3f0
[<0>] worker_thread+0x51/0x390
[<0>] kthread+0xe5/0x120
[<0>] ret_from_fork+0x31/0x50
[<0>] ret_from_fork_asm+0x1b/0x30

===125431===
    PID TTY      STAT   TIME COMMAND
 125431 pts/2    Dl+    0:00 emacs .stgit-edit.txt
[<0>] jbd2_log_wait_commit+0xd8/0x140
[<0>] ext4_sync_file+0x1cc/0x380
[<0>] __x64_sys_fsync+0x3b/0x70
[<0>] do_syscall_64+0x5d/0x90
[<0>] entry_SYSCALL_64_after_hwframe+0x6e/0xd8

===125469===
    PID TTY      STAT   TIME COMMAND
 125469 ?        D      0:00 /usr/libexec/nmh/rcvstore +kernel
[<0>] folio_wait_bit_common+0x13d/0x350
[<0>] folio_wait_writeback+0x2c/0x90
[<0>] truncate_inode_partial_folio+0x5e/0x1a0
[<0>] truncate_inode_pages_range+0x1da/0x400
[<0>] truncate_pagecache+0x47/0x60
[<0>] ext4_setattr+0x685/0xba0
[<0>] notify_change+0x1e0/0x4a0
[<0>] do_truncate+0x98/0xf0
[<0>] do_sys_ftruncate+0x15c/0x1b0
[<0>] do_syscall_64+0x5d/0x90
[<0>] entry_SYSCALL_64_after_hwframe+0x6e/0xd8


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dm-devel] Processes hung in "D" state in ext4, mm, md and dmcrypt
  2023-07-26 10:02 Processes hung in "D" state in ext4, mm, md and dmcrypt David Howells
@ 2023-07-26 15:29 ` Ming Lei
  2023-07-26 19:30   ` Andrew Morton
  2023-07-27  2:38 ` Yu Kuai
  1 sibling, 1 reply; 5+ messages in thread
From: Ming Lei @ 2023-07-26 15:29 UTC (permalink / raw)
  To: David Howells, linux-block
  Cc: Theodore Ts'o, Andrew Morton, Song Liu, Christoph Hellwig,
	Alasdair Kergon, linux-raid, linux-kernel, linux-mm, dm-devel,
	linux-ext4, Ming Lei

On Wed, Jul 26, 2023 at 6:02 PM David Howells <dhowells@redhat.com> wrote:
>
> Hi,
>
> With 6.5-rc2 (6.5.0-0.rc2.20230721gitf7e3a1bafdea.20.fc39.x86_64), I'm seeing
> a bunch of processes getting stuck in the D state on my desktop after a few
> hours of reading email and compiling stuff.  It's happened every day this week
> so far and I managed to grab stack traces of the stuck processes this morning
> (see attached).
>
> There are two blockdevs involved below, /dev/md2 and /dev/md3.  md3 is a raid1
> array with two partitions with an ext4 partition on it.  md2 is similar but
> it's dm-crypted and ext4 is on top of that.
>
...

> ===117547===
>     PID TTY      STAT   TIME COMMAND
>  117547 ?        D      5:12 [kworker/u16:8+flush-9:3]
> [<0>] blk_mq_get_tag+0x11e/0x2b0
> [<0>] __blk_mq_alloc_requests+0x1bc/0x350
> [<0>] blk_mq_submit_bio+0x2c7/0x680
> [<0>] __submit_bio+0x8b/0x170
> [<0>] submit_bio_noacct_nocheck+0x159/0x370
> [<0>] __block_write_full_folio+0x1e1/0x400
> [<0>] writepage_cb+0x1a/0x70
> [<0>] write_cache_pages+0x144/0x3b0
> [<0>] do_writepages+0x164/0x1e0
> [<0>] __writeback_single_inode+0x3d/0x360
> [<0>] writeback_sb_inodes+0x1ed/0x4b0
> [<0>] __writeback_inodes_wb+0x4c/0xf0
> [<0>] wb_writeback+0x298/0x310
> [<0>] wb_workfn+0x35b/0x510
> [<0>] process_one_work+0x1de/0x3f0
> [<0>] worker_thread+0x51/0x390
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30

BTW, -rc3 fixes one similar issue on the above code path, so please try -rc3.

106397376c03 sbitmap: fix batching wakeup

Thanks,
Ming Lei

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dm-devel] Processes hung in "D" state in ext4, mm, md and dmcrypt
  2023-07-26 15:29 ` [dm-devel] " Ming Lei
@ 2023-07-26 19:30   ` Andrew Morton
  0 siblings, 0 replies; 5+ messages in thread
From: Andrew Morton @ 2023-07-26 19:30 UTC (permalink / raw)
  To: Ming Lei
  Cc: David Howells, linux-block, Theodore Ts'o, Song Liu,
	Christoph Hellwig, Alasdair Kergon, linux-raid, linux-kernel,
	linux-mm, dm-devel, linux-ext4, Ming Lei

On Wed, 26 Jul 2023 23:29:51 +0800 Ming Lei <tom.leiming@gmail.com> wrote:

> On Wed, Jul 26, 2023 at 6:02 PM David Howells <dhowells@redhat.com> wrote:
> >
> > Hi,
> >
> > With 6.5-rc2 (6.5.0-0.rc2.20230721gitf7e3a1bafdea.20.fc39.x86_64), I'm seeing
> > a bunch of processes getting stuck in the D state on my desktop after a few
> > hours of reading email and compiling stuff.  It's happened every day this week
> > so far and I managed to grab stack traces of the stuck processes this morning
> > (see attached).
> >
> > There are two blockdevs involved below, /dev/md2 and /dev/md3.  md3 is a raid1
> > array with two partitions with an ext4 partition on it.  md2 is similar but
> > it's dm-crypted and ext4 is on top of that.
> >
> ...
> 
> > ===117547===
> >     PID TTY      STAT   TIME COMMAND
> >  117547 ?        D      5:12 [kworker/u16:8+flush-9:3]
> > [<0>] blk_mq_get_tag+0x11e/0x2b0
> > [<0>] __blk_mq_alloc_requests+0x1bc/0x350
> > [<0>] blk_mq_submit_bio+0x2c7/0x680
> > [<0>] __submit_bio+0x8b/0x170
> > [<0>] submit_bio_noacct_nocheck+0x159/0x370
> > [<0>] __block_write_full_folio+0x1e1/0x400
> > [<0>] writepage_cb+0x1a/0x70
> > [<0>] write_cache_pages+0x144/0x3b0
> > [<0>] do_writepages+0x164/0x1e0
> > [<0>] __writeback_single_inode+0x3d/0x360
> > [<0>] writeback_sb_inodes+0x1ed/0x4b0
> > [<0>] __writeback_inodes_wb+0x4c/0xf0
> > [<0>] wb_writeback+0x298/0x310
> > [<0>] wb_workfn+0x35b/0x510
> > [<0>] process_one_work+0x1de/0x3f0
> > [<0>] worker_thread+0x51/0x390
> > [<0>] kthread+0xe5/0x120
> > [<0>] ret_from_fork+0x31/0x50
> > [<0>] ret_from_fork_asm+0x1b/0x30
> 
> BTW, -rc3 fixes one similar issue on the above code path, so please try -rc3.
> 
> 106397376c03 sbitmap: fix batching wakeup

That patch really needs a Fixes:, please.  And consideration for a
-stable backport.

Looking at what has changed recently in sbitmap, it seems unlikely that
106397376c03 fixes an issue that just appeared in 6.5-rcX.  But maybe
the issue you have identified has recently become easier to hit; we'll
see.


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dm-devel] Processes hung in "D" state in ext4, mm, md and dmcrypt
  2023-07-26 10:02 Processes hung in "D" state in ext4, mm, md and dmcrypt David Howells
  2023-07-26 15:29 ` [dm-devel] " Ming Lei
@ 2023-07-27  2:38 ` Yu Kuai
  2023-08-09 18:28   ` [dm-devel] Processes hung in "D" state in ext4, mm, md and dmcrypt (lockdep output) Ben Greear
  1 sibling, 1 reply; 5+ messages in thread
From: Yu Kuai @ 2023-07-27  2:38 UTC (permalink / raw)
  To: David Howells, Theodore Ts'o, Andrew Morton, Song Liu,
	Christoph Hellwig, Alasdair Kergon
  Cc: linux-raid, linux-kernel, linux-mm, dm-devel, linux-ext4,
	yukuai (C)

Hi,

在 2023/07/26 18:02, David Howells 写道:
> Hi,
> 
> With 6.5-rc2 (6.5.0-0.rc2.20230721gitf7e3a1bafdea.20.fc39.x86_64), I'm seeing
> a bunch of processes getting stuck in the D state on my desktop after a few
> hours of reading email and compiling stuff.  It's happened every day this week
> so far and I managed to grab stack traces of the stuck processes this morning
> (see attached).
> 
> There are two blockdevs involved below, /dev/md2 and /dev/md3.  md3 is a raid1
> array with two partitions with an ext4 partition on it.  md2 is similar but
> it's dm-crypted and ext4 is on top of that.
> 
> David
> ---
> 
>     1015 ?        D      0:04 [md2_raid1]
>     1074 ?        D      0:00 [jbd2/sda6-8]
>     1138 ?        D      0:00 [jbd2/md3-8]
>     1167 ?        D      0:10 [dmcrypt_write/253:0]
>     1202 ?        D      0:03 [jbd2/dm-0-8]
>   117547 ?        D      5:12 [kworker/u16:8+flush-9:3]
>   121540 ?        D      0:46 [kworker/u16:10+flush-253:0]
>   125431 pts/2    Dl+    0:00 emacs .stgit-edit.txt
>   125469 ?        D      0:00 /usr/libexec/nmh/rcvstore +kernel
> 
> ===1015===
>      PID TTY      STAT   TIME COMMAND
>     1015 ?        D      0:04 [md2_raid1]
> [<0>] md_super_wait+0xa2/0xe0
> [<0>] md_bitmap_daemon_work+0x183/0x3b0
> [<0>] md_check_recovery+0x42/0x5a0
> [<0>] raid1d+0x87/0x16f0 [raid1]
> [<0>] md_thread+0xab/0x190
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30

This means either the io to write super_block is stuck in underlying
disks or writing super_block is broken, I think it's probably the
former case. You'll need to locate where this io is now. If it can
be sure that there is no io in underlying disks, then this problem
is related to raid.
> 
> ===1074===
>      PID TTY      STAT   TIME COMMAND
>     1074 ?        D      0:00 [jbd2/sda6-8]
> [<0>] jbd2_journal_commit_transaction+0x11a6/0x1a20
> [<0>] kjournald2+0xad/0x280
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30
> 
> ===1138===
>      PID TTY      STAT   TIME COMMAND
>     1138 ?        D      0:00 [jbd2/md3-8]
> [<0>] jbd2_journal_commit_transaction+0x162d/0x1a20
> [<0>] kjournald2+0xad/0x280
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30
> 
> ===1167===
>      PID TTY      STAT   TIME COMMAND
>     1167 ?        D      0:10 [dmcrypt_write/253:0]
> [<0>] md_super_wait+0xa2/0xe0
> [<0>] md_bitmap_unplug+0xad/0x120
> [<0>] flush_bio_list+0xf3/0x100 [raid1]
> [<0>] raid1_unplug+0x3b/0xb0 [raid1]
> [<0>] __blk_flush_plug+0xd8/0x160
> [<0>] blk_finish_plug+0x29/0x40
> [<0>] dmcrypt_write+0x132/0x140 [dm_crypt]
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30
> 
> ===1202===
>      PID TTY      STAT   TIME COMMAND
>     1202 ?        D      0:03 [jbd2/dm-0-8]
> [<0>] jbd2_journal_commit_transaction+0x162d/0x1a20
> [<0>] kjournald2+0xad/0x280
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30
> 
> ===117547===
>      PID TTY      STAT   TIME COMMAND
>   117547 ?        D      5:12 [kworker/u16:8+flush-9:3]
> [<0>] blk_mq_get_tag+0x11e/0x2b0

Is this one of raid underlying disks? If so, this looks like io is stuck
in underlying disks.

Thanks,
Kuai

> [<0>] __blk_mq_alloc_requests+0x1bc/0x350
> [<0>] blk_mq_submit_bio+0x2c7/0x680
> [<0>] __submit_bio+0x8b/0x170
> [<0>] submit_bio_noacct_nocheck+0x159/0x370
> [<0>] __block_write_full_folio+0x1e1/0x400
> [<0>] writepage_cb+0x1a/0x70
> [<0>] write_cache_pages+0x144/0x3b0
> [<0>] do_writepages+0x164/0x1e0
> [<0>] __writeback_single_inode+0x3d/0x360
> [<0>] writeback_sb_inodes+0x1ed/0x4b0
> [<0>] __writeback_inodes_wb+0x4c/0xf0
> [<0>] wb_writeback+0x298/0x310
> [<0>] wb_workfn+0x35b/0x510
> [<0>] process_one_work+0x1de/0x3f0
> [<0>] worker_thread+0x51/0x390
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30
> 
> ===121540===
>      PID TTY      STAT   TIME COMMAND
>   121540 ?        D      0:46 [kworker/u16:10+flush-253:0]
> [<0>] folio_wait_bit_common+0x13d/0x350
> [<0>] mpage_prepare_extent_to_map+0x309/0x4d0
> [<0>] ext4_do_writepages+0x25d/0xc90
> [<0>] ext4_writepages+0xad/0x180
> [<0>] do_writepages+0xcf/0x1e0
> [<0>] __writeback_single_inode+0x3d/0x360
> [<0>] writeback_sb_inodes+0x1ed/0x4b0
> [<0>] __writeback_inodes_wb+0x4c/0xf0
> [<0>] wb_writeback+0x298/0x310
> [<0>] wb_workfn+0x35b/0x510
> [<0>] process_one_work+0x1de/0x3f0
> [<0>] worker_thread+0x51/0x390
> [<0>] kthread+0xe5/0x120
> [<0>] ret_from_fork+0x31/0x50
> [<0>] ret_from_fork_asm+0x1b/0x30
> 
> ===125431===
>      PID TTY      STAT   TIME COMMAND
>   125431 pts/2    Dl+    0:00 emacs .stgit-edit.txt
> [<0>] jbd2_log_wait_commit+0xd8/0x140
> [<0>] ext4_sync_file+0x1cc/0x380
> [<0>] __x64_sys_fsync+0x3b/0x70
> [<0>] do_syscall_64+0x5d/0x90
> [<0>] entry_SYSCALL_64_after_hwframe+0x6e/0xd8
> 
> ===125469===
>      PID TTY      STAT   TIME COMMAND
>   125469 ?        D      0:00 /usr/libexec/nmh/rcvstore +kernel
> [<0>] folio_wait_bit_common+0x13d/0x350
> [<0>] folio_wait_writeback+0x2c/0x90
> [<0>] truncate_inode_partial_folio+0x5e/0x1a0
> [<0>] truncate_inode_pages_range+0x1da/0x400
> [<0>] truncate_pagecache+0x47/0x60
> [<0>] ext4_setattr+0x685/0xba0
> [<0>] notify_change+0x1e0/0x4a0
> [<0>] do_truncate+0x98/0xf0
> [<0>] do_sys_ftruncate+0x15c/0x1b0
> [<0>] do_syscall_64+0x5d/0x90
> [<0>] entry_SYSCALL_64_after_hwframe+0x6e/0xd8
> --
> dm-devel mailing list
> dm-devel@redhat.com
> https://listman.redhat.com/mailman/listinfo/dm-devel
> 
> .
> 


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dm-devel] Processes hung in "D" state in ext4, mm, md and dmcrypt (lockdep output)
  2023-07-27  2:38 ` Yu Kuai
@ 2023-08-09 18:28   ` Ben Greear
  0 siblings, 0 replies; 5+ messages in thread
From: Ben Greear @ 2023-08-09 18:28 UTC (permalink / raw)
  To: Yu Kuai, David Howells, Theodore Ts'o, Andrew Morton,
	Song Liu, Christoph Hellwig, Alasdair Kergon
  Cc: linux-raid, linux-kernel, linux-mm, dm-devel, linux-ext4,
	yukuai (C)

On 7/26/23 19:38, Yu Kuai wrote:
> Hi,
> 
> 在 2023/07/26 18:02, David Howells 写道:
>> Hi,
>>
>> With 6.5-rc2 (6.5.0-0.rc2.20230721gitf7e3a1bafdea.20.fc39.x86_64), I'm seeing
>> a bunch of processes getting stuck in the D state on my desktop after a few
>> hours of reading email and compiling stuff.  It's happened every day this week
>> so far and I managed to grab stack traces of the stuck processes this morning
>> (see attached).
>>
>> There are two blockdevs involved below, /dev/md2 and /dev/md3.  md3 is a raid1
>> array with two partitions with an ext4 partition on it.  md2 is similar but
>> it's dm-crypted and ext4 is on top of that.

Hello,

I have been hitting deadlocks in 6.5 series, and I finally managed to catch one with
lockdep kernel and serial console active.

My kernels have a lot of wifi patches on top of mainline, but I do not see indication
in this backtrace that it is network related.  A quick search of the mailing list found
this thread, and it looks like it might be similar to what I see.  I am able to test
patches and try other debug options if someone has suggestions:

INFO: task mdX_raid1:488 blocked for more than 180 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:mdX_raid1       state:D stack:0     pid:488   ppid:2      flags:0x00004000
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  schedule+0x89/0x110
  md_super_wait+0x110/0x160
  ? md_super_write+0x1b0/0x1b0
  ? print_usage_bug.part.0+0x3d0/0x3d0
  ? __mutex_lock+0xd22/0x1200
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lock_acquire+0x15a/0x3b0
  ? md_bitmap_daemon_work+0x2fc/0x6a0
  ? find_held_lock+0x83/0xa0
  md_bitmap_wait_writes+0x15f/0x170
  ? behind_writes_used_reset+0x40/0x40
  ? mark_held_locks+0x65/0x90
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irq+0x1f/0x40
  md_bitmap_daemon_work+0x483/0x6a0
  ? md_bitmap_write_all+0xe0/0xe0
  ? lock_acquire+0x15a/0x3b0
  ? lock_timer_base+0x93/0xd0
  md_check_recovery+0x7b/0xa50
  raid1d+0xdb/0x30b0 [raid1]
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? raid1_read_request+0x13b0/0x13b0 [raid1]
  ? finish_wait+0x4a/0xc0
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? md_thread+0x1fe/0x280
  ? reacquire_held_locks+0x270/0x270
  ? mark_held_locks+0x24/0x90
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  md_thread+0x175/0x280
  ? md_submit_discard_bio+0x210/0x210
  ? mark_held_locks+0x24/0x90
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? md_submit_discard_bio+0x210/0x210
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task jbd2/dm-4-8:526 blocked for more than 180 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:jbd2/dm-4-8     state:D stack:0     pid:526   ppid:2      flags:0x00004000
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? lock_sync+0xf0/0xf0
  ? find_held_lock+0x83/0xa0
  ? __wait_on_bit+0x95/0x100
  ? reacquire_held_locks+0x270/0x270
  ? submit_bio_noacct_nocheck+0x516/0x5a0
  schedule+0x89/0x110
  io_schedule+0x68/0xa0
  bit_wait_io+0x12/0x80
  __wait_on_bit+0x3a/0x100
  ? bit_wait+0x80/0x80
  out_of_line_wait_on_bit+0xd7/0x110
  ? __wait_on_bit+0x100/0x100
  ? cpuacct_css_alloc+0xd0/0xd0
  jbd2_journal_commit_transaction+0x2767/0x2ec0
  ? jbd2_journal_finish_inode_data_buffers+0x50/0x50
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? __timer_delete_sync+0x170/0x210
  ? reacquire_held_locks+0x270/0x270
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? timer_shutdown+0xe0/0xe0
  ? do_raw_read_unlock+0x80/0x80
  kjournald2+0x16c/0x560
  ? jbd2_write_superblock+0x450/0x450
  ? mark_held_locks+0x24/0x90
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? jbd2_write_superblock+0x450/0x450
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task journal-offline:190685 blocked for more than 180 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:journal-offline state:D stack:0     pid:190685 ppid:1      flags:0x00000002
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? folio_wait_bit_common+0x11e/0x470
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? reacquire_held_locks+0x270/0x270
  ? do_raw_spin_lock+0x114/0x1a0
  schedule+0x89/0x110
  io_schedule+0x68/0xa0
  folio_wait_bit_common+0x20b/0x470
  ? filemap_get_folios_contig+0x4e0/0x4e0
  ? filemap_invalidate_unlock_two+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  folio_wait_writeback+0x16/0x100
  __filemap_fdatawait_range+0xc1/0x150
  ? filemap_get_folios_tag+0x5c0/0x5c0
  ? find_held_lock+0x83/0xa0
  ? __filemap_fdatawrite_range+0xc6/0x120
  ? delete_from_page_cache_batch+0x450/0x450
  ? strscpy+0xc5/0x190
  file_write_and_wait_range+0x9b/0xc0
  ext4_sync_file+0x134/0x640
  __x64_sys_fsync+0x3e/0x60
  do_syscall_64+0x34/0xb0
  entry_SYSCALL_64_after_hwframe+0x46/0xb0
RIP: 0033:0x7f9088d0868a
RSP: 002b:00007f90869fd9f0 EFLAGS: 00000293 ORIG_RAX: 000000000000004a
RAX: ffffffffffffffda RBX: 000055efd35f5940 RCX: 00007f9088d0868a
RDX: 0000000000000000 RSI: 0000000000000000 RDI: 000000000000002a
RBP: 000055efd361a510 R08: 0000000000000000 R09: 00007fff65ac277f
R10: 00007f9088c8cc1e R11: 0000000000000293 R12: 000055efd2c527a0
R13: 0000000000000002 R14: 00007fff65ac2680 R15: 00007f90861fe000
  </TASK>
INFO: task jbd2/dm-14-8:958 blocked for more than 180 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:jbd2/dm-14-8    state:D stack:0     pid:958   ppid:2      flags:0x00004000
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? blk_mq_flush_plug_list+0x10/0x30
  ? __blk_flush_plug+0x1be/0x210
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  schedule+0x89/0x110
  md_write_start+0x45e/0x590
  ? md_seq_show+0xc80/0xc80
  ? mark_lock+0x10b/0xd30
  ? destroy_sched_domains_rcu+0x30/0x30
  ? __lock_acquire+0x910/0x3240
  raid1_make_request+0x17c/0x18d0 [raid1]
  ? __lock_acquire+0x910/0x3240
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  ? lock_acquire+0x15a/0x3b0
  ? md_handle_request+0x230/0x5e0
  ? find_held_lock+0x83/0xa0
  ? raid1_end_write_request+0x720/0x720 [raid1]
  ? lock_release+0x1d1/0x3b0
  ? md_handle_request+0x29b/0x5e0
  ? reacquire_held_locks+0x270/0x270
  ? lock_acquire+0x15a/0x3b0
  ? bio_associate_blkg_from_css+0x209/0x9d0
  ? find_held_lock+0x83/0xa0
  ? lock_is_held_type+0xa6/0x110
  md_handle_request+0x2dc/0x5e0
  ? mddev_init_writes_pending+0x60/0x60
  ? lock_release+0x1d1/0x3b0
  ? bio_associate_blkg+0x71/0x130
  ? reacquire_held_locks+0x270/0x270
  ? bio_associate_blkg_from_css+0x2e1/0x9d0
  raid_map+0x61/0x80 [dm_raid]
  __map_bio+0x85/0x290
  dm_submit_bio+0x4b9/0xca0
  ? dm_get_live_table+0xd0/0xd0
  ? reacquire_held_locks+0x270/0x270
  ? lock_release+0x1d1/0x3b0
  ? lock_is_held_type+0xa6/0x110
  __submit_bio+0x139/0x180
  ? __bio_queue_enter+0x350/0x350
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? seqcount_lockdep_reader_access.constprop.0+0x82/0x90
  ? seqcount_lockdep_reader_access.constprop.0+0x82/0x90
  ? ktime_get+0x1c/0xb0
  submit_bio_noacct_nocheck+0x2a1/0x5a0
  ? fscrypt_dio_supported+0xb0/0xb0
  ? blk_alloc_queue+0x310/0x310
  ? __probestub_jbd2_shrink_scan_enter+0x1/0x10
  ? submit_bio_noacct+0x310/0xa60
  jbd2_journal_commit_transaction+0x18ed/0x2ec0
  ? jbd2_journal_finish_inode_data_buffers+0x50/0x50
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? __timer_delete_sync+0x170/0x210
  ? reacquire_held_locks+0x270/0x270
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? timer_shutdown+0xe0/0xe0
  ? mark_held_locks+0x65/0x90
  kjournald2+0x16c/0x560
  ? jbd2_write_superblock+0x450/0x450
  ? mark_held_locks+0x24/0x90
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? jbd2_write_superblock+0x450/0x450
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task kworker/u16:0:145831 blocked for more than 181 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/u16:0   state:D stack:0     pid:145831 ppid:2      flags:0x00004000
Workqueue: writeback wb_workfn (flush-252:14)
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? blk_mq_flush_plug_list+0x10/0x30
  ? __blk_flush_plug+0x1be/0x210
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  schedule+0x89/0x110
  md_write_start+0x45e/0x590
  ? md_seq_show+0xc80/0xc80
  ? mark_lock+0x10b/0xd30
  ? __writeback_inodes_wb+0x6b/0x130
  ? destroy_sched_domains_rcu+0x30/0x30
  ? __lock_acquire+0x910/0x3240
  raid1_make_request+0x17c/0x18d0 [raid1]
  ? __lock_acquire+0x910/0x3240
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  ? lock_acquire+0x15a/0x3b0
  ? md_handle_request+0x230/0x5e0
  ? find_held_lock+0x83/0xa0
  ? raid1_end_write_request+0x720/0x720 [raid1]
  ? lock_release+0x1d1/0x3b0
  ? md_handle_request+0x29b/0x5e0
  ? reacquire_held_locks+0x270/0x270
  ? lock_acquire+0x15a/0x3b0
  ? bio_associate_blkg_from_css+0x209/0x9d0
  ? find_held_lock+0x83/0xa0
  ? lock_is_held_type+0xa6/0x110
  md_handle_request+0x2dc/0x5e0
  ? mddev_init_writes_pending+0x60/0x60
  ? lock_release+0x1d1/0x3b0
  ? bio_associate_blkg+0x71/0x130
  ? reacquire_held_locks+0x270/0x270
  ? bio_associate_blkg_from_css+0x2e1/0x9d0
  raid_map+0x61/0x80 [dm_raid]
  __map_bio+0x85/0x290
  dm_submit_bio+0x4b9/0xca0
  ? dm_get_live_table+0xd0/0xd0
  ? reacquire_held_locks+0x270/0x270
  ? lock_release+0x1d1/0x3b0
  ? lock_is_held_type+0xa6/0x110
  __submit_bio+0x139/0x180
  ? __bio_queue_enter+0x350/0x350
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? seqcount_lockdep_reader_access.constprop.0+0x82/0x90
  ? seqcount_lockdep_reader_access.constprop.0+0x82/0x90
  ? ktime_get+0x1c/0xb0
  submit_bio_noacct_nocheck+0x2a1/0x5a0
  ? fscrypt_set_bio_crypt_ctx+0x6c/0x120
  ? fscrypt_dio_supported+0xb0/0xb0
  ? blk_alloc_queue+0x310/0x310
  ? guard_bio_eod+0x51/0x310
  ? submit_bio_noacct+0x310/0xa60
  __block_write_full_folio+0x3ae/0x6d0
  ? bdev_statx_dioalign+0xe0/0xe0
  ? block_invalidate_folio+0x2d0/0x2d0
  writepage_cb+0x34/0xb0
  write_cache_pages+0x257/0x650
  ? dirty_background_bytes_handler+0x20/0x20
  ? folio_clear_dirty_for_io+0x1b0/0x1b0
  ? do_writepages+0x25e/0x300
  ? find_held_lock+0x83/0xa0
  ? reacquire_held_locks+0x270/0x270
  ? do_raw_spin_lock+0x114/0x1a0
  ? spin_bug+0xe0/0xe0
  do_writepages+0x192/0x300
  ? dirty_ratio_handler+0x50/0x50
  ? lock_release+0x1d1/0x3b0
  ? writeback_sb_inodes+0x2b2/0x7c0
  ? reacquire_held_locks+0x270/0x270
  ? do_raw_spin_lock+0x114/0x1a0
  __writeback_single_inode+0x8f/0x620
  writeback_sb_inodes+0x36d/0x7c0
  ? sync_inode_metadata+0xc0/0xc0
  ? down_read_trylock+0x16d/0x290
  ? trylock_super+0x11/0x60
  __writeback_inodes_wb+0x6b/0x130
  wb_writeback+0x447/0x540
  ? __writeback_inodes_wb+0x130/0x130
  ? mark_held_locks+0x65/0x90
  ? get_nr_inodes+0x5f/0xa0
  wb_workfn+0x5b4/0x7c0
  ? inode_wait_for_writeback+0x30/0x30
  ? lock_acquire+0x15a/0x3b0
  ? process_one_work+0x53f/0xa30
  ? lock_sync+0xf0/0xf0
  ? mark_held_locks+0x24/0x90
  process_one_work+0x622/0xa30
  ? lock_sync+0xf0/0xf0
  ? pwq_dec_nr_in_flight+0x100/0x100
  ? spin_bug+0xe0/0xe0
  ? mark_held_locks+0x24/0x90
  worker_thread+0x8a/0x6f0
  ? process_one_work+0xa30/0xa30
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task kworker/u16:3:174452 blocked for more than 181 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:kworker/u16:3   state:D stack:0     pid:174452 ppid:2      flags:0x00004000
Workqueue: writeback wb_workfn (flush-252:4)
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? blk_mq_flush_plug_list+0x10/0x30
  ? __blk_flush_plug+0x1be/0x210
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  schedule+0x89/0x110
  md_super_wait+0x110/0x160
  ? md_super_write+0x1b0/0x1b0
  ? destroy_sched_domains_rcu+0x30/0x30
  ? write_page+0x1a3/0x890
  md_bitmap_wait_writes+0x15f/0x170
  ? behind_writes_used_reset+0x40/0x40
  ? __percpu_counter_sum+0x16/0xe0
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? __percpu_counter_sum+0xca/0xe0
  md_bitmap_unplug+0x1c5/0x210
  flush_bio_list+0x19a/0x1a0 [raid1]
  raid1_unplug+0x63/0x130 [raid1]
  __blk_flush_plug+0x189/0x210
  ? blk_start_plug_nr_ios+0xd0/0xd0
  blk_finish_plug+0x37/0x50
  wb_writeback+0x48c/0x540
  ? __writeback_inodes_wb+0x130/0x130
  wb_workfn+0x51a/0x7c0
  ? inode_wait_for_writeback+0x30/0x30
  ? lock_acquire+0x15a/0x3b0
  ? process_one_work+0x53f/0xa30
  ? lock_sync+0xf0/0xf0
  ? mark_held_locks+0x24/0x90
  process_one_work+0x622/0xa30
  ? lock_sync+0xf0/0xf0
  ? pwq_dec_nr_in_flight+0x100/0x100
  ? spin_bug+0xe0/0xe0
  ? mark_held_locks+0x24/0x90
  worker_thread+0x8a/0x6f0
  ? process_one_work+0xa30/0xa30
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task btserver:188934 blocked for more than 181 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:btserver        state:D stack:0     pid:188934 ppid:188373 flags:0x00000002
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? lock_release+0x1d1/0x3b0
  ? mark_held_locks+0x65/0x90
  schedule+0x89/0x110
  io_schedule+0x68/0xa0
  bit_wait_io+0x12/0x80
  __wait_on_bit_lock+0xcf/0x120
  ? bit_wait+0x80/0x80
  out_of_line_wait_on_bit_lock+0xd7/0x110
  ? __wait_on_bit_lock+0x120/0x120
  ? cpuacct_css_alloc+0xd0/0xd0
  do_get_write_access+0x458/0x830
  jbd2_journal_get_write_access+0x76/0xa0
  __ext4_journal_get_write_access+0xea/0x210
  ext4_orphan_add+0x432/0x8b0
  ? stop_this_handle+0x210/0x210
  ? kasan_set_track+0x21/0x30
  ? __kasan_slab_alloc+0x62/0x70
  ? ext4_fc_destroy_dentry_cache+0x10/0x10
  ? kmem_cache_alloc+0x22e/0x290
  ? rcu_is_watching+0x1f/0x40
  ? jbd2__journal_start+0x22d/0x340
  ext4_truncate+0x43c/0x930
  ? ext4_punch_hole+0x7a0/0x7a0
  ? __do_fault+0x290/0x290
  ext4_setattr+0xc43/0x1100
  ? make_vfsgid+0x76/0xf0
  notify_change+0x306/0x6c0
  do_truncate+0xdc/0x160
  ? file_open_root+0x200/0x200
  ? ext4_file_write_iter+0xcd0/0xcd0
  path_openat+0xf37/0x1120
  ? path_lookupat+0x2e0/0x2e0
  ? ___slab_alloc+0x995/0x1300
  do_filp_open+0x151/0x270
  ? may_open_dev+0x50/0x50
  ? lock_is_held_type+0xa6/0x110
  ? reacquire_held_locks+0x270/0x270
  ? do_raw_spin_lock+0x114/0x1a0
  ? spin_bug+0xe0/0xe0
  ? do_raw_spin_unlock+0x8e/0xe0
  ? _raw_spin_unlock+0x1a/0x30
  ? alloc_fd+0x19b/0x320
  do_sys_openat2+0x103/0x130
  ? build_open_flags+0x240/0x240
  __x64_sys_openat+0xe0/0x130
  ? __ia32_compat_sys_open+0x120/0x120
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? syscall_enter_from_user_mode+0x1c/0x50
  do_syscall_64+0x34/0xb0
  entry_SYSCALL_64_after_hwframe+0x46/0xb0
RIP: 0033:0x7f63771018e8
RSP: 002b:00007fffd4942040 EFLAGS: 00000246 ORIG_RAX: 0000000000000101
RAX: ffffffffffffffda RBX: 0000000000000241 RCX: 00007f63771018e8
RDX: 0000000000000241 RSI: 0000000003c27440 RDI: 00000000ffffff9c
RBP: 0000000003c27440 R08: 0000000000000004 R09: 0000000000000001
R10: 00000000000001b6 R11: 0000000000000246 R12: 0000000003c27440
R13: 0000000000a316da R14: 0000000000000001 R15: 0000000000b0bbb0
  </TASK>

Showing all locks held in the system:
1 lock held by rcu_tasks_kthre/12:
  #0: ffffffff83840590 (rcu_tasks.tasks_gp_mutex){+.+.}-{4:4}, at: rcu_tasks_one_gp+0x32/0x690
1 lock held by rcu_tasks_trace/13:
  #0: ffffffff83840290 (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{4:4}, at: rcu_tasks_one_gp+0x32/0x690
1 lock held by khungtaskd/64:
  #0: ffffffff83841160 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x29/0x230
1 lock held by mdX_raid1/488:
  #0: ffff88810aeac5e8 (&mddev->bitmap_info.mutex){+.+.}-{4:4}, at: md_bitmap_daemon_work+0x83/0x6a0
1 lock held by mdX_raid1/913:
  #0: ffff888121ccc5e8 (&mddev->bitmap_info.mutex){+.+.}-{4:4}, at: md_bitmap_daemon_work+0x83/0x6a0
1 lock held by jbd2/dm-14-8/958:
  #0: ffff88810dffe680 (&md->io_barrier){.+.+}-{0:0}, at: dm_get_live_table+0x0/0xd0
1 lock held by fuse mainloop/2513:
  #0: ffff88811f35e468 (&pipe->mutex/1){+.+.}-{4:4}, at: splice_file_to_pipe+0x21/0xd0
1 lock held by fuse mainloop/2515:
  #0: ffff88811c744468 (&pipe->mutex/1){+.+.}-{4:4}, at: splice_file_to_pipe+0x21/0xd0
4 locks held by kworker/u16:0/145831:
  #0: ffff88810b797938 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #1: ffff88814efc7dc0 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #2: ffff88810a3080e0 (&type->s_umount_key#53){.+.+}-{4:4}, at: trylock_super+0x11/0x60
  #3: ffff88810dffe680 (&md->io_barrier){.+.+}-{0:0}, at: dm_get_live_table+0x0/0xd0
2 locks held by kworker/u16:3/174452:
  #0: ffff88810b797938 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #1: ffff88812bd77dc0 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
2 locks held by kworker/4:1/184807:
  #0: ffff88810dc6e138 ((wq_completion)md){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #1: ffff888145e5fdc0 ((work_completion)(&mddev->flush_work)){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
4 locks held by btserver/188934:
  #0: ffff88812f7ee400 (sb_writers#9){.+.+}-{0:0}, at: path_openat+0xd77/0x1120
  #1: ffff888147320400 (&sb->s_type->i_mutex_key#15){++++}-{4:4}, at: do_truncate+0xca/0x160
  #2: ffff888147320590 (mapping.invalidate_lock#2){++++}-{4:4}, at: ext4_setattr+0x671/0x1100
  #3: ffff88812f7ec980 (jbd2_handle){++++}-{0:0}, at: start_this_handle+0x91b/0xaf0

=============================================

INFO: task mdX_raid1:488 blocked for more than 364 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:mdX_raid1       state:D stack:0     pid:488   ppid:2      flags:0x00004000
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  schedule+0x89/0x110
  md_super_wait+0x110/0x160
  ? md_super_write+0x1b0/0x1b0
  ? print_usage_bug.part.0+0x3d0/0x3d0
  ? __mutex_lock+0xd22/0x1200
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lock_acquire+0x15a/0x3b0
  ? md_bitmap_daemon_work+0x2fc/0x6a0
  ? find_held_lock+0x83/0xa0
  md_bitmap_wait_writes+0x15f/0x170
  ? behind_writes_used_reset+0x40/0x40
  ? mark_held_locks+0x65/0x90
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irq+0x1f/0x40
  md_bitmap_daemon_work+0x483/0x6a0
  ? md_bitmap_write_all+0xe0/0xe0
  ? lock_acquire+0x15a/0x3b0
  ? lock_timer_base+0x93/0xd0
  md_check_recovery+0x7b/0xa50
  raid1d+0xdb/0x30b0 [raid1]
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? raid1_read_request+0x13b0/0x13b0 [raid1]
  ? finish_wait+0x4a/0xc0
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? md_thread+0x1fe/0x280
  ? reacquire_held_locks+0x270/0x270
  ? mark_held_locks+0x24/0x90
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  md_thread+0x175/0x280
  ? md_submit_discard_bio+0x210/0x210
  ? mark_held_locks+0x24/0x90
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? md_submit_discard_bio+0x210/0x210
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task jbd2/dm-4-8:526 blocked for more than 364 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:jbd2/dm-4-8     state:D stack:0     pid:526   ppid:2      flags:0x00004000
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? lock_sync+0xf0/0xf0
  ? find_held_lock+0x83/0xa0
  ? __wait_on_bit+0x95/0x100
  ? reacquire_held_locks+0x270/0x270
  ? submit_bio_noacct_nocheck+0x516/0x5a0
  schedule+0x89/0x110
  io_schedule+0x68/0xa0
  bit_wait_io+0x12/0x80
  __wait_on_bit+0x3a/0x100
  ? bit_wait+0x80/0x80
  out_of_line_wait_on_bit+0xd7/0x110
  ? __wait_on_bit+0x100/0x100
  ? cpuacct_css_alloc+0xd0/0xd0
  jbd2_journal_commit_transaction+0x2767/0x2ec0
  ? jbd2_journal_finish_inode_data_buffers+0x50/0x50
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? __timer_delete_sync+0x170/0x210
  ? reacquire_held_locks+0x270/0x270
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? timer_shutdown+0xe0/0xe0
  ? do_raw_read_unlock+0x80/0x80
  kjournald2+0x16c/0x560
  ? jbd2_write_superblock+0x450/0x450
  ? mark_held_locks+0x24/0x90
  ? destroy_sched_domains_rcu+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x132/0x200
  ? _raw_spin_unlock_irqrestore+0x2b/0x50
  ? jbd2_write_superblock+0x450/0x450
  kthread+0x173/0x1b0
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork+0x28/0x50
  ? kthread_complete_and_exit+0x20/0x20
  ret_from_fork_asm+0x11/0x20
  </TASK>
INFO: task journal-offline:190685 blocked for more than 365 seconds.
       Tainted: G        W          6.5.0-rc4+ #6
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
task:journal-offline state:D stack:0     pid:190685 ppid:1      flags:0x00000002
Call Trace:
  <TASK>
  __schedule+0x7aa/0x16e0
  ? io_schedule_timeout+0xb0/0xb0
  ? folio_wait_bit_common+0x11e/0x470
  ? find_held_lock+0x83/0xa0
  ? lock_release+0x1d1/0x3b0
  ? reacquire_held_locks+0x270/0x270
  ? do_raw_spin_lock+0x114/0x1a0
  schedule+0x89/0x110
  io_schedule+0x68/0xa0
  folio_wait_bit_common+0x20b/0x470
  ? filemap_get_folios_contig+0x4e0/0x4e0
  ? filemap_invalidate_unlock_two+0x30/0x30
  ? lockdep_hardirqs_on_prepare+0x200/0x200
  folio_wait_writeback+0x16/0x100
  __filemap_fdatawait_range+0xc1/0x150
  ? filemap_get_folios_tag+0x5c0/0x5c0
  ? find_held_lock+0x83/0xa0
  ? __filemap_fdatawrite_range+0xc6/0x120
  ? delete_from_page_cache_batch+0x450/0x450
  ? strscpy+0xc5/0x190
  file_write_and_wait_range+0x9b/0xc0
  ext4_sync_file+0x134/0x640
  __x64_sys_fsync+0x3e/0x60
  do_syscall_64+0x34/0xb0
  entry_SYSCALL_64_after_hwframe+0x46/0xb0
RIP: 0033:0x7f9088d0868a
RSP: 002b:00007f90869fd9f0 EFLAGS: 00000293 ORIG_RAX: 000000000000004a
RAX: ffffffffffffffda RBX: 000055efd35f5940 RCX: 00007f9088d0868a
RDX: 0000000000000000 RSI: 0000000000000000 RDI: 000000000000002a
RBP: 000055efd361a510 R08: 0000000000000000 R09: 00007fff65ac277f
R10: 00007f9088c8cc1e R11: 0000000000000293 R12: 000055efd2c527a0
R13: 0000000000000002 R14: 00007fff65ac2680 R15: 00007f90861fe000
  </TASK>
Future hung task reports are suppressed, see sysctl kernel.hung_task_warnings

Showing all locks held in the system:
1 lock held by rcu_tasks_kthre/12:
  #0: ffffffff83840590 (rcu_tasks.tasks_gp_mutex){+.+.}-{4:4}, at: rcu_tasks_one_gp+0x32/0x690
1 lock held by rcu_tasks_trace/13:
  #0: ffffffff83840290 (rcu_tasks_trace.tasks_gp_mutex){+.+.}-{4:4}, at: rcu_tasks_one_gp+0x32/0x690
1 lock held by khungtaskd/64:
  #0: ffffffff83841160 (rcu_read_lock){....}-{1:3}, at: debug_show_all_locks+0x29/0x230
1 lock held by mdX_raid1/488:
  #0: ffff88810aeac5e8 (&mddev->bitmap_info.mutex){+.+.}-{4:4}, at: md_bitmap_daemon_work+0x83/0x6a0
1 lock held by mdX_raid1/913:
  #0: ffff888121ccc5e8 (&mddev->bitmap_info.mutex){+.+.}-{4:4}, at: md_bitmap_daemon_work+0x83/0x6a0
1 lock held by jbd2/dm-14-8/958:
  #0: ffff88810dffe680 (&md->io_barrier){.+.+}-{0:0}, at: dm_get_live_table+0x0/0xd0
1 lock held by fuse mainloop/2513:
  #0: ffff88811f35e468 (&pipe->mutex/1){+.+.}-{4:4}, at: splice_file_to_pipe+0x21/0xd0
1 lock held by fuse mainloop/2515:
  #0: ffff88811c744468 (&pipe->mutex/1){+.+.}-{4:4}, at: splice_file_to_pipe+0x21/0xd0
4 locks held by kworker/u16:0/145831:
  #0: ffff88810b797938 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #1: ffff88814efc7dc0 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #2: ffff88810a3080e0 (&type->s_umount_key#53){.+.+}-{4:4}, at: trylock_super+0x11/0x60
  #3: ffff88810dffe680 (&md->io_barrier){.+.+}-{0:0}, at: dm_get_live_table+0x0/0xd0
2 locks held by kworker/u16:3/174452:
  #0: ffff88810b797938 ((wq_completion)writeback){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #1: ffff88812bd77dc0 ((work_completion)(&(&wb->dwork)->work)){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
2 locks held by kworker/4:1/184807:
  #0: ffff88810dc6e138 ((wq_completion)md){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
  #1: ffff888145e5fdc0 ((work_completion)(&mddev->flush_work)){+.+.}-{0:0}, at: process_one_work+0x53f/0xa30
4 locks held by btserver/188934:
  #0: ffff88812f7ee400 (sb_writers#9){.+.+}-{0:0}, at: path_openat+0xd77/0x1120
  #1: ffff888147320400 (&sb->s_type->i_mutex_key#15){++++}-{4:4}, at: do_truncate+0xca/0x160
  #2: ffff888147320590 (mapping.invalidate_lock#2){++++}-{4:4}, at: ext4_setattr+0x671/0x1100
  #3: ffff88812f7ec980 (jbd2_handle){++++}-{0:0}, at: start_this_handle+0x91b/0xaf0
3 locks held by grep/190700:
  #0: ffff888124a38848 (&f->f_pos_lock){+.+.}-{4:4}, at: __fdget_pos+0x40/0x50
  #1: ffff88812f7ee400 (sb_writers#9){.+.+}-{0:0}, at: filemap_read+0x5ab/0x5c0
  #2: ffff88812f7ec980 (jbd2_handle){++++}-{0:0}, at: start_this_handle+0x91b/0xaf0

=============================================

Thanks,
Ben

-- 
Ben Greear <greearb@candelatech.com>
Candela Technologies Inc  http://www.candelatech.com



^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2023-08-09 18:38 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2023-07-26 10:02 Processes hung in "D" state in ext4, mm, md and dmcrypt David Howells
2023-07-26 15:29 ` [dm-devel] " Ming Lei
2023-07-26 19:30   ` Andrew Morton
2023-07-27  2:38 ` Yu Kuai
2023-08-09 18:28   ` [dm-devel] Processes hung in "D" state in ext4, mm, md and dmcrypt (lockdep output) Ben Greear

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).