public inbox for linux-btrfs@vger.kernel.org
 help / color / mirror / Atom feed
* zbd/009 (btrfs on zone gaps) fail on current Linus tree
@ 2025-10-07  6:21 Christoph Hellwig
  2025-10-07 11:16 ` Johannes Thumshirn
  0 siblings, 1 reply; 15+ messages in thread
From: Christoph Hellwig @ 2025-10-07  6:21 UTC (permalink / raw)
  To: linux-btrfs

At least in my usual x86_64 test VM:

zbd/009 (test gap zone support with BTRFS)                  
[   21.102324] run blktests zbd/009 at 2025-10-07 06:10:44
[   21.251500] sd 3:0:0:0: [sda] Synchronizing SCSI cache
[   21.413634] scsi_debug:sdebug_driver_probe: scsi_debug: trim poll_queues to 0. poll_q/nr_hw )
[   21.414056] scsi host3: scsi_debug: version 0191 [20210520]
[   21.414056]   dev_size_mb=1024, opts=0x0, submit_queues=1, statistics=0
[   21.415307] scsi 3:0:0:0: Direct-Access-ZBC Linux    scsi_debug 0191 PQ: 0 ANSI: 7
[   21.416384] scsi 3:0:0:0: Power-on or device reset occurred
[   21.416981] sd 3:0:0:0: Attached scsi generic sg1 type 20
[   21.417533] sd 3:0:0:0: [sda] Host-managed zoned block device
[   21.418153] sd 3:0:0:0: [sda] 262144 4096-byte logical blocks: (1.07 GB/1.00 GiB)
[   21.418676] sd 3:0:0:0: [sda] Write Protect is off
[   21.419017] sd 3:0:0:0: [sda] Write cache: enabled, read cache: enabled, supports DPO and FUA
[   21.419685] sd 3:0:0:0: [sda] permanent stream count = 5
[   21.420158] sd 3:0:0:0: [sda] Preferred minimum I/O size 4096 bytes
[   21.420593] sd 3:0:0:0: [sda] Optimal transfer size 4194304 bytes
[   21.421261] sd 3:0:0:0: [sda] 256 zones of 1024 logical blocks
[   21.456700] sd 3:0:0:0: [sda] Attached SCSI disk
[   21.523845] BTRFS: device fsid 9bcd6f4c-db2e-44d7-8597-4eb5774c1460 devid 1 transid 6 /dev/s)
[   21.528211] BTRFS info (device sda): first mount of filesystem 9bcd6f4c-db2e-44d7-8597-4eb570
[   21.528623] BTRFS info (device sda): using crc32c (crc32c-lib) checksum algorithm
[   21.530206] BTRFS info (device sda): host-managed zoned block device /dev/sda, 256 zones of s
[   21.530663] BTRFS info (device sda): zoned mode enabled with zone size 4194304
[   21.532601] BTRFS info (device sda): checking UUID tree
[   21.532909] BTRFS info (device sda): enabling ssd optimizations
[   21.533145] BTRFS info (device sda): enabling free space tree

[  242.795457] INFO: task kworker/u8:4:859 blocked for more than 120 seconds.
[  242.796028]       Tainted: G                 N  6.17.0+ #4047
[  242.796426] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
[  242.796941] task:kworker/u8:4    state:D stack:0     pid:859 tgid:859   ppid:2      task_f0
[  242.797667] Workqueue: writeback wb_workfn (flush-btrfs-2)
[  242.798065] Call Trace:
[  242.798227]  <TASK>
[  242.798369]  __schedule+0x524/0xb60
[  242.798601]  schedule+0x29/0xe0
[  242.798804]  io_schedule+0x4b/0x70
[  242.799024]  folio_wait_bit_common+0x126/0x390
[  242.799300]  ? filemap_get_folios_tag+0x247/0x2a0
[  242.800054]  ? __pfx_wake_page_function+0x10/0x10
[  242.800354]  extent_write_cache_pages+0x5c6/0x9c0
[  242.800631]  ? stack_depot_save_flags+0x29/0x870
[  242.800904]  ? set_track_prepare+0x45/0x70
[  242.801145]  ? __kmalloc_noprof+0x3a7/0x4e0
[  242.801391]  ? virtqueue_add_sgs+0x308/0x720
[  242.801644]  ? virtblk_add_req+0x81/0xe0
[  242.801875]  ? virtblk_add_req_batch+0x4b/0x100
[  242.802140]  ? virtio_queue_rqs+0x133/0x180
[  242.802385]  ? blk_mq_dispatch_queue_requests+0x155/0x180
[  242.802697]  ? blk_mq_flush_plug_list+0x73/0x160
[  242.802967]  ? preempt_count_add+0x4d/0xb0
[  242.803210]  btrfs_writepages+0x70/0x120
[  242.803636]  do_writepages+0xc5/0x160
[  242.803870]  __writeback_single_inode+0x3c/0x330
[  242.804154]  writeback_sb_inodes+0x21a/0x4d0
[  242.804436]  __writeback_inodes_wb+0x47/0xe0
[  242.804758]  wb_writeback+0x19a/0x310
[  242.805029]  wb_workfn+0x348/0x440
[  242.805248]  process_one_work+0x169/0x320
[  242.805487]  worker_thread+0x246/0x390
[  242.805711]  ? _raw_spin_unlock_irqrestore+0x1d/0x40
[  242.806003]  ? __pfx_worker_thread+0x10/0x10
[  242.806253]  kthread+0x106/0x220
[  242.806454]  ? __pfx_kthread+0x10/0x10
[  242.806683]  ? __pfx_kthread+0x10/0x10
[  242.806915]  ret_from_fork+0x11d/0x160
[  242.807145]  ? __pfx_kthread+0x10/0x10
[  242.807646]  ret_from_fork_asm+0x1a/0x30
[  242.807904]  </TASK>


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-07  6:21 zbd/009 (btrfs on zone gaps) fail on current Linus tree Christoph Hellwig
@ 2025-10-07 11:16 ` Johannes Thumshirn
  2025-10-08  6:02   ` hch
  0 siblings, 1 reply; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-07 11:16 UTC (permalink / raw)
  To: hch@infradead.org, linux-btrfs@vger.kernel.org

On 10/7/25 8:21 AM, Christoph Hellwig wrote:
> At least in my usual x86_64 test VM:
>
> zbd/009 (test gap zone support with BTRFS)
> [   21.102324] run blktests zbd/009 at 2025-10-07 06:10:44
> [   21.251500] sd 3:0:0:0: [sda] Synchronizing SCSI cache
> [   21.413634] scsi_debug:sdebug_driver_probe: scsi_debug: trim poll_queues to 0. poll_q/nr_hw )
> [   21.414056] scsi host3: scsi_debug: version 0191 [20210520]
> [   21.414056]   dev_size_mb=1024, opts=0x0, submit_queues=1, statistics=0
> [   21.415307] scsi 3:0:0:0: Direct-Access-ZBC Linux    scsi_debug 0191 PQ: 0 ANSI: 7
> [   21.416384] scsi 3:0:0:0: Power-on or device reset occurred
> [   21.416981] sd 3:0:0:0: Attached scsi generic sg1 type 20
> [   21.417533] sd 3:0:0:0: [sda] Host-managed zoned block device
> [   21.418153] sd 3:0:0:0: [sda] 262144 4096-byte logical blocks: (1.07 GB/1.00 GiB)
> [   21.418676] sd 3:0:0:0: [sda] Write Protect is off
> [   21.419017] sd 3:0:0:0: [sda] Write cache: enabled, read cache: enabled, supports DPO and FUA
> [   21.419685] sd 3:0:0:0: [sda] permanent stream count = 5
> [   21.420158] sd 3:0:0:0: [sda] Preferred minimum I/O size 4096 bytes
> [   21.420593] sd 3:0:0:0: [sda] Optimal transfer size 4194304 bytes
> [   21.421261] sd 3:0:0:0: [sda] 256 zones of 1024 logical blocks
> [   21.456700] sd 3:0:0:0: [sda] Attached SCSI disk
> [   21.523845] BTRFS: device fsid 9bcd6f4c-db2e-44d7-8597-4eb5774c1460 devid 1 transid 6 /dev/s)
> [   21.528211] BTRFS info (device sda): first mount of filesystem 9bcd6f4c-db2e-44d7-8597-4eb570
> [   21.528623] BTRFS info (device sda): using crc32c (crc32c-lib) checksum algorithm
> [   21.530206] BTRFS info (device sda): host-managed zoned block device /dev/sda, 256 zones of s
> [   21.530663] BTRFS info (device sda): zoned mode enabled with zone size 4194304
> [   21.532601] BTRFS info (device sda): checking UUID tree
> [   21.532909] BTRFS info (device sda): enabling ssd optimizations
> [   21.533145] BTRFS info (device sda): enabling free space tree
>
> [  242.795457] INFO: task kworker/u8:4:859 blocked for more than 120 seconds.
> [  242.796028]       Tainted: G                 N  6.17.0+ #4047
> [  242.796426] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> [  242.796941] task:kworker/u8:4    state:D stack:0     pid:859 tgid:859   ppid:2      task_f0
> [  242.797667] Workqueue: writeback wb_workfn (flush-btrfs-2)
> [  242.798065] Call Trace:
> [  242.798227]  <TASK>
> [  242.798369]  __schedule+0x524/0xb60
> [  242.798601]  schedule+0x29/0xe0
> [  242.798804]  io_schedule+0x4b/0x70
> [  242.799024]  folio_wait_bit_common+0x126/0x390
> [  242.799300]  ? filemap_get_folios_tag+0x247/0x2a0
> [  242.800054]  ? __pfx_wake_page_function+0x10/0x10
> [  242.800354]  extent_write_cache_pages+0x5c6/0x9c0
> [  242.800631]  ? stack_depot_save_flags+0x29/0x870
> [  242.800904]  ? set_track_prepare+0x45/0x70
> [  242.801145]  ? __kmalloc_noprof+0x3a7/0x4e0
> [  242.801391]  ? virtqueue_add_sgs+0x308/0x720
> [  242.801644]  ? virtblk_add_req+0x81/0xe0
> [  242.801875]  ? virtblk_add_req_batch+0x4b/0x100
> [  242.802140]  ? virtio_queue_rqs+0x133/0x180
> [  242.802385]  ? blk_mq_dispatch_queue_requests+0x155/0x180
> [  242.802697]  ? blk_mq_flush_plug_list+0x73/0x160
> [  242.802967]  ? preempt_count_add+0x4d/0xb0
> [  242.803210]  btrfs_writepages+0x70/0x120
> [  242.803636]  do_writepages+0xc5/0x160
> [  242.803870]  __writeback_single_inode+0x3c/0x330
> [  242.804154]  writeback_sb_inodes+0x21a/0x4d0
> [  242.804436]  __writeback_inodes_wb+0x47/0xe0
> [  242.804758]  wb_writeback+0x19a/0x310
> [  242.805029]  wb_workfn+0x348/0x440
> [  242.805248]  process_one_work+0x169/0x320
> [  242.805487]  worker_thread+0x246/0x390
> [  242.805711]  ? _raw_spin_unlock_irqrestore+0x1d/0x40
> [  242.806003]  ? __pfx_worker_thread+0x10/0x10
> [  242.806253]  kthread+0x106/0x220
> [  242.806454]  ? __pfx_kthread+0x10/0x10
> [  242.806683]  ? __pfx_kthread+0x10/0x10
> [  242.806915]  ret_from_fork+0x11d/0x160
> [  242.807145]  ? __pfx_kthread+0x10/0x10
> [  242.807646]  ret_from_fork_asm+0x1a/0x30
> [  242.807904]  </TASK>
>
>
>
hmm how reproducible is it on your side? I cannot reproduce it (yet)

zbd/009 (test gap zone support with BTRFS)  [passed]
     runtime    ...  12.516s

Will run the test in a loop (note I've tested both today's tree from 
Linus as well as btrfs/for-next from today)


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-07 11:16 ` Johannes Thumshirn
@ 2025-10-08  6:02   ` hch
  2025-10-08  6:06     ` Johannes Thumshirn
  0 siblings, 1 reply; 15+ messages in thread
From: hch @ 2025-10-08  6:02 UTC (permalink / raw)
  To: Johannes Thumshirn; +Cc: hch@infradead.org, linux-btrfs@vger.kernel.org

[-- Attachment #1: Type: text/plain, Size: 269 bytes --]

On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
> hmm how reproducible is it on your side? I cannot reproduce it (yet)

100% over about a dozen runs, a few of those including unrelated
patches.

My kernel .config and qemu command line are attached.

[-- Attachment #2: config.gz --]
[-- Type: application/gzip, Size: 26392 bytes --]

[-- Attachment #3: kvm.sh --]
[-- Type: application/x-sh, Size: 838 bytes --]

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-08  6:02   ` hch
@ 2025-10-08  6:06     ` Johannes Thumshirn
  2025-10-08 14:09       ` Johannes Thumshirn
  0 siblings, 1 reply; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-08  6:06 UTC (permalink / raw)
  To: hch@infradead.org; +Cc: linux-btrfs@vger.kernel.org

On 10/8/25 8:02 AM, hch@infradead.org wrote:
> On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
>> hmm how reproducible is it on your side? I cannot reproduce it (yet)
> 100% over about a dozen runs, a few of those including unrelated
> patches.
>
> My kernel .config and qemu command line are attached.
>
OK I'll give it a shot. For my config + qemu it survived 250 runs of 
zbd/009 yesterday without a hang :(


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-08  6:06     ` Johannes Thumshirn
@ 2025-10-08 14:09       ` Johannes Thumshirn
  2025-10-10  5:25         ` hch
  2025-10-16  6:59         ` Christoph Hellwig
  0 siblings, 2 replies; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-08 14:09 UTC (permalink / raw)
  To: hch@infradead.org; +Cc: linux-btrfs@vger.kernel.org

On 10/8/25 8:07 AM, Johannes Thumshirn wrote:
> On 10/8/25 8:02 AM, hch@infradead.org wrote:
>> On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
>>> hmm how reproducible is it on your side? I cannot reproduce it (yet)
>> 100% over about a dozen runs, a few of those including unrelated
>> patches.
>>
>> My kernel .config and qemu command line are attached.
>>
> OK I'll give it a shot. For my config + qemu it survived 250 runs of
> zbd/009 yesterday without a hang :(
>
>
Nope, even with your kconfig no success on recreating the bug.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-08 14:09       ` Johannes Thumshirn
@ 2025-10-10  5:25         ` hch
  2025-10-16  6:59         ` Christoph Hellwig
  1 sibling, 0 replies; 15+ messages in thread
From: hch @ 2025-10-10  5:25 UTC (permalink / raw)
  To: Johannes Thumshirn; +Cc: hch@infradead.org, linux-btrfs@vger.kernel.org

On Wed, Oct 08, 2025 at 02:09:00PM +0000, Johannes Thumshirn wrote:
> On 10/8/25 8:07 AM, Johannes Thumshirn wrote:
> > On 10/8/25 8:02 AM, hch@infradead.org wrote:
> >> On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
> >>> hmm how reproducible is it on your side? I cannot reproduce it (yet)
> >> 100% over about a dozen runs, a few of those including unrelated
> >> patches.
> >>
> >> My kernel .config and qemu command line are attached.
> >>
> > OK I'll give it a shot. For my config + qemu it survived 250 runs of
> > zbd/009 yesterday without a hang :(
> >
> >
> Nope, even with your kconfig no success on recreating the bug.

Weird.  I'll see if I can find some time to bisect it.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-08 14:09       ` Johannes Thumshirn
  2025-10-10  5:25         ` hch
@ 2025-10-16  6:59         ` Christoph Hellwig
  2025-10-17 13:33           ` Johannes Thumshirn
  1 sibling, 1 reply; 15+ messages in thread
From: Christoph Hellwig @ 2025-10-16  6:59 UTC (permalink / raw)
  To: Johannes Thumshirn; +Cc: Naohiro Aota, linux-btrfs

I've bisected the hang to:

commit 04147d8394e80acaaebf0365f112339e8b606c05 (HEAD)
Author: Naohiro Aota <naohiro.aota@wdc.com>
Date:   Wed Jul 16 16:59:55 2025 +0900

    btrfs: zoned: limit active zones to max_open_zones

with that patch zbd/009 hangs 100% for my config, and without it,
it works fine 100%.

On Wed, Oct 08, 2025 at 02:09:00PM +0000, Johannes Thumshirn wrote:
> On 10/8/25 8:07 AM, Johannes Thumshirn wrote:
> > On 10/8/25 8:02 AM, hch@infradead.org wrote:
> >> On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
> >>> hmm how reproducible is it on your side? I cannot reproduce it (yet)
> >> 100% over about a dozen runs, a few of those including unrelated
> >> patches.
> >>
> >> My kernel .config and qemu command line are attached.
> >>
> > OK I'll give it a shot. For my config + qemu it survived 250 runs of
> > zbd/009 yesterday without a hang :(
> >
> >
> Nope, even with your kconfig no success on recreating the bug.
> 
---end quoted text---

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-16  6:59         ` Christoph Hellwig
@ 2025-10-17 13:33           ` Johannes Thumshirn
  2025-10-20  6:47             ` hch
  2025-10-20  9:03             ` Naohiro Aota
  0 siblings, 2 replies; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-17 13:33 UTC (permalink / raw)
  To: hch@infradead.org; +Cc: Naohiro Aota, linux-btrfs@vger.kernel.org

On 10/16/25 8:59 AM, Christoph Hellwig wrote:
> I've bisected the hang to:
>
> commit 04147d8394e80acaaebf0365f112339e8b606c05 (HEAD)
> Author: Naohiro Aota <naohiro.aota@wdc.com>
> Date:   Wed Jul 16 16:59:55 2025 +0900
>
>      btrfs: zoned: limit active zones to max_open_zones
>
> with that patch zbd/009 hangs 100% for my config, and without it,
> it works fine 100%.

I still can't reproduce it. We seen a mount error as fallout of it 
though, can you check if you have 53de7ee4e28f ("btrfs: zoned: don't 
fail mount needlessly due to too many active zones")?

@Naohiro can you have a look if you can reproduce it?


Thanks,

     Johannes


> On Wed, Oct 08, 2025 at 02:09:00PM +0000, Johannes Thumshirn wrote:
>> On 10/8/25 8:07 AM, Johannes Thumshirn wrote:
>>> On 10/8/25 8:02 AM, hch@infradead.org wrote:
>>>> On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
>>>>> hmm how reproducible is it on your side? I cannot reproduce it (yet)
>>>> 100% over about a dozen runs, a few of those including unrelated
>>>> patches.
>>>>
>>>> My kernel .config and qemu command line are attached.
>>>>
>>> OK I'll give it a shot. For my config + qemu it survived 250 runs of
>>> zbd/009 yesterday without a hang :(
>>>
>>>
>> Nope, even with your kconfig no success on recreating the bug.
>>
> ---end quoted text---
>


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-17 13:33           ` Johannes Thumshirn
@ 2025-10-20  6:47             ` hch
  2025-10-20  6:52               ` Johannes Thumshirn
  2025-10-20  9:03             ` Naohiro Aota
  1 sibling, 1 reply; 15+ messages in thread
From: hch @ 2025-10-20  6:47 UTC (permalink / raw)
  To: Johannes Thumshirn
  Cc: hch@infradead.org, Naohiro Aota, linux-btrfs@vger.kernel.org

On Fri, Oct 17, 2025 at 01:33:08PM +0000, Johannes Thumshirn wrote:
> On 10/16/25 8:59 AM, Christoph Hellwig wrote:
> > I've bisected the hang to:
> >
> > commit 04147d8394e80acaaebf0365f112339e8b606c05 (HEAD)
> > Author: Naohiro Aota <naohiro.aota@wdc.com>
> > Date:   Wed Jul 16 16:59:55 2025 +0900
> >
> >      btrfs: zoned: limit active zones to max_open_zones
> >
> > with that patch zbd/009 hangs 100% for my config, and without it,
> > it works fine 100%.
> 
> I still can't reproduce it. We seen a mount error as fallout of it 
> though, can you check if you have 53de7ee4e28f ("btrfs: zoned: don't 
> fail mount needlessly due to too many active zones")?

Still hanging on -rc2 that has it.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-20  6:47             ` hch
@ 2025-10-20  6:52               ` Johannes Thumshirn
  2025-10-20  6:53                 ` hch
  0 siblings, 1 reply; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-20  6:52 UTC (permalink / raw)
  To: hch@infradead.org; +Cc: Naohiro Aota, linux-btrfs@vger.kernel.org

On 10/20/25 8:47 AM, hch@infradead.org wrote:
> On Fri, Oct 17, 2025 at 01:33:08PM +0000, Johannes Thumshirn wrote:
>> On 10/16/25 8:59 AM, Christoph Hellwig wrote:
>>> I've bisected the hang to:
>>>
>>> commit 04147d8394e80acaaebf0365f112339e8b606c05 (HEAD)
>>> Author: Naohiro Aota <naohiro.aota@wdc.com>
>>> Date:   Wed Jul 16 16:59:55 2025 +0900
>>>
>>>       btrfs: zoned: limit active zones to max_open_zones
>>>
>>> with that patch zbd/009 hangs 100% for my config, and without it,
>>> it works fine 100%.
>> I still can't reproduce it. We seen a mount error as fallout of it
>> though, can you check if you have 53de7ee4e28f ("btrfs: zoned: don't
>> fail mount needlessly due to too many active zones")?
> Still hanging on -rc2 that has it.
>
>
OK, maybe I tested wrong. Does it also hang if you only run zbd/009 or 
do you need to run the other zbd tests before?


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-20  6:52               ` Johannes Thumshirn
@ 2025-10-20  6:53                 ` hch
  2025-10-20  6:55                   ` Johannes Thumshirn
  0 siblings, 1 reply; 15+ messages in thread
From: hch @ 2025-10-20  6:53 UTC (permalink / raw)
  To: Johannes Thumshirn
  Cc: hch@infradead.org, Naohiro Aota, linux-btrfs@vger.kernel.org

On Mon, Oct 20, 2025 at 06:52:04AM +0000, Johannes Thumshirn wrote:
> OK, maybe I tested wrong. Does it also hang if you only run zbd/009 or 
> do you need to run the other zbd tests before?

Just running zbd/009 is fine.

Also make sure to try to use my kvm script as-is as it gets, given
the bisected commit I might be sensitive to the open zone limits or
zone append size or something like that.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-20  6:53                 ` hch
@ 2025-10-20  6:55                   ` Johannes Thumshirn
  2025-10-20  7:21                     ` hch
  0 siblings, 1 reply; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-20  6:55 UTC (permalink / raw)
  To: hch@infradead.org; +Cc: Naohiro Aota, linux-btrfs@vger.kernel.org

On 10/20/25 8:53 AM, hch@infradead.org wrote:
> On Mon, Oct 20, 2025 at 06:52:04AM +0000, Johannes Thumshirn wrote:
>> OK, maybe I tested wrong. Does it also hang if you only run zbd/009 or
>> do you need to run the other zbd tests before?
> Just running zbd/009 is fine.
>
> Also make sure to try to use my kvm script as-is as it gets, given
> the bisected commit I might be sensitive to the open zone limits or
> zone append size or something like that.

Yeah the issue on my side could be that I don't have virtio-blk devices 
attached to the VM as I don't have debian images lying aroudn for the 
rootfs (I'm using my hosts' root as a RO virtiofs mount).


I try to find something.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-20  6:55                   ` Johannes Thumshirn
@ 2025-10-20  7:21                     ` hch
  2025-10-20  7:43                       ` Johannes Thumshirn
  0 siblings, 1 reply; 15+ messages in thread
From: hch @ 2025-10-20  7:21 UTC (permalink / raw)
  To: Johannes Thumshirn
  Cc: hch@infradead.org, Naohiro Aota, linux-btrfs@vger.kernel.org

On Mon, Oct 20, 2025 at 06:55:46AM +0000, Johannes Thumshirn wrote:
> On 10/20/25 8:53 AM, hch@infradead.org wrote:
> > On Mon, Oct 20, 2025 at 06:52:04AM +0000, Johannes Thumshirn wrote:
> >> OK, maybe I tested wrong. Does it also hang if you only run zbd/009 or
> >> do you need to run the other zbd tests before?
> > Just running zbd/009 is fine.
> >
> > Also make sure to try to use my kvm script as-is as it gets, given
> > the bisected commit I might be sensitive to the open zone limits or
> > zone append size or something like that.
> 
> Yeah the issue on my side could be that I don't have virtio-blk devices 
> attached to the VM as I don't have debian images lying aroudn for the 
> rootfs (I'm using my hosts' root as a RO virtiofs mount).

Actually, the zbd tests emulate the zoned devices on top of the regular
ones, so it should not matter.  Here is my blktests config, btw:

TEST_DEVS=(/dev/vdb)
nvme_trtype=loop

so nothing special.

^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-20  7:21                     ` hch
@ 2025-10-20  7:43                       ` Johannes Thumshirn
  0 siblings, 0 replies; 15+ messages in thread
From: Johannes Thumshirn @ 2025-10-20  7:43 UTC (permalink / raw)
  To: hch@infradead.org; +Cc: Naohiro Aota, linux-btrfs@vger.kernel.org

On 10/20/25 9:22 AM, hch@infradead.org wrote:
> On Mon, Oct 20, 2025 at 06:55:46AM +0000, Johannes Thumshirn wrote:
>> On 10/20/25 8:53 AM, hch@infradead.org wrote:
>>> On Mon, Oct 20, 2025 at 06:52:04AM +0000, Johannes Thumshirn wrote:
>>>> OK, maybe I tested wrong. Does it also hang if you only run zbd/009 or
>>>> do you need to run the other zbd tests before?
>>> Just running zbd/009 is fine.
>>>
>>> Also make sure to try to use my kvm script as-is as it gets, given
>>> the bisected commit I might be sensitive to the open zone limits or
>>> zone append size or something like that.
>> Yeah the issue on my side could be that I don't have virtio-blk devices
>> attached to the VM as I don't have debian images lying aroudn for the
>> rootfs (I'm using my hosts' root as a RO virtiofs mount).
> Actually, the zbd tests emulate the zoned devices on top of the regular
> ones, so it should not matter.  Here is my blktests config, btw:
yes
> TEST_DEVS=(/dev/vdb)
> nvme_trtype=loop
>
> so nothing special.
>
and yes.


Can you generate me a vmcore and send me the vmlinux (with debug info) 
and core? Maybe I can find out what's going on that way.


^ permalink raw reply	[flat|nested] 15+ messages in thread

* Re: zbd/009 (btrfs on zone gaps) fail on current Linus tree
  2025-10-17 13:33           ` Johannes Thumshirn
  2025-10-20  6:47             ` hch
@ 2025-10-20  9:03             ` Naohiro Aota
  1 sibling, 0 replies; 15+ messages in thread
From: Naohiro Aota @ 2025-10-20  9:03 UTC (permalink / raw)
  To: Johannes Thumshirn, hch@infradead.org; +Cc: linux-btrfs@vger.kernel.org

On Fri Oct 17, 2025 at 10:33 PM JST, Johannes Thumshirn wrote:
> On 10/16/25 8:59 AM, Christoph Hellwig wrote:
>> I've bisected the hang to:
>>
>> commit 04147d8394e80acaaebf0365f112339e8b606c05 (HEAD)
>> Author: Naohiro Aota <naohiro.aota@wdc.com>
>> Date:   Wed Jul 16 16:59:55 2025 +0900
>>
>>      btrfs: zoned: limit active zones to max_open_zones
>>
>> with that patch zbd/009 hangs 100% for my config, and without it,
>> it works fine 100%.
>
> I still can't reproduce it. We seen a mount error as fallout of it 
> though, can you check if you have 53de7ee4e28f ("btrfs: zoned: don't 
> fail mount needlessly due to too many active zones")?
>
> @Naohiro can you have a look if you can reproduce it?

I'm running the test case 100 times on my virtme setup, and it all
passed so far.

>
>
> Thanks,
>
>      Johannes
>
>
>> On Wed, Oct 08, 2025 at 02:09:00PM +0000, Johannes Thumshirn wrote:
>>> On 10/8/25 8:07 AM, Johannes Thumshirn wrote:
>>>> On 10/8/25 8:02 AM, hch@infradead.org wrote:
>>>>> On Tue, Oct 07, 2025 at 11:16:08AM +0000, Johannes Thumshirn wrote:
>>>>>> hmm how reproducible is it on your side? I cannot reproduce it (yet)
>>>>> 100% over about a dozen runs, a few of those including unrelated
>>>>> patches.
>>>>>
>>>>> My kernel .config and qemu command line are attached.
>>>>>
>>>> OK I'll give it a shot. For my config + qemu it survived 250 runs of
>>>> zbd/009 yesterday without a hang :(
>>>>
>>>>
>>> Nope, even with your kconfig no success on recreating the bug.
>>>
>> ---end quoted text---
>>

^ permalink raw reply	[flat|nested] 15+ messages in thread

end of thread, other threads:[~2025-10-20  9:03 UTC | newest]

Thread overview: 15+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-10-07  6:21 zbd/009 (btrfs on zone gaps) fail on current Linus tree Christoph Hellwig
2025-10-07 11:16 ` Johannes Thumshirn
2025-10-08  6:02   ` hch
2025-10-08  6:06     ` Johannes Thumshirn
2025-10-08 14:09       ` Johannes Thumshirn
2025-10-10  5:25         ` hch
2025-10-16  6:59         ` Christoph Hellwig
2025-10-17 13:33           ` Johannes Thumshirn
2025-10-20  6:47             ` hch
2025-10-20  6:52               ` Johannes Thumshirn
2025-10-20  6:53                 ` hch
2025-10-20  6:55                   ` Johannes Thumshirn
2025-10-20  7:21                     ` hch
2025-10-20  7:43                       ` Johannes Thumshirn
2025-10-20  9:03             ` Naohiro Aota

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox