From: DanglingPointer <danglingpointerexception@gmail.com>
To: "johnh1214@gmail.com" <johnh1214@gmail.com>, linux-btrfs@vger.kernel.org
Subject: Re: BTRFS RAID 5 array ran out of space, switched to R/O mode with errors
Date: Wed, 13 Aug 2025 09:28:53 +1000 [thread overview]
Message-ID: <1ffdaf61-e90b-40df-963a-224831f03163@gmail.com> (raw)
In-Reply-To: <b4efb68d-28d9-44cd-8190-6870d0dce098@gmail.com>
yes I do have a backup. There's a nightly backup via duplicacy backup.
The 8TB x 3 used to be 6TB x 3. I've been slowly upgrading them over
the last 4 years. However I never got close to 75% usage on them in
that time. I'll see when I can find time to upgrade the 6 to an 8.
Have you upgraded the kernel and btrfs-progs yet? Perhaps the newer
stuff fixes whatever caused the issue. That server of mine with the 75%
usage has 6.12.40 LTS kernel with btrfs-progs v6.6.3 from ubuntu 24.04
by default.
From what I understand, scrub is now in kernel so should be at the
version of the kernel and not btrfs-progs.
There's also btrfs.static which is a downloadable all in one utility
file tagged per version on kdave's github. Here's the counterpart for
my kernel version for example...
https://github.com/kdave/btrfs-progs/releases/tag/v6.12
thanks for your response!
On 13/8/25 06:23, johnh1214@gmail.com wrote:
> When I added an extra 8TB drive to the array, and upgraded from space
> cache v1 to v2, I had enough backup storage space to make a full copy
> (unfortunately, most of the full backup has been replaced with other
> data). The addition of an extra drive and space cache upgrade went
> smoothly, but it was only with a full backup available, that I was
> willing to give it a try. If I were you, I'd make sure I had a full
> backup of everything important, before adding on extra storage space.
>
> I knew better than to copy more data over to the RAID array running
> low, but that day, I was impatient, and thought it improbable
> something would go so horribly wrong.
>
> It's so ridiculous, I want to take a look at the source code to figure
> out how running low on storage space, can mess up an entire RAID
> array. I was hoping a dev already in the know, could shed some light
> on it, and what will be the best set of steps to take to try and recover.
>
> On 2025-08-10 19:02, DanglingPointer wrote:
>> > Are you in a similar position with a corrupted RAID 5 array?
>>
>> No I'm not, but I am close to 75% full on one of my arrays of 30TB.
>> (8TBx3, + 6TB x1) with RAID5 data and RAID1 metadata/system.
>>
>> I'm not sure when I'll have a window to swap out the 6TB for an
>> 8TB. Just curious on what could go wrong if Murphy's law happens.
>> And how to recover.
>>
>>
>> On 11/8/25 06:16, johnh1214@gmail.com wrote:
>>> I did not get any help from the devs mailing list yet.
>>>
>>> For now, I've left things as-is (auto-mounted in RO degraded mode)
>>> and I'm securing a new NAS server, with sufficient drive space to
>>> backup whatever is still readable and error free, then I'll attempt
>>> a recovery using whatever best practice is advised, which currently
>>> is uncertain. I'll also upgrade the kernel to a more recent version
>>> which may have improved BTRFS recovery tools.
>>>
>>> Are you in a similar position with a corrupted RAID 5 array?
>>>
>>> In my case, it's not a HW failure, all the drives checkout fine,
>>> there are no SMART errors, the SATA cables are secured properly,
>>> etc. The problem was entirely due to running low on drive space,
>>> which of course, should not result in a corruption like I'm
>>> experiencing. I had another BTRFS RAID 5 array that ran out of space
>>> before this event, it had no issue, that's why I wasn't concerned
>>> with the low storage space.
>>>
>>> It's possible there was a problem somewhere due to an unclean
>>> shutdown, which became a bigger problem when the drive ran low on
>>> space. My mistake for not running scrubs more regularly.
>>>
>>> I read that a degraded RO mounted array, may cause random read
>>> errors, however the read errors in my case are not random, they are
>>> very specific to certain files.
>>>
>>> The best advice so far, was to copy all the data as is before
>>> performing a recovery. After that, if you can duplicate each drive
>>> exactly, then you can attempt multiple recovery's, however the array
>>> is too big for that much effort, so I have only one shot at it. The
>>> idea is to try mounting in "recovery" mode, if it tries to auto
>>> repair, cancel it, then add extra temporary storage space, because
>>> it was low on data. A balance operation is needed to free up space,
>>> but that seems dangerous given there's corruption, however a scrub
>>> may fail if there's not enough space. The temp storage should be
>>> large enough, very reliable (eg not a plugged in USB), and should
>>> have two partitions for duplicating the metadata.
>>>
>>> One mistake I made, was using RAID 1 for the metadata, when I had
>>> enough drives to use RAID 1C4. Maybe if I had more metadata
>>> redundancy, I'd be able to read everything.
>>>
>>> There are a few possible different steps to take when recovering, as
>>> I said, I'm compiling a list, securing sufficient storage for a full
>>> backup of what's still readable, and trying to determine the best
>>> order of steps to take before giving it a try.
>>>
>>> On 2025-08-10 01:10, DanglingPointer wrote:
>>>> Did anyone help you Johnh1214?
>>>>
>>>> Were you able to recover and get it readable?
>>>>
>>>>
>>>> On 13/7/25 08:02, johnh1214@gmail.com wrote:
>>>>> It was suggested to me, that I contact the btrfs devs group to
>>>>> inform the group of this situation I've encountered.
>>>>>
>>>>> Any advice on what I should do to try and recover from the
>>>>> situation will be appreciated.
>>>>>
>>>>> Fortunately I have a good copy of the most important data backed
>>>>> up, however there are a few recent files not yet backed up that I
>>>>> hope to recover. A lot of the data was for temporary archival
>>>>> reasons, while moving systems around, most of it is no longer
>>>>> required, however, I'd like to try and recover from this situation
>>>>> if possible.
>>>>>
>>>>> The RAID5 array is 5 x 8TB, using RAID 1 for metatdata. I was told
>>>>> I could have instead used RAID1c4 for metatdata, but did not know
>>>>> this beforehand.
>>>>>
>>>>> I was copying data to the raid array, when btrfs ran out of free
>>>>> space, and went into read-only mode. The main issue is not that
>>>>> it's gone into read-only mode, it's that I can no longer read most
>>>>> of the files on the entire array, despite there being no
>>>>> noticeable issues reading the data before the low space problem
>>>>> was encountered. I never imagined, that running out of space could
>>>>> cause the entire file system to become mostly unreadable.
>>>>>
>>>>> I have not yet attempted a recovery, and have left things exactly
>>>>> as they are after the error was encountered. The server has not
>>>>> been rebooted, and I've commented out the mounting of the array
>>>>> from fstab in case there's an unplanned reboot. What I want to do
>>>>> first, is copy what I can from the drive "as-is" in read-only
>>>>> mode, before I attempt a rescue. I can list all the directories
>>>>> and files OK, there's no sign of file system structure corruption,
>>>>> but many of the underlying files are either entirely unreadable,
>>>>> or are only partly readable, with maybe 20% that are 100% readable.
>>>>>
>>>>> The file system reported it was running low on free space, however
>>>>> there was enough listed for the data transfer, about 2x more than
>>>>> I required. I know that when space is low, btrfs will sometimes
>>>>> run out early, despite it claiming there's enough. In the case at
>>>>> hand, I think it claimed to have about 150GB free, and I was
>>>>> copying about 56GB to it. I should have known better than to
>>>>> bother doing the copy, but expected no worse than the copy
>>>>> aborting due to a lack of free space. On single drives BTRFS FS,
>>>>> I've run out of space many times before without issues.
>>>>>
>>>>> After successfully transferring about 50% of the data, the copy
>>>>> aborted. There's no space left, and the FS encountered an error,
>>>>> switching to read-only mode to prevent further damage.
>>>>>
>>>>> Here are the vitals:
>>>>>
>>>>> Kernel: 6.1.0-37-amd64
>>>>>
>>>>> RAID 5 array is 5 x 8TB, using RAID 1 for metatdata
>>>>> Note: someone told me I could have used RAID1cx4 for metatdata, is
>>>>> that true?
>>>>>
>>>>> All 5 Toshiba HDD's appear to be operating normally, and SMART
>>>>> reports all 5 drives are 100% healthy.
>>>>>
>>>>> btrfs fi show
>>>>>
>>>>> Label: 'pl.8000.00' uuid: 5c3747c7-cc6a-45f2-a1e6-860095d0f7cd
>>>>> Total devices 5 FS bytes used 29.02TiB
>>>>> devid 1 size 7.28TiB used 7.28TiB path /dev/sdb
>>>>> devid 2 size 7.28TiB used 7.28TiB path /dev/sdd
>>>>> devid 3 size 7.28TiB used 7.28TiB path /dev/sdc
>>>>> devid 4 size 7.28TiB used 7.28TiB path /dev/sde
>>>>> devid 5 size 7.28TiB used 7.28TiB path /dev/sda
>>>>>
>>>>> Here is a /kern.log dump showing when the problem first appeared.
>>>>> The dump indicates the array ran out of space, and was re-mounted
>>>>> read-only, however the read-only state appears to be bad, with
>>>>> missing "mirrors".
>>>>>
>>>>> In the dump below, the line, "2025-07-07T19:30:28.806270-07:00"
>>>>> reports the array went into read-only mode.
>>>>> Afterwards, it's complaining about mirror 1 and 2 wanted, which
>>>>> appear to be unavailable.
>>>>>
>>>>> /var/log/kern.log
>>>>> 2025-07-07T19:30:28.806185-07:00 host kernel: [3777897.708336]
>>>>> ------------[ cut here ]------------
>>>>> 2025-07-07T19:30:28.806200-07:00 host kernel: [3777897.708339]
>>>>> BTRFS: Transaction aborted (error -28)
>>>>> 2025-07-07T19:30:28.806200-07:00 host kernel: [3777897.708366]
>>>>> WARNING: CPU: 8 PID: 3044847 at fs/btrfs/extent-tree.c:3092
>>>>> __btrfs_free_extent.cold+0x661/0xa0b [btrfs]
>>>>> 2025-07-07T19:30:28.806202-07:00 host kernel: [3777897.708409]
>>>>> Modules linked in: xt_MASQUERADE iptable_nat tcp_diag inet_diag
>>>>> unix_diag nft_nat nft_masq nft_chain_nat nf_nat nf_conntrack
>>>>> nf_defrag_ipv6 nf_defrag_ipv4 vhost_net vho>
>>>>> 2025-07-07T19:30:28.806204-07:00 host kernel: [3777897.708493]
>>>>> drm_kms_helper ahci libahci xhci_pci libata xhci_hcd nvme ixgbe
>>>>> drm nvme_core xfrm_algo t10_pi usbcore scsi_mod crc32_pclmul
>>>>> mdio_devres igb crc32c_intel i2c_piix4 lib>
>>>>> 2025-07-07T19:30:28.806205-07:00 host kernel: [3777897.708524]
>>>>> CPU: 8 PID: 3044847 Comm: kworker/u64:43 Tainted: G X
>>>>> 6.1.0-37-amd64 #1 Debian 6.1.140-1
>>>>> 2025-07-07T19:30:28.806205-07:00 host kernel: [3777897.708528]
>>>>> Hardware name: To Be Filled By O.E.M. To Be Filled By
>>>>> O.E.M./X570D4U-2L2T, BIOS P1.30 12/04/2020
>>>>> 2025-07-07T19:30:28.806206-07:00 host kernel: [3777897.708530]
>>>>> Workqueue: events_unbound btrfs_async_reclaim_metadata_space [btrfs]
>>>>> 2025-07-07T19:30:28.806207-07:00 host kernel: [3777897.708565]
>>>>> RIP: 0010:__btrfs_free_extent.cold+0x661/0xa0b [btrfs]
>>>>> 2025-07-07T19:30:28.806208-07:00 host kernel: [3777897.708596]
>>>>> Code: 78 50 e8 65 de ff ff eb 8e 8b 7c 24 20 e8 35 ee ff ff 84 c0
>>>>> 0f 84 fb 01 00 00 8b 74 24 20 48 c7 c7 b8 94 b3 c0 e8 50 61 38 da
>>>>> <0f> 0b c6 44 24 30 01 44 8b 44 24 3>
>>>>> 2025-07-07T19:30:28.806209-07:00 host kernel: [3777897.708598]
>>>>> RSP: 0018:ffff9cbf1df13a80 EFLAGS: 00010286
>>>>> 2025-07-07T19:30:28.806210-07:00 host kernel: [3777897.708600]
>>>>> RAX: 0000000000000000 RBX: 00002f6f91030000 RCX: 0000000000000027
>>>>> 2025-07-07T19:30:28.806210-07:00 host kernel: [3777897.708602]
>>>>> RDX: ffff8b5b8ec203e8 RSI: 0000000000000001 RDI: ffff8b5b8ec203e0
>>>>> 2025-07-07T19:30:28.806210-07:00 host kernel: [3777897.708604]
>>>>> RBP: 0000000000000000 R08: 0000000000000000 R09: ffff9cbf1df138f8
>>>>> 2025-07-07T19:30:28.806211-07:00 host kernel: [3777897.708605]
>>>>> R10: 0000000000000003 R11: ffff8b5c0f25de90 R12: 0000000000000001
>>>>> 2025-07-07T19:30:28.806211-07:00 host kernel: [3777897.708606]
>>>>> R13: 0000000000000000 R14: 000000000000000a R15: ffff8b448dc605b0
>>>>> 2025-07-07T19:30:28.806212-07:00 host kernel: [3777897.708608] FS:
>>>>> 0000000000000000(0000) GS:ffff8b5b8ec00000(0000)
>>>>> knlGS:0000000000000000
>>>>> 2025-07-07T19:30:28.806212-07:00 host kernel: [3777897.708610] CS:
>>>>> 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
>>>>> 2025-07-07T19:30:28.806213-07:00 host kernel: [3777897.708611]
>>>>> CR2: 0000557ac7d0a438 CR3: 00000001ead0c000 CR4: 0000000000350ee0
>>>>> 2025-07-07T19:30:28.806215-07:00 host kernel: [3777897.708613]
>>>>> Call Trace:
>>>>> 2025-07-07T19:30:28.806215-07:00 host kernel: [3777897.708617] <TASK>
>>>>> 2025-07-07T19:30:28.806216-07:00 host kernel: [3777897.708621] ?
>>>>> __slab_free+0x11/0x2d0
>>>>> 2025-07-07T19:30:28.806234-07:00 host kernel: [3777897.708629]
>>>>> __btrfs_run_delayed_refs+0x2be/0x10b0 [btrfs]
>>>>> 2025-07-07T19:30:28.806234-07:00 host kernel: [3777897.708660] ?
>>>>> btrfs_search_slot+0x8a7/0xc90 [btrfs]
>>>>> 2025-07-07T19:30:28.806235-07:00 host kernel: [3777897.708688] ?
>>>>> srso_return_thunk+0x5/0x10
>>>>> 2025-07-07T19:30:28.806236-07:00 host kernel: [3777897.708693] ?
>>>>> set_extent_buffer_dirty+0x15/0x130 [btrfs]
>>>>> 2025-07-07T19:30:28.806236-07:00 host kernel: [3777897.708723] ?
>>>>> srso_return_thunk+0x5/0x10
>>>>> 2025-07-07T19:30:28.806236-07:00 host kernel: [3777897.708725] ?
>>>>> release_extent_buffer+0x99/0xb0 [btrfs]
>>>>> 2025-07-07T19:30:28.806237-07:00 host kernel: [3777897.708754]
>>>>> btrfs_run_delayed_refs+0x76/0x1a0 [btrfs]
>>>>> 2025-07-07T19:30:28.806237-07:00 host kernel: [3777897.708783]
>>>>> btrfs_start_dirty_block_groups+0x305/0x540 [btrfs]
>>>>> 2025-07-07T19:30:28.806238-07:00 host kernel: [3777897.708817]
>>>>> btrfs_commit_transaction+0xab/0xc50 [btrfs]
>>>>> 2025-07-07T19:30:28.806238-07:00 host kernel: [3777897.708847] ?
>>>>> srso_return_thunk+0x5/0x10
>>>>> 2025-07-07T19:30:28.806239-07:00 host kernel: [3777897.708849] ?
>>>>> start_transaction+0xc3/0x5f0 [btrfs]
>>>>> 2025-07-07T19:30:28.806254-07:00 host kernel: [3777897.708878]
>>>>> flush_space+0xfd/0x5f0 [btrfs]
>>>>> 2025-07-07T19:30:28.806254-07:00 host kernel: [3777897.708911]
>>>>> btrfs_async_reclaim_metadata_space+0x1cc/0x2d0 [btrfs]
>>>>> 2025-07-07T19:30:28.806255-07:00 host kernel: [3777897.708940]
>>>>> process_one_work+0x1c7/0x380
>>>>> 2025-07-07T19:30:28.806255-07:00 host kernel: [3777897.708946]
>>>>> worker_thread+0x4d/0x380
>>>>> 2025-07-07T19:30:28.806255-07:00 host kernel: [3777897.708949] ?
>>>>> rescuer_thread+0x3a0/0x3a0
>>>>> 2025-07-07T19:30:28.806256-07:00 host kernel: [3777897.708950]
>>>>> kthread+0xda/0x100
>>>>> 2025-07-07T19:30:28.806256-07:00 host kernel: [3777897.708954] ?
>>>>> kthread_complete_and_exit+0x20/0x20
>>>>> 2025-07-07T19:30:28.806257-07:00 host kernel: [3777897.708957]
>>>>> ret_from_fork+0x22/0x30
>>>>> 2025-07-07T19:30:28.806258-07:00 host kernel: [3777897.708964]
>>>>> </TASK>
>>>>> 2025-07-07T19:30:28.806258-07:00 host kernel: [3777897.708965]
>>>>> ---[ end trace 0000000000000000 ]---
>>>>> 2025-07-07T19:30:28.806261-07:00 host kernel: [3777897.708967]
>>>>> BTRFS info (device sdb: state A): dumping space info:
>>>>> 2025-07-07T19:30:28.806262-07:00 host kernel: [3777897.708970]
>>>>> BTRFS info (device sdb: state A): space_info DATA has 60921196544
>>>>> free, is full
>>>>> 2025-07-07T19:30:28.806262-07:00 host kernel: [3777897.708972]
>>>>> BTRFS info (device sdb: state A): space_info total=31937462009856,
>>>>> used=31873161326592, pinned=4096, reserved=3377864704,
>>>>> may_use=45056, readonly=1572864 zone_unusable=0
>>>>> 2025-07-07T19:30:28.806263-07:00 host kernel: [3777897.708975]
>>>>> BTRFS info (device sdb: state A): space_info METADATA has -2326528
>>>>> free, is full
>>>>> 2025-07-07T19:30:28.806263-07:00 host kernel: [3777897.708977]
>>>>> BTRFS info (device sdb: state A): space_info total=41875931136,
>>>>> used=39246086144, pinned=926269440, reserved=47579136,
>>>>> may_use=1658191872, readonly=131072 zone_unusable>
>>>>> 2025-07-07T19:30:28.806266-07:00 host kernel: [3777897.708980]
>>>>> BTRFS info (device sdb: state A): space_info SYSTEM has 31686656
>>>>> free, is not full
>>>>> 2025-07-07T19:30:28.806266-07:00 host kernel: [3777897.708982]
>>>>> BTRFS info (device sdb: state A): space_info total=33554432,
>>>>> used=1867776, pinned=0, reserved=0, may_use=0, readonly=0
>>>>> zone_unusable=0
>>>>> 2025-07-07T19:30:28.806267-07:00 host kernel: [3777897.708985]
>>>>> BTRFS info (device sdb: state A): global_block_rsv: size 536870912
>>>>> reserved 0
>>>>> 2025-07-07T19:30:28.806268-07:00 host kernel: [3777897.708986]
>>>>> BTRFS info (device sdb: state A): trans_block_rsv: size 0 reserved 0
>>>>> 2025-07-07T19:30:28.806268-07:00 host kernel: [3777897.708988]
>>>>> BTRFS info (device sdb: state A): chunk_block_rsv: size 0 reserved 0
>>>>> 2025-07-07T19:30:28.806269-07:00 host kernel: [3777897.708990]
>>>>> BTRFS info (device sdb: state A): delayed_block_rsv: size 262144
>>>>> reserved 262144
>>>>> 2025-07-07T19:30:28.806269-07:00 host kernel: [3777897.708992]
>>>>> BTRFS info (device sdb: state A): delayed_refs_rsv: size
>>>>> 2375024640 reserved 0
>>>>> 2025-07-07T19:30:28.806270-07:00 host kernel: [3777897.708994]
>>>>> BTRFS: error (device sdb: state A) in __btrfs_free_extent:3092:
>>>>> errno=-28 No space left
>>>>> 2025-07-07T19:30:28.806270-07:00 host kernel: [3777897.709670]
>>>>> BTRFS info (device sdb: state EA): forced readonly
>>>>> 2025-07-07T19:30:28.806271-07:00 host kernel: [3777897.709672]
>>>>> BTRFS error (device sdb: state EA): failed to run delayed ref for
>>>>> logical 52156220768256 num_bytes 16384 type 176 action 2 ref_mod
>>>>> 1: -28
>>>>> 2025-07-07T19:30:28.806274-07:00 host kernel: [3777897.710683]
>>>>> BTRFS: error (device sdb: state EA) in
>>>>> btrfs_run_delayed_refs:2165: errno=-28 No space left
>>>>> 2025-07-07T19:31:37.922065-07:00 host kernel: [3777966.824892]
>>>>> BTRFS warning (device sdb: state EA): Skipping commit of aborted
>>>>> transaction.
>>>>> 2025-07-07T19:31:37.922082-07:00 host kernel: [3777966.824906]
>>>>> BTRFS: error (device sdb: state EA) in btrfs_sync_log:3161:
>>>>> errno=-5 IO failure
>>>>> 2025-07-07T20:06:21.662072-07:00 host kernel: [3780050.499696]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 2 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.682076-07:00 host kernel: [3780050.519936]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 1 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.682100-07:00 host kernel: [3780050.522082]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 2 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.686053-07:00 host kernel: [3780050.524439]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 1 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.698380-07:00 host kernel: [3780050.526396]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 2 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.698408-07:00 host kernel: [3780050.531760]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 1 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.702623-07:00 host kernel: [3780050.539105]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 2 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.708871-07:00 host kernel: [3780050.543772]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 1 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.717935-07:00 host kernel: [3780050.549936]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 2 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:21.726076-07:00 host kernel: [3780050.558423]
>>>>> BTRFS error (device sdb: state EA): parent transid verify failed
>>>>> on logical 54213629362176 mirror 1 wanted 4093212 found 4093166
>>>>> 2025-07-07T20:06:31.106118-07:00 host kernel: [3780059.943621]
>>>>> verify_parent_transid: 278 callbacks suppressed
>>>>>
>>>>> :
>>>>> NOTE: the parent transit errors, as shown above, repeat forever
>>>>>
>>>>> There are also checksum verify errors logged, like the one shown
>>>>> below, most of them are mirror 2, with a few mirror 1 logged.
>>>>>
>>>>> 2025-07-07T20:29:14.078068-07:00 host kernel: [3781422.870343]
>>>>> BTRFS warning (device sdb: state EA): checksum verify failed on
>>>>> logical 55129635373056 mirror 2 wanted 0x076ea9a7 found 0xf5e8fb26
>>>>> level 1
>>>>>
>>>>> Are the metadata mirrors, 1 and 2, entirely missing, or are two of
>>>>> the 8TB drives involved in the array, no longer available after it
>>>>> went into read-only mode?
>>>>>
>>>>> I use xfce4 for emergency use on this server, and the file browser
>>>>> Thunar, is now showing 4 or the 5 drives as being available for
>>>>> mounting, whereas before this situation happened, none of the
>>>>> drives were shown as being available.
>>>>>
>>>>> This is a very confusing situation. Any assistance on how to
>>>>> proceed from here is appreciated.
>>>>>
>>>>>
>>>
>
next prev parent reply other threads:[~2025-08-12 23:28 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-07-12 22:02 BTRFS RAID 5 array ran out of space, switched to R/O mode with errors johnh1214
2025-08-10 8:10 ` DanglingPointer
2025-08-10 20:16 ` johnh1214
2025-08-11 2:02 ` DanglingPointer
2025-08-12 20:23 ` johnh1214
2025-08-12 23:28 ` DanglingPointer [this message]
2025-08-13 0:43 ` johnh1214
2025-08-13 7:29 ` Paul Jones
2025-08-12 11:19 ` Jani Partanen
2025-08-12 20:55 ` johnh1214
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1ffdaf61-e90b-40df-963a-224831f03163@gmail.com \
--to=danglingpointerexception@gmail.com \
--cc=johnh1214@gmail.com \
--cc=linux-btrfs@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).