* xfs_repair hangs at "process newly discovered inodes..."
@ 2022-11-17 18:40 iamdooser
2022-11-17 18:48 ` Eric Sandeen
0 siblings, 1 reply; 6+ messages in thread
From: iamdooser @ 2022-11-17 18:40 UTC (permalink / raw)
To: linux-xfs
Hello,
I'm not sure this is the correct forum; if not I'd appreciate guidance.
I have a Unraid machine that experienced an unmountable file system on
an array disc. Running:
xfs_repair -nv /dev/md3
works, however when running
xfs_repair -v /dev/md3
it stops at "process newly discovered inodes..." and doesn't seem to be
doing anything.
I've asked in the unraid forum and they've directed me to the xfs
mailing list.
Appreciate any help.
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..."
2022-11-17 18:40 xfs_repair hangs at "process newly discovered inodes..." iamdooser
@ 2022-11-17 18:48 ` Eric Sandeen
2022-11-19 17:24 ` iamdooser
0 siblings, 1 reply; 6+ messages in thread
From: Eric Sandeen @ 2022-11-17 18:48 UTC (permalink / raw)
To: iamdooser, linux-xfs
On 11/17/22 12:40 PM, iamdooser wrote:
> Hello,
>
> I'm not sure this is the correct forum; if not I'd appreciate guidance.
>
> I have a Unraid machine that experienced an unmountable file system on an array disc. Running:
>
> xfs_repair -nv /dev/md3
Did that find errors?
> works, however when running
>
> xfs_repair -v /dev/md3
>
> it stops at "process newly discovered inodes..." and doesn't seem to be doing anything.
>
> I've asked in the unraid forum and they've directed me to the xfs mailing list.
>
> Appreciate any help.
Please tell us the version of xfsprogs you're using, and provide the full xfs_repair
output (with and without -n).
If it really looks like a bug, and not simply a slow repair, providing an xfs_metadump
may help us evaluate the problem further.
-Eric
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..."
2022-11-17 18:48 ` Eric Sandeen
@ 2022-11-19 17:24 ` iamdooser
2022-11-21 16:14 ` Carlos Maiolino
2022-11-21 20:48 ` Dave Chinner
0 siblings, 2 replies; 6+ messages in thread
From: iamdooser @ 2022-11-19 17:24 UTC (permalink / raw)
To: Eric Sandeen, linux-xfs
Thank you for responding.
Yes that found errors, although I'm not accustomed to interpreting the
output.
xfs_repair version 5.18.0
The output of xfs_repair -nv was quite large, as was the
xfs_metadump...not sure that's indicative of something, but I've
uploaded them here:
https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing
There doesn't seem to be much activity once it hangs at "process newly
discovered inodes..." so it doesn't seem like just a slow repair.
Desipte there being no sign of activity, I've let it run for 24+ hours
and saw no changes..
On 11/17/22 13:48, Eric Sandeen wrote:
> On 11/17/22 12:40 PM, iamdooser wrote:
>> Hello,
>>
>> I'm not sure this is the correct forum; if not I'd appreciate guidance.
>>
>> I have a Unraid machine that experienced an unmountable file system on an array disc. Running:
>>
>> xfs_repair -nv /dev/md3
>
> Did that find errors?
>
>> works, however when running
>>
>> xfs_repair -v /dev/md3
>>
>> it stops at "process newly discovered inodes..." and doesn't seem to be doing anything.
>>
>> I've asked in the unraid forum and they've directed me to the xfs mailing list.
>>
>> Appreciate any help.
>
> Please tell us the version of xfsprogs you're using, and provide the full xfs_repair
> output (with and without -n).
>
> If it really looks like a bug, and not simply a slow repair, providing an xfs_metadump
> may help us evaluate the problem further.
>
> -Eric
>
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..."
2022-11-19 17:24 ` iamdooser
@ 2022-11-21 16:14 ` Carlos Maiolino
2022-11-21 16:24 ` Eric Sandeen
2022-11-21 20:48 ` Dave Chinner
1 sibling, 1 reply; 6+ messages in thread
From: Carlos Maiolino @ 2022-11-21 16:14 UTC (permalink / raw)
To: iamdooser; +Cc: Eric Sandeen, linux-xfs
Hi.
On Sat, Nov 19, 2022 at 12:24:18PM -0500, iamdooser wrote:
> Thank you for responding.
>
> Yes that found errors, although I'm not accustomed to interpreting the
> output.
>
> xfs_repair version 5.18.0
>
> The output of xfs_repair -nv was quite large, as was the
> xfs_metadump...not sure that's indicative of something, but I've
> uploaded them here:
> https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing
>
>
> There doesn't seem to be much activity once it hangs at "process newly
> discovered inodes..." so it doesn't seem like just a slow repair.
> Desipte there being no sign of activity, I've let it run for 24+ hours
> and saw no changes..
>
Before anything else, could you please try to run the latest xfsprogs from:
https://git.kernel.org/pub/scm/fs/xfs/xfsprogs-dev.git/log/?h=master
A quick test in my laptop using the metadump you provided, finished the repair
in about 5 mintues:
Maximum metadata LSN (2138201505:-135558109) is ahead of log (96:0).
Format log to cycle 2138201508.
XFS_REPAIR Summary Mon Nov 21 17:04:44 2022
Phase Start End Duration
Phase 1: 11/21 16:59:36 11/21 16:59:36
Phase 2: 11/21 16:59:36 11/21 16:59:37 1 second
Phase 3: 11/21 16:59:37 11/21 17:03:47 4 minutes, 10 seconds
Phase 4: 11/21 17:03:47 11/21 17:04:06 19 seconds
Phase 5: 11/21 17:04:06 11/21 17:04:07 1 second
Phase 6: 11/21 17:04:07 11/21 17:04:38 31 seconds
Phase 7: 11/21 17:04:38 11/21 17:04:38
Total run time: 5 minutes, 2 seconds
done
Also, feel free to compress any file you need to share with us :)
Cheers.
>
> On 11/17/22 13:48, Eric Sandeen wrote:
> > On 11/17/22 12:40 PM, iamdooser wrote:
> >> Hello,
> >>
> >> I'm not sure this is the correct forum; if not I'd appreciate guidance.
> >>
> >> I have a Unraid machine that experienced an unmountable file system on an array disc. Running:
> >>
> >> xfs_repair -nv /dev/md3
> >
> > Did that find errors?
> >
> >> works, however when running
> >>
> >> xfs_repair -v /dev/md3
> >>
> >> it stops at "process newly discovered inodes..." and doesn't seem to be doing anything.
> >>
> >> I've asked in the unraid forum and they've directed me to the xfs mailing list.
> >>
> >> Appreciate any help.
> >
> > Please tell us the version of xfsprogs you're using, and provide the full xfs_repair
> > output (with and without -n).
> >
> > If it really looks like a bug, and not simply a slow repair, providing an xfs_metadump
> > may help us evaluate the problem further.
> >
> > -Eric
> >
--
Carlos Maiolino
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..."
2022-11-21 16:14 ` Carlos Maiolino
@ 2022-11-21 16:24 ` Eric Sandeen
0 siblings, 0 replies; 6+ messages in thread
From: Eric Sandeen @ 2022-11-21 16:24 UTC (permalink / raw)
To: Carlos Maiolino, iamdooser; +Cc: linux-xfs
On 11/21/22 10:14 AM, Carlos Maiolino wrote:
> Hi.
>
>
> On Sat, Nov 19, 2022 at 12:24:18PM -0500, iamdooser wrote:
>> Thank you for responding.
>>
>> Yes that found errors, although I'm not accustomed to interpreting the
>> output.
>>
>> xfs_repair version 5.18.0
>>
>> The output of xfs_repair -nv was quite large, as was the
>> xfs_metadump...not sure that's indicative of something, but I've
>> uploaded them here:
>> https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing
>>
>>
>> There doesn't seem to be much activity once it hangs at "process newly
>> discovered inodes..." so it doesn't seem like just a slow repair.
>> Desipte there being no sign of activity, I've let it run for 24+ hours
>> and saw no changes..
>>
>
> Before anything else, could you please try to run the latest xfsprogs from:
>
> https://git.kernel.org/pub/scm/fs/xfs/xfsprogs-dev.git/log/?h=master
>
> A quick test in my laptop using the metadump you provided, finished the repair
> in about 5 mintues:
>
> Maximum metadata LSN (2138201505:-135558109) is ahead of log (96:0).
> Format log to cycle 2138201508.
>
> XFS_REPAIR Summary Mon Nov 21 17:04:44 2022
>
> Phase Start End Duration
> Phase 1: 11/21 16:59:36 11/21 16:59:36
> Phase 2: 11/21 16:59:36 11/21 16:59:37 1 second
> Phase 3: 11/21 16:59:37 11/21 17:03:47 4 minutes, 10 seconds
> Phase 4: 11/21 17:03:47 11/21 17:04:06 19 seconds
> Phase 5: 11/21 17:04:06 11/21 17:04:07 1 second
> Phase 6: 11/21 17:04:07 11/21 17:04:38 31 seconds
> Phase 7: 11/21 17:04:38 11/21 17:04:38
>
> Total run time: 5 minutes, 2 seconds
> done
>
> Also, feel free to compress any file you need to share with us :)
Yup. All those files should compress quite well.
So, the "-nv" output shows many, many errors. While xfs_repair should eventually
make the filesystem metadata consistent again, that's not the same thing as data
recovery.
What happened to this filesystem and/or its storage that prompted the need for
repair?
-Eric
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..."
2022-11-19 17:24 ` iamdooser
2022-11-21 16:14 ` Carlos Maiolino
@ 2022-11-21 20:48 ` Dave Chinner
1 sibling, 0 replies; 6+ messages in thread
From: Dave Chinner @ 2022-11-21 20:48 UTC (permalink / raw)
To: iamdooser; +Cc: Eric Sandeen, linux-xfs
On Sat, Nov 19, 2022 at 12:24:18PM -0500, iamdooser wrote:
> Thank you for responding.
>
> Yes that found errors, although I'm not accustomed to interpreting the
> output.
>
> xfs_repair version 5.18.0
>
> The output of xfs_repair -nv was quite large, as was the xfs_metadump...not
> sure that's indicative of something, but I've uploaded them here:
> https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing
Ok....
According to the the "-nv" output, you a clean log and widespread
per-AG btree corruptions and inconsistencies. Free inodes not found
in the finobt, free space only found in on free sapce btree, records
in btrees out of order, multiply-claimed blocks (cross linked files
and cross linked free space!), etc.
Every AG shows the same corruption pattern - I've never seen a
filesystem with a clean log in this state before. This sort of
widespread lack of consistency in btree structures isn't a result of
an isolated storage media or I/O error - something major has
happened here.
The first thing I have to ask: did you zero the log with xfs_repair
because you couldn't repair it and then take these repair output
dumps? This *smells* zeroing the log with xfs_repair and throwing
away all the metadata in the log after removing a bunch of files
and the system crashing immediately afterwards. Log recovery in that
case would have made the btrees and inode states mostly
consistent...
Can you please explain how the filesystem got into this state in the
first place? What storage you have, what kernel you are running,
what distro/appliance this filesystem is hosted on, what operations
were being performed when it all went wrong, etc? We really need to
know how the fs got into this state so that we can determine if
other users are at risk of this sort of thing...
> There doesn't seem to be much activity once it hangs at "process newly
> discovered inodes..." so it doesn't seem like just a slow repair. Desipte
> there being no sign of activity, I've let it run for 24+ hours and saw no
> changes..
Use "-t 300" for xfs_repair to output a progress report every 5
minutes. Likely the operation is slow because it is IO bound moving
one inode at a time to lost+found...
Cheers,
Dave.
--
Dave Chinner
david@fromorbit.com
^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2022-11-21 20:48 UTC | newest]
Thread overview: 6+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2022-11-17 18:40 xfs_repair hangs at "process newly discovered inodes..." iamdooser
2022-11-17 18:48 ` Eric Sandeen
2022-11-19 17:24 ` iamdooser
2022-11-21 16:14 ` Carlos Maiolino
2022-11-21 16:24 ` Eric Sandeen
2022-11-21 20:48 ` Dave Chinner
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox