* xfs_repair hangs at "process newly discovered inodes..." @ 2022-11-17 18:40 iamdooser 2022-11-17 18:48 ` Eric Sandeen 0 siblings, 1 reply; 6+ messages in thread From: iamdooser @ 2022-11-17 18:40 UTC (permalink / raw) To: linux-xfs Hello, I'm not sure this is the correct forum; if not I'd appreciate guidance. I have a Unraid machine that experienced an unmountable file system on an array disc. Running: xfs_repair -nv /dev/md3 works, however when running xfs_repair -v /dev/md3 it stops at "process newly discovered inodes..." and doesn't seem to be doing anything. I've asked in the unraid forum and they've directed me to the xfs mailing list. Appreciate any help. ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..." 2022-11-17 18:40 xfs_repair hangs at "process newly discovered inodes..." iamdooser @ 2022-11-17 18:48 ` Eric Sandeen 2022-11-19 17:24 ` iamdooser 0 siblings, 1 reply; 6+ messages in thread From: Eric Sandeen @ 2022-11-17 18:48 UTC (permalink / raw) To: iamdooser, linux-xfs On 11/17/22 12:40 PM, iamdooser wrote: > Hello, > > I'm not sure this is the correct forum; if not I'd appreciate guidance. > > I have a Unraid machine that experienced an unmountable file system on an array disc. Running: > > xfs_repair -nv /dev/md3 Did that find errors? > works, however when running > > xfs_repair -v /dev/md3 > > it stops at "process newly discovered inodes..." and doesn't seem to be doing anything. > > I've asked in the unraid forum and they've directed me to the xfs mailing list. > > Appreciate any help. Please tell us the version of xfsprogs you're using, and provide the full xfs_repair output (with and without -n). If it really looks like a bug, and not simply a slow repair, providing an xfs_metadump may help us evaluate the problem further. -Eric ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..." 2022-11-17 18:48 ` Eric Sandeen @ 2022-11-19 17:24 ` iamdooser 2022-11-21 16:14 ` Carlos Maiolino 2022-11-21 20:48 ` Dave Chinner 0 siblings, 2 replies; 6+ messages in thread From: iamdooser @ 2022-11-19 17:24 UTC (permalink / raw) To: Eric Sandeen, linux-xfs Thank you for responding. Yes that found errors, although I'm not accustomed to interpreting the output. xfs_repair version 5.18.0 The output of xfs_repair -nv was quite large, as was the xfs_metadump...not sure that's indicative of something, but I've uploaded them here: https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing There doesn't seem to be much activity once it hangs at "process newly discovered inodes..." so it doesn't seem like just a slow repair. Desipte there being no sign of activity, I've let it run for 24+ hours and saw no changes.. On 11/17/22 13:48, Eric Sandeen wrote: > On 11/17/22 12:40 PM, iamdooser wrote: >> Hello, >> >> I'm not sure this is the correct forum; if not I'd appreciate guidance. >> >> I have a Unraid machine that experienced an unmountable file system on an array disc. Running: >> >> xfs_repair -nv /dev/md3 > > Did that find errors? > >> works, however when running >> >> xfs_repair -v /dev/md3 >> >> it stops at "process newly discovered inodes..." and doesn't seem to be doing anything. >> >> I've asked in the unraid forum and they've directed me to the xfs mailing list. >> >> Appreciate any help. > > Please tell us the version of xfsprogs you're using, and provide the full xfs_repair > output (with and without -n). > > If it really looks like a bug, and not simply a slow repair, providing an xfs_metadump > may help us evaluate the problem further. > > -Eric > ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..." 2022-11-19 17:24 ` iamdooser @ 2022-11-21 16:14 ` Carlos Maiolino 2022-11-21 16:24 ` Eric Sandeen 2022-11-21 20:48 ` Dave Chinner 1 sibling, 1 reply; 6+ messages in thread From: Carlos Maiolino @ 2022-11-21 16:14 UTC (permalink / raw) To: iamdooser; +Cc: Eric Sandeen, linux-xfs Hi. On Sat, Nov 19, 2022 at 12:24:18PM -0500, iamdooser wrote: > Thank you for responding. > > Yes that found errors, although I'm not accustomed to interpreting the > output. > > xfs_repair version 5.18.0 > > The output of xfs_repair -nv was quite large, as was the > xfs_metadump...not sure that's indicative of something, but I've > uploaded them here: > https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing > > > There doesn't seem to be much activity once it hangs at "process newly > discovered inodes..." so it doesn't seem like just a slow repair. > Desipte there being no sign of activity, I've let it run for 24+ hours > and saw no changes.. > Before anything else, could you please try to run the latest xfsprogs from: https://git.kernel.org/pub/scm/fs/xfs/xfsprogs-dev.git/log/?h=master A quick test in my laptop using the metadump you provided, finished the repair in about 5 mintues: Maximum metadata LSN (2138201505:-135558109) is ahead of log (96:0). Format log to cycle 2138201508. XFS_REPAIR Summary Mon Nov 21 17:04:44 2022 Phase Start End Duration Phase 1: 11/21 16:59:36 11/21 16:59:36 Phase 2: 11/21 16:59:36 11/21 16:59:37 1 second Phase 3: 11/21 16:59:37 11/21 17:03:47 4 minutes, 10 seconds Phase 4: 11/21 17:03:47 11/21 17:04:06 19 seconds Phase 5: 11/21 17:04:06 11/21 17:04:07 1 second Phase 6: 11/21 17:04:07 11/21 17:04:38 31 seconds Phase 7: 11/21 17:04:38 11/21 17:04:38 Total run time: 5 minutes, 2 seconds done Also, feel free to compress any file you need to share with us :) Cheers. > > On 11/17/22 13:48, Eric Sandeen wrote: > > On 11/17/22 12:40 PM, iamdooser wrote: > >> Hello, > >> > >> I'm not sure this is the correct forum; if not I'd appreciate guidance. > >> > >> I have a Unraid machine that experienced an unmountable file system on an array disc. Running: > >> > >> xfs_repair -nv /dev/md3 > > > > Did that find errors? > > > >> works, however when running > >> > >> xfs_repair -v /dev/md3 > >> > >> it stops at "process newly discovered inodes..." and doesn't seem to be doing anything. > >> > >> I've asked in the unraid forum and they've directed me to the xfs mailing list. > >> > >> Appreciate any help. > > > > Please tell us the version of xfsprogs you're using, and provide the full xfs_repair > > output (with and without -n). > > > > If it really looks like a bug, and not simply a slow repair, providing an xfs_metadump > > may help us evaluate the problem further. > > > > -Eric > > -- Carlos Maiolino ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..." 2022-11-21 16:14 ` Carlos Maiolino @ 2022-11-21 16:24 ` Eric Sandeen 0 siblings, 0 replies; 6+ messages in thread From: Eric Sandeen @ 2022-11-21 16:24 UTC (permalink / raw) To: Carlos Maiolino, iamdooser; +Cc: linux-xfs On 11/21/22 10:14 AM, Carlos Maiolino wrote: > Hi. > > > On Sat, Nov 19, 2022 at 12:24:18PM -0500, iamdooser wrote: >> Thank you for responding. >> >> Yes that found errors, although I'm not accustomed to interpreting the >> output. >> >> xfs_repair version 5.18.0 >> >> The output of xfs_repair -nv was quite large, as was the >> xfs_metadump...not sure that's indicative of something, but I've >> uploaded them here: >> https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing >> >> >> There doesn't seem to be much activity once it hangs at "process newly >> discovered inodes..." so it doesn't seem like just a slow repair. >> Desipte there being no sign of activity, I've let it run for 24+ hours >> and saw no changes.. >> > > Before anything else, could you please try to run the latest xfsprogs from: > > https://git.kernel.org/pub/scm/fs/xfs/xfsprogs-dev.git/log/?h=master > > A quick test in my laptop using the metadump you provided, finished the repair > in about 5 mintues: > > Maximum metadata LSN (2138201505:-135558109) is ahead of log (96:0). > Format log to cycle 2138201508. > > XFS_REPAIR Summary Mon Nov 21 17:04:44 2022 > > Phase Start End Duration > Phase 1: 11/21 16:59:36 11/21 16:59:36 > Phase 2: 11/21 16:59:36 11/21 16:59:37 1 second > Phase 3: 11/21 16:59:37 11/21 17:03:47 4 minutes, 10 seconds > Phase 4: 11/21 17:03:47 11/21 17:04:06 19 seconds > Phase 5: 11/21 17:04:06 11/21 17:04:07 1 second > Phase 6: 11/21 17:04:07 11/21 17:04:38 31 seconds > Phase 7: 11/21 17:04:38 11/21 17:04:38 > > Total run time: 5 minutes, 2 seconds > done > > Also, feel free to compress any file you need to share with us :) Yup. All those files should compress quite well. So, the "-nv" output shows many, many errors. While xfs_repair should eventually make the filesystem metadata consistent again, that's not the same thing as data recovery. What happened to this filesystem and/or its storage that prompted the need for repair? -Eric ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: xfs_repair hangs at "process newly discovered inodes..." 2022-11-19 17:24 ` iamdooser 2022-11-21 16:14 ` Carlos Maiolino @ 2022-11-21 20:48 ` Dave Chinner 1 sibling, 0 replies; 6+ messages in thread From: Dave Chinner @ 2022-11-21 20:48 UTC (permalink / raw) To: iamdooser; +Cc: Eric Sandeen, linux-xfs On Sat, Nov 19, 2022 at 12:24:18PM -0500, iamdooser wrote: > Thank you for responding. > > Yes that found errors, although I'm not accustomed to interpreting the > output. > > xfs_repair version 5.18.0 > > The output of xfs_repair -nv was quite large, as was the xfs_metadump...not > sure that's indicative of something, but I've uploaded them here: > https://drive.google.com/drive/folders/1OyQOZNsTS1w1Utx1ZfQEH-bS_Cyj8-F2?usp=sharing Ok.... According to the the "-nv" output, you a clean log and widespread per-AG btree corruptions and inconsistencies. Free inodes not found in the finobt, free space only found in on free sapce btree, records in btrees out of order, multiply-claimed blocks (cross linked files and cross linked free space!), etc. Every AG shows the same corruption pattern - I've never seen a filesystem with a clean log in this state before. This sort of widespread lack of consistency in btree structures isn't a result of an isolated storage media or I/O error - something major has happened here. The first thing I have to ask: did you zero the log with xfs_repair because you couldn't repair it and then take these repair output dumps? This *smells* zeroing the log with xfs_repair and throwing away all the metadata in the log after removing a bunch of files and the system crashing immediately afterwards. Log recovery in that case would have made the btrees and inode states mostly consistent... Can you please explain how the filesystem got into this state in the first place? What storage you have, what kernel you are running, what distro/appliance this filesystem is hosted on, what operations were being performed when it all went wrong, etc? We really need to know how the fs got into this state so that we can determine if other users are at risk of this sort of thing... > There doesn't seem to be much activity once it hangs at "process newly > discovered inodes..." so it doesn't seem like just a slow repair. Desipte > there being no sign of activity, I've let it run for 24+ hours and saw no > changes.. Use "-t 300" for xfs_repair to output a progress report every 5 minutes. Likely the operation is slow because it is IO bound moving one inode at a time to lost+found... Cheers, Dave. -- Dave Chinner david@fromorbit.com ^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2022-11-21 20:48 UTC | newest] Thread overview: 6+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2022-11-17 18:40 xfs_repair hangs at "process newly discovered inodes..." iamdooser 2022-11-17 18:48 ` Eric Sandeen 2022-11-19 17:24 ` iamdooser 2022-11-21 16:14 ` Carlos Maiolino 2022-11-21 16:24 ` Eric Sandeen 2022-11-21 20:48 ` Dave Chinner
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox