From: Brian Foster <bfoster@redhat.com>
To: Christian Kujau <lists@nerdbynature.de>
Cc: linux-xfs@vger.kernel.org
Subject: Re: xfs_repair: couldn't map inode 2089979520, err = 117
Date: Thu, 18 Jan 2018 09:18:27 -0500 [thread overview]
Message-ID: <20180118141827.GA56446@bfoster.bfoster> (raw)
In-Reply-To: <alpine.DEB.2.21.9.1801172204210.14759@trent.utfs.org>
On Wed, Jan 17, 2018 at 10:27:19PM -0800, Christian Kujau wrote:
> Hi,
>
> after a(nother) power outage this disk enclosure (containing two seperate
> disks, connected via USB) was acting up and while one of the disks seems
> to have died, the other one still works and no more hardware errors are
> reported for the enclosure or the disk.
>
> The XFS file system on this disk can be mounted (!) and data can be read,
> but an xfs_repair fails to complete: http://nerdbynature.de/bits/4.14/xfs/
>
> I have (compressed) xfs_metadump images available if anyone is interested.
>
> A timeline of events:
>
> * disk enclosure[0] connected to a Raspbery Pi (aarch64)
> * power failure, and possible power spike after power came back
> * RPI and disk enclosure disconnected from power.
> * disk enclosure connected to an x86-64 machine with lots of RAM
> * xfs_repair (Fedora 27, xfsprogs-4.12) attempted, but the disk enclosure
> was still trying to handle the other (failing) disk and the repair
> failed after some USB resets.
> * failed disk was removed from the enclosure, no more hardware errors
> since, but still xfs_repair is unable to complete.
>
> After a chat on #xfs, Eric and Dave remarked:
>
> > error 117 means the inode is corrupted; probably shouldn't be at that
> > stage, probably indicates a repair bug? just looking at the first few
> > errors
> > bad magic # 0x49414233 in btbno block 28/134141
> > bad magic # 0x46494233 in btcnt block 30/870600
> > the first magic is IAB3 the 2nd is FIB3 those are magic numbers for
> > xfs, but not for the type of block it thought it was checking
>
> ...and also:
>
> > cross linked btrees does tend to indicate something went badly wrong
> > at the hardware level
>
> So, with all that (failed xfs_repair runs that were interrupted by
> hardware faults and also possibly flaky USB controller[0]) - has anybody
> an idea on how to convince xfs_repair to still clean up this mess? Or is
> there no other way than to restore from backup?
>
I suspect, as intimated by the irc snippet above, there's a bug in
xfs_repair where we've run into an on-disk corruption that was expected
to have been resolved one way or another before phase 7. Note that
xfs_repair is not a data recovery tool, so it has full license to simply
throw objects away that are considered beyond repair or cannot be made
sense of. For that reason, it's usually considered a bug for repair to
exit/crash as shown in your logs. I think you'll need to make your
metadump(s) available for anybody to make progress beyond that.
Brian
> Thanks,
> Christian.
>
> [0] When the disk enclosure is connected to the Raspberry Pi 3, the kernel
> usually recognizes it as follows:
>
> usb 1-1.4: new high-speed USB device number 4 using dwc2
> usb 1-1.4: New USB device found, idVendor=7825, idProduct=a2a8
> usb 1-1.4: New USB device strings: Mfr=1, Product=2, SerialNumber=5
> usb 1-1.4: Product: ElitePro Dual U3FW
> usb 1-1.4: Manufacturer: OWC
> usb 1-1.4: SerialNumber: DB9876543211160
> usb 1-1.4: The driver for the USB controller dwc2_hsotg does not support scatter-gather which is
> usb 1-1.4: required by the UAS driver. Please try an other USB controller if you wish to use UAS.
> usb 1-1.4: The driver for the USB controller dwc2_hsotg does not support scatter-gather which is
> usb 1-1.4: required by the UAS driver. Please try an other USB controller if you wish to use UAS.
> usb-storage 1-1.4:1.0: USB Mass Storage device detected
> scsi host0: usb-storage 1-1.4:1.0
> scsi 0:0:0:0: Direct-Access ElitePro Dual U3FW-1 0006 PQ: 0 ANSI: 6
> scsi 0:0:0:1: Direct-Access ElitePro Dual U3FW-2 0006 PQ: 0 ANSI: 6
> sd 0:0:0:0: Attached scsi generic sg0 type 0
> sd 0:0:0:0: [sda] Very big device. Trying to use READ CAPACITY(16).
> sd 0:0:0:0: [sda] 7814037168 512-byte logical blocks: (4.00 TB/3.64 TiB)
> sd 0:0:0:0: [sda] Write Protect is off
> sd 0:0:0:0: [sda] Mode Sense: 47 00 10 08
> sd 0:0:0:0: [sda] No Caching mode page found
> sd 0:0:0:0: [sda] Assuming drive cache: write through
> sd 0:0:0:0: [sda] Very big device. Trying to use READ CAPACITY(16).
> [...]
>
>
> --
> BOFH excuse #449:
>
> greenpeace free'd the mallocs
> --
> To unsubscribe from this list: send the line "unsubscribe linux-xfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
next prev parent reply other threads:[~2018-01-18 14:18 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-01-18 6:27 xfs_repair: couldn't map inode 2089979520, err = 117 Christian Kujau
2018-01-18 14:18 ` Brian Foster [this message]
2018-01-18 18:36 ` Brian Foster
2018-01-18 18:55 ` Darrick J. Wong
2018-01-18 19:59 ` Brian Foster
2018-01-29 3:22 ` Christian Kujau
2018-01-18 21:59 ` Christian Kujau
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180118141827.GA56446@bfoster.bfoster \
--to=bfoster@redhat.com \
--cc=linux-xfs@vger.kernel.org \
--cc=lists@nerdbynature.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox