From: Kai Krakow <hurikhan77@gmail.com>
To: linux-btrfs@vger.kernel.org
Subject: Re: A Big Thank You, and some Notes on Current Recovery Tools.
Date: Mon, 1 Jan 2018 13:15:29 +0100 [thread overview]
Message-ID: <1clphe-a7q.ln1@hurikhan77.spdns.de> (raw)
In-Reply-To: c3d54bf6-20a8-0b98-baef-d2353b97849e@gmx.com
Am Mon, 01 Jan 2018 18:13:10 +0800 schrieb Qu Wenruo:
> On 2018年01月01日 08:48, Stirling Westrup wrote:
>> Okay, I want to start this post with a HUGE THANK YOU THANK YOU THANK
>> YOU to Nikolay Borisov and most especially to Qu Wenruo!
>>
>> Thanks to their tireless help in answering all my dumb questions I have
>> managed to get my BTRFS working again! As I speak I have the full,
>> non-degraded, quad of drives mounted and am updating my latest backup
>> of their contents.
>>
>> I had a 4-drive setup with 2x4T and 2x2T drives and one of the 2T
>> drives failed, and with help I was able to make a 100% recovery of the
>> lost data. I do have some observations on what I went through though.
>> Take this as constructive criticism, or as a point for discussing
>> additions to the recovery tools:
>>
>> 1) I had a 2T drive die with exactly 3 hard-sector errors and those 3
>> errors exactly coincided with the 3 super-blocks on the drive.
>
> WTF, why all these corruption all happens at btrfs super blocks?!
>
> What a coincident.
Maybe it's a hybrid drive with flash? Or something that went wrong in the
drive-internal cache memory the very time when superblocks where updated?
I bet that the sectors aren't really broken, just the on-disk checksum
didn't match the sector. I remember such things happening to me more than
once back in the days when drives where still connected by molex power
connectors. Those connectors started to get loose over time, due to
thermals or repeated disconnect and connect. That is, drives sometimes
started to no longer have a reliable power source which let to all sorts
of very strange problems, mostly resulting in pseudo-defective sectors.
That said, the OP would like to check the power supply after this
coincidence... Maybe it's aging and no longer able to support all four
drives, CPU, GPU and stuff with stable power.
--
Regards,
Kai
Replies to list-only preferred.
next prev parent reply other threads:[~2018-01-01 12:15 UTC|newest]
Thread overview: 12+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-01-01 0:48 A Big Thank You, and some Notes on Current Recovery Tools Stirling Westrup
2018-01-01 5:21 ` Duncan
2018-01-01 10:13 ` Qu Wenruo
2018-01-01 12:15 ` Kai Krakow [this message]
2018-01-01 19:44 ` Stirling Westrup
2018-01-02 2:03 ` Duncan
2018-01-02 10:02 ` ein
2018-01-02 11:15 ` Paul Jones
2018-01-02 12:45 ` Marat Khalili
2018-01-02 14:45 ` ein
2018-01-01 22:50 ` waxhead
2018-01-02 0:57 ` Qu Wenruo
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1clphe-a7q.ln1@hurikhan77.spdns.de \
--to=hurikhan77@gmail.com \
--cc=linux-btrfs@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox