* mdadm raid5 dropped 2 disks
@ 2016-02-05 10:30 André Teichert
2016-02-06 9:45 ` Wols Lists
2016-02-06 10:54 ` André Teichert
0 siblings, 2 replies; 5+ messages in thread
From: André Teichert @ 2016-02-05 10:30 UTC (permalink / raw)
To: linux-raid
Hi,
I had a raid5 (mdadm V. 3.2.5) with 3 disks. Within an hour 2 disks dropped.
Both disks show smart error 184, but I can still read them.
First I did a full dd-copy of each disk to imagefile image[123] and
wrote it back to a large 4tb disk with 3 partitions.
mdadm -E /dev/sda1
/dev/sda1:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x0
Array UUID : 8bf0a3b8:a98e95fd:6a0884e6:fbe6ab09
Name : server:0 (local to host server)
Creation Time : Sun Nov 24 04:21:09 2013
Raid Level : raid5
Raid Devices : 3
Avail Dev Size : 1953262961 (931.39 GiB 1000.07 GB)
Array Size : 1953262592 (1862.78 GiB 2000.14 GB)
Used Dev Size : 1953262592 (931.39 GiB 1000.07 GB)
Data Offset : 262144 sectors
Super Offset : 8 sectors
State : active
Device UUID : 6f793025:415d8c8b:e7d37bbb:19524380
Update Time : Wed Feb 3 10:16:27 2016
Checksum : 74a4a730 - correct
Events : 311
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 2
Array State : .AA ('A' == active, '.' == missing)
mdadm -E /dev/sda2
/dev/sda2:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x0
Array UUID : 8bf0a3b8:a98e95fd:6a0884e6:fbe6ab09
Name : server:0 (local to host server)
Creation Time : Sun Nov 24 04:21:09 2013
Raid Level : raid5
Raid Devices : 3
Avail Dev Size : 1953262961 (931.39 GiB 1000.07 GB)
Array Size : 1953262592 (1862.78 GiB 2000.14 GB)
Used Dev Size : 1953262592 (931.39 GiB 1000.07 GB)
Data Offset : 262144 sectors
Super Offset : 8 sectors
State : clean
Device UUID : fc963d80:307b6345:c95b6d94:162c7c7c
Update Time : Wed Feb 3 10:16:40 2016
Checksum : 5eaf449a - correct
Events : 314
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 1
Array State : .A. ('A' == active, '.' == missing)
mdadm -E /dev/sda3
/dev/sda3:
Magic : a92b4efc
Version : 1.2
Feature Map : 0x0
Array UUID : 8bf0a3b8:a98e95fd:6a0884e6:fbe6ab09
Name : server:0 (local to host server)
Creation Time : Sun Nov 24 04:21:09 2013
Raid Level : raid5
Raid Devices : 3
Avail Dev Size : 1953262961 (931.39 GiB 1000.07 GB)
Array Size : 1953262592 (1862.78 GiB 2000.14 GB)
Used Dev Size : 1953262592 (931.39 GiB 1000.07 GB)
Data Offset : 262144 sectors
Super Offset : 8 sectors
State : active
Device UUID : 73b1275f:8600a6b4:51234150:e035eef3
Update Time : Wed Feb 3 09:37:09 2016
Checksum : e024ac15 - correct
Events : 217
Layout : left-symmetric
Chunk Size : 512K
Device Role : Active device 0
Array State : AAA ('A' == active, '.' == missing)
Seems like sda3 dropped first and big difference in events so I started
only sda1+sda2
mdadm --assemble --force /dev/md0 /dev/sda1 /dev/sda2
Seemed to work and assembled the raid with 2/3 disks clean.
The filesystem is ext4.
Runnung "fsck -y /dev/md0" with lots of errors.
mount -t ext4 /dev/md0 /mnt didnt recognize the filesystem
Should I try --create --assmume-clean sda1 sda2 missing?
I try to stay calm and pray for help.
thx a lot
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: mdadm raid5 dropped 2 disks
2016-02-05 10:30 mdadm raid5 dropped 2 disks André Teichert
@ 2016-02-06 9:45 ` Wols Lists
2016-02-06 10:54 ` André Teichert
1 sibling, 0 replies; 5+ messages in thread
From: Wols Lists @ 2016-02-06 9:45 UTC (permalink / raw)
To: André Teichert, linux-raid
On 05/02/16 10:30, André Teichert wrote:
> Should I try --create --assmume-clean sda1 sda2 missing?
> I try to stay calm and pray for help.
> thx a lot
Find and download lsdrv by Phil Turmel. You'll need Python 2.7 to run it
iirc.
DO NOT EVEN CONSIDER running "mdadm --create" until you've got all the
data you can out of the drive headers (lsdrv is intended to get this data).
Post the data from lsdrv to the list, and then someone else will chime
in and help you further.
(If you get --create wrong, you've probably just permanently lost the
contents of your array ...)
Cheers,
Wol
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: mdadm raid5 dropped 2 disks
2016-02-05 10:30 mdadm raid5 dropped 2 disks André Teichert
2016-02-06 9:45 ` Wols Lists
@ 2016-02-06 10:54 ` André Teichert
2016-02-06 11:46 ` Wols Lists
1 sibling, 1 reply; 5+ messages in thread
From: André Teichert @ 2016-02-06 10:54 UTC (permalink / raw)
To: linux-raid
Hi Wol,
the raid is save now. I didnt work on the disks itself only with images.
dd if=/dev/sd[abc]1 of=/mnt/image[123].dd bs=1
losetup /dev/loop1 /mnt/image1.dd
losetup /dev/loop2 /mnt/image2.dd
mdadm --create /dev/md0 --raid-devices=3 --level=5 --assume-clean
--size=976631296 --raid-devices=3 missing /dev/loop2 /dev/loop1
fsck without any errors. Data is back!
What I maybe did wrong in the first run?
1) I wrote the first backup of the disks to partitions, that were a
little bigger than the original one. These were fresh drives with no
data writen so far.
2) I did the create with the right order (missing loo2 loop1) from
mdadm -E "Device Role"
glad it worked. Will save lsdrv and hope never to use it. Create new
backupplan!
cheers
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: mdadm raid5 dropped 2 disks
2016-02-06 10:54 ` André Teichert
@ 2016-02-06 11:46 ` Wols Lists
2016-02-06 14:03 ` Phil Turmel
0 siblings, 1 reply; 5+ messages in thread
From: Wols Lists @ 2016-02-06 11:46 UTC (permalink / raw)
To: André Teichert, linux-raid
On 06/02/16 10:54, André Teichert wrote:
> glad it worked. Will save lsdrv and hope never to use it. Create new
> backupplan!
Please DO use lsdrv. EVERY time you create or change the config of an
array, run it and save the info somewhere safe.
Neil has just stepped down as maintainer and Phil is now "one of the
team", he wrote that utility specifically to help in recovering crashed
arrays. If anything goes wrong and you've got that data to hand, it'll
make their lives much easier to help you.
Cheers,
Wol
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: mdadm raid5 dropped 2 disks
2016-02-06 11:46 ` Wols Lists
@ 2016-02-06 14:03 ` Phil Turmel
0 siblings, 0 replies; 5+ messages in thread
From: Phil Turmel @ 2016-02-06 14:03 UTC (permalink / raw)
To: Wols Lists, André Teichert, linux-raid
Good morning André, Wol,
On 02/06/2016 06:46 AM, Wols Lists wrote:
> On 06/02/16 10:54, André Teichert wrote:
>> glad it worked. Will save lsdrv and hope never to use it. Create new
>> backupplan!
>
> Please DO use lsdrv. EVERY time you create or change the config of an
> array, run it and save the info somewhere safe.
Yes, lsdrv is intended to document a running system for later use when
not running. Especially serial numbers, device names, UUIDs, and layer
relationships. It is still helpful later, but naturally some info can
be missed.
> Neil has just stepped down as maintainer and Phil is now "one of the
> team",
That's a bit of a stretch. I'm not maintaining any code and I'm
volunteering on the list as I have in the past.
> he wrote that utility specifically to help in recovering crashed
> arrays. If anything goes wrong and you've got that data to hand, it'll
> make their lives much easier to help you.
Yes, array configuration documentation is critical to recovery.
Phil
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2016-02-06 14:03 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2016-02-05 10:30 mdadm raid5 dropped 2 disks André Teichert
2016-02-06 9:45 ` Wols Lists
2016-02-06 10:54 ` André Teichert
2016-02-06 11:46 ` Wols Lists
2016-02-06 14:03 ` Phil Turmel
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).