* RAID5 faild while in degraded mode, need help @ 2012-07-08 19:05 Dietrich Heise 2012-07-09 0:12 ` NeilBrown 0 siblings, 1 reply; 5+ messages in thread From: Dietrich Heise @ 2012-07-08 19:05 UTC (permalink / raw) To: linux-raid Hi, the following Problem, One of four drives has S.M.A.R.T. errors, so I removed it and replaced, with a new one. In the time the drive was rebuilding, one of the three left devices has an I/O error (sdd1) (sdc1 was the replaced drive an was syncing). Now the following happends (two drives are spare drives :( ) p3 disks # mdadm -D /dev/md1 /dev/md1: Version : 1.2 Creation Time : Mon Feb 28 19:57:56 2011 Raid Level : raid5 Used Dev Size : 1465126400 (1397.25 GiB 1500.29 GB) Raid Devices : 4 Total Devices : 4 Persistence : Superblock is persistent Update Time : Sun Jul 8 20:37:12 2012 State : active, FAILED, Not Started Active Devices : 2 Working Devices : 4 Failed Devices : 0 Spare Devices : 2 Layout : left-symmetric Chunk Size : 512K Name : p3:0 (local to host p3) UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 Events : 121205 Number Major Minor RaidDevice State 0 8 81 0 active sync /dev/sdf1 1 8 65 1 active sync /dev/sde1 2 0 0 2 removed 3 0 0 3 removed 4 8 49 - spare /dev/sdd1 5 8 33 - spare /dev/sdc1 here is more information: p3 disks # mdadm -E /dev/sdc1 /dev/sdc1: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 Name : p3:0 (local to host p3) Creation Time : Mon Feb 28 19:57:56 2011 Raid Level : raid5 Raid Devices : 4 Avail Dev Size : 2930275057 (1397.26 GiB 1500.30 GB) Array Size : 8790758400 (4191.76 GiB 4500.87 GB) Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) Data Offset : 2048 sectors Super Offset : 8 sectors State : active Device UUID : caefb029:526187ef:2051f578:db2b82b7 Update Time : Sun Jul 8 20:37:12 2012 Checksum : 18e2bfe1 - correct Events : 121205 Layout : left-symmetric Chunk Size : 512K Device Role : spare Array State : AA.. ('A' == active, '.' == missing) p3 disks # mdadm -E /dev/sdd1 /dev/sdd1: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 Name : p3:0 (local to host p3) Creation Time : Mon Feb 28 19:57:56 2011 Raid Level : raid5 Raid Devices : 4 Avail Dev Size : 2930269954 (1397.26 GiB 1500.30 GB) Array Size : 8790758400 (4191.76 GiB 4500.87 GB) Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) Data Offset : 2048 sectors Super Offset : 8 sectors State : active Device UUID : 4231e244:60e27ed4:eff405d0:2e615493 Update Time : Sun Jul 8 20:37:12 2012 Checksum : 4bec6e25 - correct Events : 0 Layout : left-symmetric Chunk Size : 512K Device Role : spare Array State : AA.. ('A' == active, '.' == missing) p3 disks # mdadm -E /dev/sde1 /dev/sde1: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 Name : p3:0 (local to host p3) Creation Time : Mon Feb 28 19:57:56 2011 Raid Level : raid5 Raid Devices : 4 Avail Dev Size : 2930253889 (1397.25 GiB 1500.29 GB) Array Size : 8790758400 (4191.76 GiB 4500.87 GB) Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) Data Offset : 2048 sectors Super Offset : 8 sectors State : active Device UUID : 28b08f44:4cc24663:84d39337:94c35d67 Update Time : Sun Jul 8 20:37:12 2012 Checksum : 15faa8a1 - correct Events : 121205 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 1 Array State : AA.. ('A' == active, '.' == missing) p3 disks # mdadm -E /dev/sdf1 /dev/sdf1: Magic : a92b4efc Version : 1.2 Feature Map : 0x0 Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 Name : p3:0 (local to host p3) Creation Time : Mon Feb 28 19:57:56 2011 Raid Level : raid5 Raid Devices : 4 Avail Dev Size : 2930269954 (1397.26 GiB 1500.30 GB) Array Size : 8790758400 (4191.76 GiB 4500.87 GB) Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) Data Offset : 2048 sectors Super Offset : 8 sectors State : active Device UUID : 78d5600a:91927758:f78a1cea:3bfa3f5b Update Time : Sun Jul 8 20:37:12 2012 Checksum : 7767cb10 - correct Events : 121205 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 0 Array State : AA.. ('A' == active, '.' == missing) Is there a way to repair the raid? thanks! Dietrich ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: RAID5 faild while in degraded mode, need help 2012-07-08 19:05 RAID5 faild while in degraded mode, need help Dietrich Heise @ 2012-07-09 0:12 ` NeilBrown 2012-07-09 11:02 ` Dietrich Heise 0 siblings, 1 reply; 5+ messages in thread From: NeilBrown @ 2012-07-09 0:12 UTC (permalink / raw) To: Dietrich Heise; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 7071 bytes --] On Sun, 8 Jul 2012 21:05:02 +0200 Dietrich Heise <dh@dhde.de> wrote: > Hi, > > the following Problem, > One of four drives has S.M.A.R.T. errors, so I removed it and > replaced, with a new one. > > In the time the drive was rebuilding, one of the three left devices > has an I/O error (sdd1) (sdc1 was the replaced drive an was syncing). > > Now the following happends (two drives are spare drives :( ) It looks like you tried to --add /dev/sdd1 back in after it failed, and mdadm let new. Newer versions of mdadm will refuse as that is not a good thing to do but it shouldn't stop you getting your data back. First thing to realise is that you could have data corruption. There is at least one block in the array which cannot be recovered, possibly more. i.e. any block on sdd1 which is bad, and any block at the same offset in sdc1. These blocks may not be in files which would be lucky, or they may contain important metadata which might mean you've lost lots of files. If you hadn't tried to --add /dev/sdd1 you could just force-assemble the array back to degraded mode (without sdc1) and back up any critical data. As sdd1 now thinks it is a spare you need to re-create the array instead: mdadm -S /dev/md1 mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 /dev/sdd1 missing or mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 missing /dev/sdd1 depending on whether sdd1 as the 3rd or 4th device in the array - I cannot tell from the output here. You should then be able to mount the array and backup stuff. You then want to use 'ddrescue' to copy sdd1 onto a device with no bad blocks, and assemble the array using the device instead of sdd1. Finally, you can add the new spare (sdc1) to the array and it should rebuild successfully - providing there are no bad blocks on sdf1 or sde1. I hope that makes sense. Do ask if anything is unclear. NeilBrown > > p3 disks # mdadm -D /dev/md1 > /dev/md1: > Version : 1.2 > Creation Time : Mon Feb 28 19:57:56 2011 > Raid Level : raid5 > Used Dev Size : 1465126400 (1397.25 GiB 1500.29 GB) > Raid Devices : 4 > Total Devices : 4 > Persistence : Superblock is persistent > > Update Time : Sun Jul 8 20:37:12 2012 > State : active, FAILED, Not Started > Active Devices : 2 > Working Devices : 4 > Failed Devices : 0 > Spare Devices : 2 > > Layout : left-symmetric > Chunk Size : 512K > > Name : p3:0 (local to host p3) > UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > Events : 121205 > > Number Major Minor RaidDevice State > 0 8 81 0 active sync /dev/sdf1 > 1 8 65 1 active sync /dev/sde1 > 2 0 0 2 removed > 3 0 0 3 removed > > 4 8 49 - spare /dev/sdd1 > 5 8 33 - spare /dev/sdc1 > > here is more information: > > p3 disks # mdadm -E /dev/sdc1 > /dev/sdc1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > Name : p3:0 (local to host p3) > Creation Time : Mon Feb 28 19:57:56 2011 > Raid Level : raid5 > Raid Devices : 4 > > Avail Dev Size : 2930275057 (1397.26 GiB 1500.30 GB) > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : active > Device UUID : caefb029:526187ef:2051f578:db2b82b7 > > Update Time : Sun Jul 8 20:37:12 2012 > Checksum : 18e2bfe1 - correct > Events : 121205 > > Layout : left-symmetric > Chunk Size : 512K > > Device Role : spare > Array State : AA.. ('A' == active, '.' == missing) > p3 disks # mdadm -E /dev/sdd1 > /dev/sdd1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > Name : p3:0 (local to host p3) > Creation Time : Mon Feb 28 19:57:56 2011 > Raid Level : raid5 > Raid Devices : 4 > > Avail Dev Size : 2930269954 (1397.26 GiB 1500.30 GB) > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : active > Device UUID : 4231e244:60e27ed4:eff405d0:2e615493 > > Update Time : Sun Jul 8 20:37:12 2012 > Checksum : 4bec6e25 - correct > Events : 0 > > Layout : left-symmetric > Chunk Size : 512K > > Device Role : spare > Array State : AA.. ('A' == active, '.' == missing) > p3 disks # mdadm -E /dev/sde1 > /dev/sde1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > Name : p3:0 (local to host p3) > Creation Time : Mon Feb 28 19:57:56 2011 > Raid Level : raid5 > Raid Devices : 4 > > Avail Dev Size : 2930253889 (1397.25 GiB 1500.29 GB) > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : active > Device UUID : 28b08f44:4cc24663:84d39337:94c35d67 > > Update Time : Sun Jul 8 20:37:12 2012 > Checksum : 15faa8a1 - correct > Events : 121205 > > Layout : left-symmetric > Chunk Size : 512K > > Device Role : Active device 1 > Array State : AA.. ('A' == active, '.' == missing) > p3 disks # mdadm -E /dev/sdf1 > /dev/sdf1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > Name : p3:0 (local to host p3) > Creation Time : Mon Feb 28 19:57:56 2011 > Raid Level : raid5 > Raid Devices : 4 > > Avail Dev Size : 2930269954 (1397.26 GiB 1500.30 GB) > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : active > Device UUID : 78d5600a:91927758:f78a1cea:3bfa3f5b > > Update Time : Sun Jul 8 20:37:12 2012 > Checksum : 7767cb10 - correct > Events : 121205 > > Layout : left-symmetric > Chunk Size : 512K > > Device Role : Active device 0 > Array State : AA.. ('A' == active, '.' == missing) > > Is there a way to repair the raid? > > thanks! > Dietrich > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html [-- Attachment #2: signature.asc --] [-- Type: application/pgp-signature, Size: 828 bytes --] ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: RAID5 faild while in degraded mode, need help 2012-07-09 0:12 ` NeilBrown @ 2012-07-09 11:02 ` Dietrich Heise 2012-07-09 23:02 ` NeilBrown 0 siblings, 1 reply; 5+ messages in thread From: Dietrich Heise @ 2012-07-09 11:02 UTC (permalink / raw) To: NeilBrown; +Cc: linux-raid Hello, thanks for the hint. I do a backup with dd before that, I hope I can get back the data of the raid. The following is in the syslog: Jul 8 19:21:15 p3 kernel: Buffer I/O error on device dm-1, logical block 365625856 Jul 8 19:21:15 p3 kernel: Buffer I/O error on device dm-1, logical block 365625856 Jul 8 19:21:15 p3 kernel: lost page write due to I/O error on dm-1 Jul 8 19:21:15 p3 kernel: lost page write due to I/O error on dm-1 Jul 8 19:21:15 p3 kernel: JBD: I/O error detected when updating journal superblock for dm-1. Jul 8 19:21:15 p3 kernel: JBD: I/O error detected when updating journal superblock for dm-1. Jul 8 19:21:15 p3 kernel: RAID conf printout: Jul 8 19:21:15 p3 kernel: RAID conf printout: Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 Jul 8 19:21:15 p3 kernel: disk 3, o:0, dev:sdd1 Jul 8 19:21:15 p3 kernel: disk 3, o:0, dev:sdd1 Jul 8 19:21:15 p3 kernel: RAID conf printout: Jul 8 19:21:15 p3 kernel: RAID conf printout: Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 Jul 8 19:21:15 p3 kernel: md: recovery of RAID array md0 Jul 8 19:21:15 p3 kernel: md: recovery of RAID array md0 Jul 8 19:21:15 p3 kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk. Jul 8 19:21:15 p3 kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk. Jul 8 19:21:15 p3 kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery. Jul 8 19:21:15 p3 kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery. Jul 8 19:21:15 p3 kernel: md: using 128k window, over a total of 1465126400k. Jul 8 19:21:15 p3 kernel: md: using 128k window, over a total of 1465126400k. Jul 8 19:21:15 p3 kernel: md: resuming recovery of md0 from checkpoint. Jul 8 19:21:15 p3 kernel: md: resuming recovery of md0 from checkpoint. I think the right order is sdf1 sde1 sdc1 sdd1, I am right? So I have to do: mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 missing /dev/sdd1 The question is: sould I also add --assume-clean Thanks! Dietrich Am 09.07.2012 02:12 schrieb "NeilBrown" <neilb@suse.de>: > > On Sun, 8 Jul 2012 21:05:02 +0200 Dietrich Heise <dh@dhde.de> wrote: > > > Hi, > > > > the following Problem, > > One of four drives has S.M.A.R.T. errors, so I removed it and > > replaced, with a new one. > > > > In the time the drive was rebuilding, one of the three left devices > > has an I/O error (sdd1) (sdc1 was the replaced drive an was syncing). > > > > Now the following happends (two drives are spare drives :( ) > > It looks like you tried to --add /dev/sdd1 back in after it failed, and mdadm > let new. Newer versions of mdadm will refuse as that is not a good thing to > do but it shouldn't stop you getting your data back. > > First thing to realise is that you could have data corruption. There is at > least one block in the array which cannot be recovered, possibly more. i.e. > any block on sdd1 which is bad, and any block at the same offset in sdc1. > These blocks may not be in files which would be lucky, or they may contain > important metadata which might mean you've lost lots of files. > > If you hadn't tried to --add /dev/sdd1 you could just force-assemble the > array back to degraded mode (without sdc1) and back up any critical data. > As sdd1 now thinks it is a spare you need to re-create the array instead: > > mdadm -S /dev/md1 > mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 /dev/sdd1 missing > or > mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 missing /dev/sdd1 > > depending on whether sdd1 as the 3rd or 4th device in the array - I cannot > tell from the output here. > > You should then be able to mount the array and backup stuff. > > You then want to use 'ddrescue' to copy sdd1 onto a device with no bad > blocks, and assemble the array using the device instead of sdd1. > > Finally, you can add the new spare (sdc1) to the array and it should rebuild > successfully - providing there are no bad blocks on sdf1 or sde1. > > I hope that makes sense. Do ask if anything is unclear. > > NeilBrown > > > > > > p3 disks # mdadm -D /dev/md1 > > /dev/md1: > > Version : 1.2 > > Creation Time : Mon Feb 28 19:57:56 2011 > > Raid Level : raid5 > > Used Dev Size : 1465126400 (1397.25 GiB 1500.29 GB) > > Raid Devices : 4 > > Total Devices : 4 > > Persistence : Superblock is persistent > > > > Update Time : Sun Jul 8 20:37:12 2012 > > State : active, FAILED, Not Started > > Active Devices : 2 > > Working Devices : 4 > > Failed Devices : 0 > > Spare Devices : 2 > > > > Layout : left-symmetric > > Chunk Size : 512K > > > > Name : p3:0 (local to host p3) > > UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > > Events : 121205 > > > > Number Major Minor RaidDevice State > > 0 8 81 0 active sync /dev/sdf1 > > 1 8 65 1 active sync /dev/sde1 > > 2 0 0 2 removed > > 3 0 0 3 removed > > > > 4 8 49 - spare /dev/sdd1 > > 5 8 33 - spare /dev/sdc1 > > > > here is more information: > > > > p3 disks # mdadm -E /dev/sdc1 > > /dev/sdc1: > > Magic : a92b4efc > > Version : 1.2 > > Feature Map : 0x0 > > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > > Name : p3:0 (local to host p3) > > Creation Time : Mon Feb 28 19:57:56 2011 > > Raid Level : raid5 > > Raid Devices : 4 > > > > Avail Dev Size : 2930275057 (1397.26 GiB 1500.30 GB) > > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > > Data Offset : 2048 sectors > > Super Offset : 8 sectors > > State : active > > Device UUID : caefb029:526187ef:2051f578:db2b82b7 > > > > Update Time : Sun Jul 8 20:37:12 2012 > > Checksum : 18e2bfe1 - correct > > Events : 121205 > > > > Layout : left-symmetric > > Chunk Size : 512K > > > > Device Role : spare > > Array State : AA.. ('A' == active, '.' == missing) > > p3 disks # mdadm -E /dev/sdd1 > > /dev/sdd1: > > Magic : a92b4efc > > Version : 1.2 > > Feature Map : 0x0 > > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > > Name : p3:0 (local to host p3) > > Creation Time : Mon Feb 28 19:57:56 2011 > > Raid Level : raid5 > > Raid Devices : 4 > > > > Avail Dev Size : 2930269954 (1397.26 GiB 1500.30 GB) > > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > > Data Offset : 2048 sectors > > Super Offset : 8 sectors > > State : active > > Device UUID : 4231e244:60e27ed4:eff405d0:2e615493 > > > > Update Time : Sun Jul 8 20:37:12 2012 > > Checksum : 4bec6e25 - correct > > Events : 0 > > > > Layout : left-symmetric > > Chunk Size : 512K > > > > Device Role : spare > > Array State : AA.. ('A' == active, '.' == missing) > > p3 disks # mdadm -E /dev/sde1 > > /dev/sde1: > > Magic : a92b4efc > > Version : 1.2 > > Feature Map : 0x0 > > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > > Name : p3:0 (local to host p3) > > Creation Time : Mon Feb 28 19:57:56 2011 > > Raid Level : raid5 > > Raid Devices : 4 > > > > Avail Dev Size : 2930253889 (1397.25 GiB 1500.29 GB) > > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > > Data Offset : 2048 sectors > > Super Offset : 8 sectors > > State : active > > Device UUID : 28b08f44:4cc24663:84d39337:94c35d67 > > > > Update Time : Sun Jul 8 20:37:12 2012 > > Checksum : 15faa8a1 - correct > > Events : 121205 > > > > Layout : left-symmetric > > Chunk Size : 512K > > > > Device Role : Active device 1 > > Array State : AA.. ('A' == active, '.' == missing) > > p3 disks # mdadm -E /dev/sdf1 > > /dev/sdf1: > > Magic : a92b4efc > > Version : 1.2 > > Feature Map : 0x0 > > Array UUID : 6d4ebfd4:491bcb50:d98d5e67:f226f362 > > Name : p3:0 (local to host p3) > > Creation Time : Mon Feb 28 19:57:56 2011 > > Raid Level : raid5 > > Raid Devices : 4 > > > > Avail Dev Size : 2930269954 (1397.26 GiB 1500.30 GB) > > Array Size : 8790758400 (4191.76 GiB 4500.87 GB) > > Used Dev Size : 2930252800 (1397.25 GiB 1500.29 GB) > > Data Offset : 2048 sectors > > Super Offset : 8 sectors > > State : active > > Device UUID : 78d5600a:91927758:f78a1cea:3bfa3f5b > > > > Update Time : Sun Jul 8 20:37:12 2012 > > Checksum : 7767cb10 - correct > > Events : 121205 > > > > Layout : left-symmetric > > Chunk Size : 512K > > > > Device Role : Active device 0 > > Array State : AA.. ('A' == active, '.' == missing) > > > > Is there a way to repair the raid? > > > > thanks! > > Dietrich > > -- > > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > > the body of a message to majordomo@vger.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html > ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: RAID5 faild while in degraded mode, need help 2012-07-09 11:02 ` Dietrich Heise @ 2012-07-09 23:02 ` NeilBrown 2012-07-11 18:50 ` Dietrich Heise 0 siblings, 1 reply; 5+ messages in thread From: NeilBrown @ 2012-07-09 23:02 UTC (permalink / raw) To: Dietrich Heise; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 3287 bytes --] On Mon, 9 Jul 2012 13:02:04 +0200 Dietrich Heise <dh@dhde.de> wrote: > Hello, > > thanks for the hint. > > I do a backup with dd before that, I hope I can get back the data of the raid. > > The following is in the syslog: > > Jul 8 19:21:15 p3 kernel: Buffer I/O error on device dm-1, logical > block 365625856 > Jul 8 19:21:15 p3 kernel: Buffer I/O error on device dm-1, logical > block 365625856 > Jul 8 19:21:15 p3 kernel: lost page write due to I/O error on dm-1 > Jul 8 19:21:15 p3 kernel: lost page write due to I/O error on dm-1 > Jul 8 19:21:15 p3 kernel: JBD: I/O error detected when updating > journal superblock for dm-1. > Jul 8 19:21:15 p3 kernel: JBD: I/O error detected when updating > journal superblock for dm-1. > Jul 8 19:21:15 p3 kernel: RAID conf printout: > Jul 8 19:21:15 p3 kernel: RAID conf printout: > Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 > Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 > Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 > Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 > Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 > Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 > Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 > Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 > Jul 8 19:21:15 p3 kernel: disk 3, o:0, dev:sdd1 > Jul 8 19:21:15 p3 kernel: disk 3, o:0, dev:sdd1 > Jul 8 19:21:15 p3 kernel: RAID conf printout: > Jul 8 19:21:15 p3 kernel: RAID conf printout: > Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 > Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 > Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 > Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 > Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 > Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 > Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 > Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 > Jul 8 19:21:15 p3 kernel: md: recovery of RAID array md0 > Jul 8 19:21:15 p3 kernel: md: recovery of RAID array md0 > Jul 8 19:21:15 p3 kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk. > Jul 8 19:21:15 p3 kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk. > Jul 8 19:21:15 p3 kernel: md: using maximum available idle IO > bandwidth (but not more than 200000 KB/sec) for recovery. > Jul 8 19:21:15 p3 kernel: md: using maximum available idle IO > bandwidth (but not more than 200000 KB/sec) for recovery. > Jul 8 19:21:15 p3 kernel: md: using 128k window, over a total of 1465126400k. > Jul 8 19:21:15 p3 kernel: md: using 128k window, over a total of 1465126400k. > Jul 8 19:21:15 p3 kernel: md: resuming recovery of md0 from checkpoint. > Jul 8 19:21:15 p3 kernel: md: resuming recovery of md0 from checkpoint. > > I think the right order is sdf1 sde1 sdc1 sdd1, I am right? Yes, that looks right. > > So I have to do: > > mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 missing /dev/sdd1 > > The question is: sould I also add --assume-clean --assume-clean makes no difference to a degraded raid5 so it doesn't really matter. However I always suggest using --assume-clean when re-creating an array so on principle I would say "yes - you should add --assume-clean". NeilBrown > > Thanks! > Dietrich > [-- Attachment #2: signature.asc --] [-- Type: application/pgp-signature, Size: 828 bytes --] ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: RAID5 faild while in degraded mode, need help 2012-07-09 23:02 ` NeilBrown @ 2012-07-11 18:50 ` Dietrich Heise 0 siblings, 0 replies; 5+ messages in thread From: Dietrich Heise @ 2012-07-11 18:50 UTC (permalink / raw) To: NeilBrown; +Cc: linux-raid Hey, GREAT, it looks very good! Everything is there :) Thanks for help! Dietrich 2012/7/10 NeilBrown <neilb@suse.de>: > On Mon, 9 Jul 2012 13:02:04 +0200 Dietrich Heise <dh@dhde.de> wrote: > >> Hello, >> >> thanks for the hint. >> >> I do a backup with dd before that, I hope I can get back the data of the raid. >> >> The following is in the syslog: >> >> Jul 8 19:21:15 p3 kernel: Buffer I/O error on device dm-1, logical >> block 365625856 >> Jul 8 19:21:15 p3 kernel: Buffer I/O error on device dm-1, logical >> block 365625856 >> Jul 8 19:21:15 p3 kernel: lost page write due to I/O error on dm-1 >> Jul 8 19:21:15 p3 kernel: lost page write due to I/O error on dm-1 >> Jul 8 19:21:15 p3 kernel: JBD: I/O error detected when updating >> journal superblock for dm-1. >> Jul 8 19:21:15 p3 kernel: JBD: I/O error detected when updating >> journal superblock for dm-1. >> Jul 8 19:21:15 p3 kernel: RAID conf printout: >> Jul 8 19:21:15 p3 kernel: RAID conf printout: >> Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 >> Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 >> Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 >> Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 >> Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 >> Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 >> Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 >> Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 >> Jul 8 19:21:15 p3 kernel: disk 3, o:0, dev:sdd1 >> Jul 8 19:21:15 p3 kernel: disk 3, o:0, dev:sdd1 >> Jul 8 19:21:15 p3 kernel: RAID conf printout: >> Jul 8 19:21:15 p3 kernel: RAID conf printout: >> Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 >> Jul 8 19:21:15 p3 kernel: --- level:5 rd:4 wd:2 >> Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 >> Jul 8 19:21:15 p3 kernel: disk 0, o:1, dev:sdf1 >> Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 >> Jul 8 19:21:15 p3 kernel: disk 1, o:1, dev:sde1 >> Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 >> Jul 8 19:21:15 p3 kernel: disk 2, o:1, dev:sdc1 >> Jul 8 19:21:15 p3 kernel: md: recovery of RAID array md0 >> Jul 8 19:21:15 p3 kernel: md: recovery of RAID array md0 >> Jul 8 19:21:15 p3 kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk. >> Jul 8 19:21:15 p3 kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk. >> Jul 8 19:21:15 p3 kernel: md: using maximum available idle IO >> bandwidth (but not more than 200000 KB/sec) for recovery. >> Jul 8 19:21:15 p3 kernel: md: using maximum available idle IO >> bandwidth (but not more than 200000 KB/sec) for recovery. >> Jul 8 19:21:15 p3 kernel: md: using 128k window, over a total of 1465126400k. >> Jul 8 19:21:15 p3 kernel: md: using 128k window, over a total of 1465126400k. >> Jul 8 19:21:15 p3 kernel: md: resuming recovery of md0 from checkpoint. >> Jul 8 19:21:15 p3 kernel: md: resuming recovery of md0 from checkpoint. >> >> I think the right order is sdf1 sde1 sdc1 sdd1, I am right? > > Yes, that looks right. > >> >> So I have to do: >> >> mdadm -C /dev/md1 -l5 -n4 -e 1.2 -c 512 /dev/sdf1 /dev/sde1 missing /dev/sdd1 >> >> The question is: sould I also add --assume-clean > > --assume-clean makes no difference to a degraded raid5 so it doesn't really > matter. > However I always suggest using --assume-clean when re-creating an array so > on principle I would say "yes - you should add --assume-clean". > > NeilBrown > > >> >> Thanks! >> Dietrich >> ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2012-07-11 18:50 UTC | newest] Thread overview: 5+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2012-07-08 19:05 RAID5 faild while in degraded mode, need help Dietrich Heise 2012-07-09 0:12 ` NeilBrown 2012-07-09 11:02 ` Dietrich Heise 2012-07-09 23:02 ` NeilBrown 2012-07-11 18:50 ` Dietrich Heise
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).