* Interesting double failure
@ 2009-11-26 5:03 Tamas Vincze
2009-11-26 9:25 ` Justin Piszcz
0 siblings, 1 reply; 2+ messages in thread
From: Tamas Vincze @ 2009-11-26 5:03 UTC (permalink / raw)
To: linux-raid
sda3 and sdf3 are the two members of a RAID-1.
Both disks are the same model (300GB WD velociraptors), have the same
partition layout and are on two different controllers.
This is what happened yesterday: the same sector failed on both disks at the
same time. I have a feeling that it's a bug somewhere or perhaps a power
spike? The box is on a UPS.
Running CentOS 5.4 stock kernel: 2.6.18-164.6.1
Any thoughts?
Nov 24 08:08:12 sm kernel: sd 1:0:0:0: SCSI error: return code = 0x08000002
Nov 24 08:08:12 sm kernel: sdf: Current: sense key: Medium Error
Nov 24 08:08:12 sm kernel: Add. Sense: Record not found
Nov 24 08:08:12 sm kernel:
Nov 24 08:08:12 sm kernel: Info fld=0x1f2d8240
Nov 24 08:08:12 sm kernel: end_request: I/O error, dev sdf, sector 523076160
Nov 24 08:08:12 sm kernel: raid1: Disk failure on sdf3, disabling device.
Nov 24 08:08:12 sm kernel: Operation continuing on 1 devices
Nov 24 08:08:12 sm kernel: sd 0:0:0:0: SCSI error: return code = 0x08000002
Nov 24 08:08:12 sm kernel: sda: Current: sense key: Medium Error
Nov 24 08:08:12 sm kernel: Add. Sense: Record not found
Nov 24 08:08:12 sm kernel:
Nov 24 08:08:12 sm kernel: Info fld=0x1f2d8240
Nov 24 08:08:12 sm kernel: end_request: I/O error, dev sda, sector 523076160
Nov 24 08:08:12 sm kernel: RAID1 conf printout:
Nov 24 08:08:12 sm kernel: --- wd:1 rd:2
Nov 24 08:08:12 sm kernel: disk 0, wo:0, o:1, dev:sda3
Nov 24 08:08:12 sm kernel: disk 1, wo:1, o:0, dev:sdf3
Nov 24 08:08:12 sm kernel: RAID1 conf printout:
Nov 24 08:08:12 sm kernel: --- wd:1 rd:2
Nov 24 08:08:12 sm kernel: disk 0, wo:0, o:1, dev:sda3
Disk /dev/sdf: 300.0 GB, 300069052416 bytes
255 heads, 63 sectors/track, 36481 cylinders, total 586072368 sectors
Units = sectors of 1 * 512 = 512 bytes
Device Boot Start End Blocks Id System
/dev/sdf1 * 63 514079 257008+ fd Linux raid autodetect
/dev/sdf2 514080 1028159 257040 fd Linux raid autodetect
/dev/sdf3 1028160 523076399 261024120 fd Linux raid autodetect
/dev/sdf4 523076400 586067264 31495432+ fd Linux raid autodetect
The failed sector is near the end of the partition.
There's LVM on top of the raid and only the first 30GB or so is allocated,
so it wasn't any filesystem that made the request to that sector.
It must have been either the LVM or raid metadata, but I'm not sure where
they are allocated.
Oh, and it continued to operate on sda3 even though that also seem to have
failed.
-Tamas
^ permalink raw reply [flat|nested] 2+ messages in thread
* Re: Interesting double failure
2009-11-26 5:03 Interesting double failure Tamas Vincze
@ 2009-11-26 9:25 ` Justin Piszcz
0 siblings, 0 replies; 2+ messages in thread
From: Justin Piszcz @ 2009-11-26 9:25 UTC (permalink / raw)
To: Tamas Vincze; +Cc: linux-raid
On Thu, 26 Nov 2009, Tamas Vincze wrote:
> sda3 and sdf3 are the two members of a RAID-1.
> Both disks are the same model (300GB WD velociraptors), have the same
> partition layout and are on two different controllers.
> This is what happened yesterday: the same sector failed on both disks at the
> same time. I have a feeling that it's a bug somewhere or perhaps a power
> spike? The box is on a UPS.
> Running CentOS 5.4 stock kernel: 2.6.18-164.6.1
> Any thoughts?
>
> Nov 24 08:08:12 sm kernel: sd 1:0:0:0: SCSI error: return code = 0x08000002
> Nov 24 08:08:12 sm kernel: sdf: Current: sense key: Medium Error
> Nov 24 08:08:12 sm kernel: Add. Sense: Record not found
> Nov 24 08:08:12 sm kernel:
> Nov 24 08:08:12 sm kernel: Info fld=0x1f2d8240
> Nov 24 08:08:12 sm kernel: end_request: I/O error, dev sdf, sector 523076160
> Nov 24 08:08:12 sm kernel: raid1: Disk failure on sdf3, disabling device.
> Nov 24 08:08:12 sm kernel: Operation continuing on 1 devices
> Nov 24 08:08:12 sm kernel: sd 0:0:0:0: SCSI error: return code = 0x08000002
> Nov 24 08:08:12 sm kernel: sda: Current: sense key: Medium Error
> Nov 24 08:08:12 sm kernel: Add. Sense: Record not found
> Nov 24 08:08:12 sm kernel:
> Nov 24 08:08:12 sm kernel: Info fld=0x1f2d8240
> Nov 24 08:08:12 sm kernel: end_request: I/O error, dev sda, sector 523076160
> Nov 24 08:08:12 sm kernel: RAID1 conf printout:
> Nov 24 08:08:12 sm kernel: --- wd:1 rd:2
> Nov 24 08:08:12 sm kernel: disk 0, wo:0, o:1, dev:sda3
> Nov 24 08:08:12 sm kernel: disk 1, wo:1, o:0, dev:sdf3
> Nov 24 08:08:12 sm kernel: RAID1 conf printout:
> Nov 24 08:08:12 sm kernel: --- wd:1 rd:2
> Nov 24 08:08:12 sm kernel: disk 0, wo:0, o:1, dev:sda3
>
> Disk /dev/sdf: 300.0 GB, 300069052416 bytes
> 255 heads, 63 sectors/track, 36481 cylinders, total 586072368 sectors
> Units = sectors of 1 * 512 = 512 bytes
>
> Device Boot Start End Blocks Id System
> /dev/sdf1 * 63 514079 257008+ fd Linux raid autodetect
> /dev/sdf2 514080 1028159 257040 fd Linux raid autodetect
> /dev/sdf3 1028160 523076399 261024120 fd Linux raid autodetect
> /dev/sdf4 523076400 586067264 31495432+ fd Linux raid autodetect
>
> The failed sector is near the end of the partition.
> There's LVM on top of the raid and only the first 30GB or so is allocated,
> so it wasn't any filesystem that made the request to that sector.
> It must have been either the LVM or raid metadata, but I'm not sure where
> they are allocated.
>
> Oh, and it continued to operate on sda3 even though that also seem to have
> failed.
>
Hi,
Do not use velociraptors in RAID, they do not work, I have confirmed this
finding here:
http://forums.storagereview.net/index.php?showtopic=27303
In addition, just for the sake of it I tried a RAID with Windows 7 with
four-drives and the same problem happened.
What you are seeing is completely normal, these drives are broken for a
RAID configuration and WD will not replace them with other models.
Justin.
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2009-11-26 9:25 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-11-26 5:03 Interesting double failure Tamas Vincze
2009-11-26 9:25 ` Justin Piszcz
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).