From: "Stefan /*St0fF*/ Hübner" <stefan.huebner@stud.tu-ilmenau.de>
To: Ian Dall <ian@beware.dropbear.id.au>
Cc: Linux RAID <linux-raid@vger.kernel.org>
Subject: Re: Device kicked from raid too easilly
Date: Sat, 05 Jun 2010 09:22:38 +0200 [thread overview]
Message-ID: <4C09FB3E.1090302@stud.tu-ilmenau.de> (raw)
In-Reply-To: <1275702714.3740.55.camel@sibyl.beware.dropbear.id.au>
Hi Ian,
I do not think this is md-related, nor related to the other dropout
problem. Here we have a write-error, which correctly makes the disk
drop. So if you say the error is kind of a soft error, then actually
either the disks firmware or the scsi layer should be handling this.
But as always on write errors: mostly there are more than one write
request in the queue, so it's probably hard to find out which data
couldn't be written, or the data has already been discarded. Well,
write errors are those that should be handled in Firmware...
Stefan
P.S.: maybe you should check for firmware updates of the disks?
Am 05.06.2010 03:51, schrieb Ian Dall:
> I think this is different to the similarly titled long thread on SATA
> timeouts.
>
> I have an array of U320 scsi disks with similar characteristics from two
> different manufacturers.
>
> On two disks I see occasional scsi parity errors. I don't think this is
> a cabling or termination issue since I never see the parity errors on
> the other brand disks. smartctl shows a number of "non-medium errors"
> which I take to be the paroty errors.
>
> Now, when I have a raid 10 of these disks, the scsi parity error causes
> the first disk to be failed. The array then continues degraded with no
> apparent problems. If I read-add the failed disk, it always fails before
> the re-sync is complete. Eg:
>
> Jun 3 23:35:02 fs kernel: md: recovery of RAID array md5
> Jun 3 23:35:02 fs kernel: md: minimum _guaranteed_ speed: 1000 KB/sec/disk.
> Jun 3 23:35:02 fs kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for recovery.
> Jun 3 23:35:02 fs kernel: md: using 128k window, over a total of 29291904 blocks.
> Jun 3 23:35:07 fs kernel: sd 6:0:0:0: [sde] Result: hostbyte=DID_OK driverbyte=DRIVER_SENSE
> Jun 3 23:35:07 fs kernel: sd 6:0:0:0: [sde] Sense Key : Aborted Command [current]
> Jun 3 23:35:07 fs kernel: sd 6:0:0:0: [sde] Add. Sense: Scsi parity error
> Jun 3 23:35:07 fs kernel: sd 6:0:0:0: [sde] CDB: Write(10): 2a 00 00 05 9e 00 00 01 00 00
> Jun 3 23:35:07 fs kernel: end_request: I/O error, dev sde, sector 368128
> Jun 3 23:35:07 fs kernel: raid10: Disk failure on sde, disabling device.
> Jun 3 23:35:07 fs kernel: raid10: Operation continuing on 3 devices.
> Jun 3 23:35:07 fs kernel: md: md5: recovery done.
>
> Now I can test this disk in isolation (using iozone) pretty heavily and
> never see a problem. I can also use it in a raid0 and never see a
> problem.
>
> I think some of the strangeness is explained by the comment in the
> raid10 error handler: "else if it is the last working disks, ignore the
> error".
>
> Parity errors seem to me like they should be treated as transient
> errors. Maybe if there are multiple consecutive parity errors it could
> be assumed there is a hard fault in the transport layer. U320 uses
> "information units" with (stronger than parity) CRC checking. Although
> these errors are not reported as CRC errors that could just be a
> reporting issue (the lack of an "additional sense code qualifier").
> Given the complexity of the clock recovery de-skewing etc which goes on
> for U320, it is not surprising some disks would do it better than
> others, but a non zero error rate probably shouldn't be considered
> fatal.
>
> I don't really know where this should be fixed. Maybe the scsi layer
> should be retrying the scsi command, since it knows most about what sort
> of error it is. But equally it could be the responsibility of upper
> layers to do any retrying (which gives upper layers the option to not
> retry if they don't want to). But if the scsi layer is not responsible
> for retrying these sorts of errors, then the md layer is over-reacting
> by throwing disks out too easily.
>
>
> Regards,
> Ian
>
>
next prev parent reply other threads:[~2010-06-05 7:22 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-06-05 1:51 Device kicked from raid too easilly Ian Dall
2010-06-05 7:22 ` Stefan /*St0fF*/ Hübner [this message]
2010-06-08 5:15 ` Ian Dall
2010-06-08 5:56 ` Stefan /*St0fF*/ Hübner
2010-06-08 6:59 ` Tim Small
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4C09FB3E.1090302@stud.tu-ilmenau.de \
--to=stefan.huebner@stud.tu-ilmenau.de \
--cc=ian@beware.dropbear.id.au \
--cc=linux-raid@vger.kernel.org \
--cc=st0ff@npl.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).