* when does it become faulty disk @ 2005-06-18 11:47 Raz Ben-Jehuda(caro) 2005-06-19 19:10 ` Molle Bestefich 2005-06-21 1:53 ` More tales of horror from the linux (HW) raid crypt Harry Mangalam 0 siblings, 2 replies; 34+ messages in thread From: Raz Ben-Jehuda(caro) @ 2005-06-18 11:47 UTC (permalink / raw) To: linux-raid I have managed to make the kernel remove a disk from my raid even if this raid is "/" . I did it by adding a line in ata_scsi_error that remove the ata disk from the raid array. This means that when the first error ocurrs on a disk It is removed from the array. Well, this is not the best thing to do.. Question is : When does a disk become faulty ? Is when you have N errors in T time ? New ideas would be welcomed. -- Raz Long Live the Penguin ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: when does it become faulty disk 2005-06-18 11:47 when does it become faulty disk Raz Ben-Jehuda(caro) @ 2005-06-19 19:10 ` Molle Bestefich 2005-06-20 6:43 ` raz ben jehuda 2005-06-20 13:45 ` Michael Tokarev 2005-06-21 1:53 ` More tales of horror from the linux (HW) raid crypt Harry Mangalam 1 sibling, 2 replies; 34+ messages in thread From: Molle Bestefich @ 2005-06-19 19:10 UTC (permalink / raw) To: Raz Ben-Jehuda(caro); +Cc: linux-raid Raz Ben-Jehuda(caro) wrote: > I have managed to make the kernel remove a disk > from my raid even if this raid is "/" . I did it by adding a line > in ata_scsi_error that remove the ata disk from the raid array. > This means that when the first error ocurrs on a disk It is removed > from the array. > Well, this is not the best thing to do.. > Question is : > When does a disk become faulty ? When trying to read sectors from a disk and the disk fails the read: 1.) Read the data from the other disks in the RAID and 2.) Overwrite the sectors where the read error occur. If this write also fails, then the disk has used up it's spare sectors area. The RAID array is now by definition in a degraded state since that sector no longer exists in a redundant (readable at least) way. The disk should therefore be kicked, in order to notify the user that it should be replaced immediately. > Is when you have N errors in T time ? Nah, it's when you run out of spare sectors and your data redundancy is thereby lost that you have to fault the disk to prevent future data loss. Don't try to second-guess when the disk is going to get faulty based on how many errors occurred. If you want to do something like that, read out the SMART data from the disk. The manufacturer's data about the disks health should be your data source. > New ideas would be welcomed. HTH.. ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: when does it become faulty disk 2005-06-19 19:10 ` Molle Bestefich @ 2005-06-20 6:43 ` raz ben jehuda 2005-06-20 7:55 ` Molle Bestefich 2005-06-20 13:45 ` Michael Tokarev 1 sibling, 1 reply; 34+ messages in thread From: raz ben jehuda @ 2005-06-20 6:43 UTC (permalink / raw) To: Molle Bestefich; +Cc: linux-raid So, what about write errors ? from what you are saying i understand that when a write error occurs the disk is faulty. On Sun, 2005-06-19 at 22:10, Molle Bestefich wrote: > Raz Ben-Jehuda(caro) wrote: > > I have managed to make the kernel remove a disk > > from my raid even if this raid is "/" . I did it by adding a line > > in ata_scsi_error that remove the ata disk from the raid array. > > This means that when the first error ocurrs on a disk It is removed > > from the array. > > Well, this is not the best thing to do.. > > Question is : > > When does a disk become faulty ? > > When trying to read sectors from a disk and the disk fails the read: > 1.) Read the data from the other disks in the RAID and > 2.) Overwrite the sectors where the read error occur. > If this write also fails, then the disk has used up it's spare sectors > area. The RAID array is now by definition in a degraded state since > that sector no longer exists in a redundant (readable at least) way. > The disk should therefore be kicked, in order to notify the user that > it should be replaced immediately. > > Is when you have N errors in T time ? > > Nah, it's when you run out of spare sectors and your data redundancy > is thereby lost that you have to fault the disk to prevent future data > loss. Don't try to second-guess when the disk is going to get faulty > based on how many errors occurred. If you want to do something like > that, read out the SMART data from the disk. The manufacturer's data > about the disks health should be your data source. > > > New ideas would be welcomed. > > HTH.. > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Raz Long Live The Penguin ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: when does it become faulty disk 2005-06-20 6:43 ` raz ben jehuda @ 2005-06-20 7:55 ` Molle Bestefich 2005-06-20 10:09 ` raz ben jehuda 0 siblings, 1 reply; 34+ messages in thread From: Molle Bestefich @ 2005-06-20 7:55 UTC (permalink / raw) To: raz ben jehuda; +Cc: linux-raid raz ben jehuda wrote: > So, what about write errors ? > from what you are saying i understand that when a write error occurs > the disk is faulty. Yes.. If you are serious about your data redundancy, yes. A sector _read_ error is a notification from the disk that a sector has gone bad and that some particular data is lost. A sector _write_ error is the disk telling you that: 1. The sector has gone bad. 2. The disk failed to relocate the sector to the spare area, probably because it's full. The above are slight simplifications, since other kinds of read and write errors may in very rare cases occur. That's OK though, since you DO want sick disks with strange internal errors that are causing read or write errors to get kicked. In rare cases a disk could get sick in a way where writes to a bad sector succeeds but subsequent reads fail. Never seen it happen... But just in case, you might want to re-read a failed sector after you have written to it, just to check that the disk actually correctly relocated it. Once a disk has been kicked, you may want to instruct the user to check that the disk's spare sector count has indeed reached 0, by using smartctl -a /dev/xxx. That command will also tell of other disk failures. ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: when does it become faulty disk 2005-06-20 7:55 ` Molle Bestefich @ 2005-06-20 10:09 ` raz ben jehuda 0 siblings, 0 replies; 34+ messages in thread From: raz ben jehuda @ 2005-06-20 10:09 UTC (permalink / raw) To: Molle Bestefich; +Cc: linux-raid thank you molle. This was a very helpful information. On Mon, 2005-06-20 at 10:55, Molle Bestefich wrote: > raz ben jehuda wrote: > > So, what about write errors ? > > from what you are saying i understand that when a write error occurs > > the disk is faulty. > > Yes.. If you are serious about your data redundancy, yes. > > A sector _read_ error is a notification from the disk that a sector > has gone bad and that some particular data is lost. > > A sector _write_ error is the disk telling you that: > 1. The sector has gone bad. > 2. The disk failed to relocate the sector to the spare area, probably > because it's full. > > The above are slight simplifications, since other kinds of read and > write errors may in very rare cases occur. That's OK though, since > you DO want sick disks with strange internal errors that are causing > read or write errors to get kicked. > > In rare cases a disk could get sick in a way where writes to a bad > sector succeeds but subsequent reads fail. Never seen it happen... > But just in case, you might want to re-read a failed sector after you > have written to it, just to check that the disk actually correctly > relocated it. > > Once a disk has been kicked, you may want to instruct the user to > check that the disk's spare sector count has indeed reached 0, by > using smartctl -a /dev/xxx. That command will also tell of other disk > failures. -- Raz Long Live The Penguin ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: when does it become faulty disk 2005-06-19 19:10 ` Molle Bestefich 2005-06-20 6:43 ` raz ben jehuda @ 2005-06-20 13:45 ` Michael Tokarev 2005-06-20 15:35 ` raz ben jehuda 1 sibling, 1 reply; 34+ messages in thread From: Michael Tokarev @ 2005-06-20 13:45 UTC (permalink / raw) To: Molle Bestefich; +Cc: Raz Ben-Jehuda(caro), linux-raid Molle Bestefich wrote: > Raz Ben-Jehuda(caro) wrote: > >>I have managed to make the kernel remove a disk >>from my raid even if this raid is "/" . I did it by adding a line >>in ata_scsi_error that remove the ata disk from the raid array. >>This means that when the first error ocurrs on a disk It is removed >>from the array. >>Well, this is not the best thing to do.. >>Question is : >>When does a disk become faulty ? > > > When trying to read sectors from a disk and the disk fails the read: > 1.) Read the data from the other disks in the RAID and > 2.) Overwrite the sectors where the read error occur. Note: this is NOT how current linux softraid code works, it's how it *supposed* to work. And right now, linux raid code kicks a drive out of the array after *any* error (read or write), without trying to "understand" what happened. /mjt ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: when does it become faulty disk 2005-06-20 13:45 ` Michael Tokarev @ 2005-06-20 15:35 ` raz ben jehuda 0 siblings, 0 replies; 34+ messages in thread From: raz ben jehuda @ 2005-06-20 15:35 UTC (permalink / raw) To: Michael Tokarev; +Cc: Molle Bestefich, Raz Ben-Jehuda(caro), linux-raid On Mon, 2005-06-20 at 16:45, Michael Tokarev wrote: > Molle Bestefich wrote: > > Raz Ben-Jehuda(caro) wrote: > > > >>I have managed to make the kernel remove a disk > >>from my raid even if this raid is "/" . I did it by adding a line > >>in ata_scsi_error that remove the ata disk from the raid array. > >>This means that when the first error ocurrs on a disk It is removed > >>from the array. > >>Well, this is not the best thing to do.. > >>Question is : > >>When does a disk become faulty ? > > > > > > When trying to read sectors from a disk and the disk fails the read: > > 1.) Read the data from the other disks in the RAID and > > 2.) Overwrite the sectors where the read error occur. > > Note: this is NOT how current linux softraid code works, it's > how it *supposed* to work. And right now, linux raid code > kicks a drive out of the array after *any* error (read or > write), without trying to "understand" what happened. no accurate. I have raid1 over "/" and whenever I pull out one of the disks the system hangs. in order to fix it I added to ata_scsi_error() ata_port_disable(). > /mjt > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Raz Long Live The Penguin ^ permalink raw reply [flat|nested] 34+ messages in thread
* More tales of horror from the linux (HW) raid crypt 2005-06-18 11:47 when does it become faulty disk Raz Ben-Jehuda(caro) 2005-06-19 19:10 ` Molle Bestefich @ 2005-06-21 1:53 ` Harry Mangalam 2005-06-22 19:33 ` Mike Hardy 1 sibling, 1 reply; 34+ messages in thread From: Harry Mangalam @ 2005-06-21 1:53 UTC (permalink / raw) To: linux-raid Hi All, From the traffic, this list seems to be heavily slanted towards the SW aspect of Linux RAID, but there have been a few postings (other than mine) about the HW aspects of it. So, apologies for the verbarea on the HW aspects, but at least a few people have told me that this running monologue of raid troubles has been useful, so herein, some more. If I'm reiterating what is part of a FAQ, please let me know, but I read a lot of them and didn't stumble across much of this. short version: test ALL your disks before you use them, especially in a RAID set, especially the 'recertified' ones. long version: An FYI for anyone who uses disks in their computers... and definitely for those thinking of setting up either software or hardware raids I refer you to previous posts to this list about the detailed background, but it's briefly alluded to at the end of this list - more failures of the HW RAID on a dual opteron running ubuntu linux amd64-k8-SMP. I pulled the 3ware controller and used the onboard Silicon Image controller to run diagnostic SMART tests on all the 'recertified' SD-series disks that came back from WD. It's probably possible to use the 3ware controller (someone of the linux raid list indicated he did so), but I wanted the 3ware controller out of the loop because we suspected it as well. I used the linux command 'smartctl -t long /dev/sdx (where x=a-d). smartctl is part of the smartmontools package and can be used to test SATA disks as well as PATA disks (altho I've been told that the kernel has to be patched to do this - I'm using a Ubuntu-supplied kernel which works out-of-the-box). The long test lasts about 90 minutes for a 250GB disk and can be performed in parallel on each disk. 5 (FIVE!) of them (out of 9 returned from WD) failed that test - either they already had logged errors (SMART devices store their last 5 errors to their onboard memory), or they failed the short test (~2 minutes) or they failed the long test with unrecoverable errors. They're on their way back to WD for other replacements However, these are a superset of the disks that the 3ware controller failed on when being used for an array (see the message below) - I now think that the problem is either with the power supply (possible, but unlikely) or the disks (definitely), as well as the hotswap cages (definitely). I'm pretty sure that the controller is fine - it's been running with 5 disks in RAID5 for several days now with no errors or warnings at all. That makes me extremely suspicious of WD's 'recertified' drives, but that's the only avenue we have to get replacements right now. And I'll be dang sure to test ALL of them before I store data on them. I do have to reiterate, now that I've been running bonnie++ on both the SW RAID5 (on 4 disks - all that the onboard controller would control) and on the 3ware-controlled RAID5 that the SW is slightly faster and actually seems to use about as much CPU time as the 3ware in these tests. It's also more flexible in terms of how you set up and partition the devices. It's also so MUCH cheaper - using the onboard SATA controller and a $20 4 port SATA controller, I could control the same number of disks as the 3ware (8) but the 3ware costs $500. The big advantage of the 3ware controller is (relative) simplicity. Plug in controller, plug in disks, hit power switch, go into 3ware BIOS, allocate disks to a RAID unit, boot the OS, make a filesystem on the /dev/sdx and mount it. You can set/get some basic configuration and information from the 3ware utilities but not to the extent that you can with mdadm and the related utils. post below is to a 3ware support tech. > > As a reminder of the system, it's currently > > - dual opteron IWILL DK8X Mobo, Gb ethernet > > - Silicon image 4port SATA controller onboard (now disabled), > > - 3ware 9500S 8 port card running 8x250GB WD 2500SD disks in RAID5. > > - Disks are in 2x 4slot Chenbro hotswap RAID cages. > > - running kubuntu Linux in pure 64bit mode (altho bringing up KDE > > currently locks the system in some configurations) > > - using kernel image 2.6.11-1-amd64-k8-smp as a kubuntu debian install > > (NOT custom-compiled) > > - OS is running from a separate WD 200GB IDE disk > > (which recently bonked at 3 months old, replaced by WD without > > complaint.) - on an APC UPS (runnning apcupsd communicating thru a usb > > cable) > > > > The 9500 that you sent to me was put into service as soon as we got > > enough SD disks to make a raid5 - 3 of them, on ports 0-2, in the 1st > > hotswap cage. > > > > During that time, the array stayed up and seemed to be stable over about > > 1 week of heavy testing. Once we got all the disks replace with SD > > disks, I set it up as 8 disks in a RAID5 and things seemed to be fine for > > about a day. Then, the disk on port 3 had problems. I replaced it and > > again it appeared to go bad. I then disconnected it from the hotswap > > cage and connected it directly to the controller. That seemed to solve > > that problem, so there definitely is a problem with one hotswap cage - > > it's being replaced. > > > > > > However after that incident, there have been 3 more with disks on the > > other hotswap cage, on different ports, one of port 6 (4 warnings of: > > Sector repair completed: port=6, LBA=0x622CE39, and then the error: > > Degraded unit detected: unit=0, port=6. I wasn't sure if it was a > > seating error or a real disk error, so I pulled the disk and re-seated it > > (and the controller accepted it fine) but then after it rebuilt the > > array, it failed again on that port. OK, I replaced the drive. Then > > port 7 reported:(0x04:0x0023): Sector repair completed: port=7, > > LBA=0x2062320 > > > > I started a series of copy and read/write tests to make sure the array > > was stable under load, and then just as the array filled up, it failed > > again, this time again on port 3: (0x04:0x0002): Degraded unit detected: > > unit=0, port=3 (this port is connected directly to the controller). > > > > And this morning, I saw that yet another drive looks like it has failed > > or at least is unresponsive:(0x04:0x0009): Drive timeout detected: port=5 > > > > Discounting the incidents that seem to be related to the bad hotswap > > cage, that's still 4 disks (with MTBF of 1Mhr) that have gone bad in 2 > > days. > > > > I then connected all the disks directly to the controller to remove all > > hotswap cage influence, and the disk on port 3 almost immediately was > > marked bad - I have to say that this again sounds like a controller > > problem. An amazing statistical convergence of random disk failures? > > Electrical failure (the system is on a relatively good APC UPS (SMART UPS > > 1000), so the voltage supply should be good, and no other problems have > > been seen. I guess I could throw the Power supply on a scope to see if > > it's stable, but there have been no other such glitches (unless it's an > > uneven power supply that is causing the disks to die). > > > > Currently most of the disks that were marked bad by the 3ware controller > > are being tested under the onboard silicon image controller in a raid5 > > config. I'll test over the weekend to see what they do. At this point I tested the disks using the smartctl, and found out the bad ones. The SW RAID stayed up without errors until I brought it down to install the 3ware controller. -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> _______________________________________________ List-Info: https://maillists.uci.edu/mailman/listinfo/uci-linux -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-21 1:53 ` More tales of horror from the linux (HW) raid crypt Harry Mangalam @ 2005-06-22 19:33 ` Mike Hardy 2005-06-22 20:16 ` Harry Mangalam 2005-06-22 20:54 ` Dan Stromberg 0 siblings, 2 replies; 34+ messages in thread From: Mike Hardy @ 2005-06-22 19:33 UTC (permalink / raw) To: Harry Mangalam, linux-raid Not that I wish bad luck on anyone, but I always enjoy reading about problems others have had and how they solved them, so I don't have to learn the hard way. Hopefully your situation gets sorted out. I can only second (and third, etc) the motion to do SMART tests on all drives before putting them in service, and add that you should really do a short test daily and a long test at least weekly if possibly. Basically you just can't trust consumer drives at all these days. smartmontools and rsync are probably my most-loved open source packages these days. I usually get 1 out of 10 bad out of the box now (some remappable at least) and a handful then fail quickly too. Most of them haven't gone three years so I can't say if they fail completely, but they seem to be lasting ok with occasional bad blocks. I'm very interested in the relative SW raid / HW raid performance. I have both in service (two raid 5 sets are actually the same size with the same number of components) and see roughly the same as you mention. One difference that I see is that HW raid should generate fewer interrupts and lower bus traffic. In the one area I used HW raid (a 3ware 8 port PATA, 8x250GB Maxtor, 2xOpteron), it was specifically because the motherboard chipset (or its interaction with Linux at least) was crap, and couldn't handle the interrupt load under bonnie++. So this could be a factor. It also goes to show that burning the machine in (with bonnie++ or similar) is a very good step. At least you catch these things before they're in service... Anyway, good luck with the new drives. -Mike Harry Mangalam wrote: > Hi All, > >>From the traffic, this list seems to be heavily slanted towards the SW aspect > of Linux RAID, but there have been a few postings (other than mine) about the > HW aspects of it. So, apologies for the verbarea on the HW aspects, but at > least a few people have told me that this running monologue of raid troubles > has been useful, so herein, some more. If I'm reiterating what is part of a > FAQ, please let me know, but I read a lot of them and didn't stumble across > much of this. > > > short version: test ALL your disks before you use them, especially in a RAID > set, especially the 'recertified' ones. ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 19:33 ` Mike Hardy @ 2005-06-22 20:16 ` Harry Mangalam 2005-06-22 20:38 ` Jure Pecar ` (2 more replies) 2005-06-22 20:54 ` Dan Stromberg 1 sibling, 3 replies; 34+ messages in thread From: Harry Mangalam @ 2005-06-22 20:16 UTC (permalink / raw) To: linux-raid Thanks for the note - I'd also be interested in what drives people people use that seem to be robust and which are having problems. The 3ware tech that I spoke to recommended the WD-SD series, but a local senior sysadmin has warned his entire dept off WD drives for the reasons I've mentioned. He rec'ed Seagate (5 year warr) and Hitachi. So far the only good thing that I can say about WD is that their return policy has been relatively smooth (tho they smoothly keep returning bad drives :( ) I personally buy IBM (now Hitachi I think) and Seagate and have always avoided WD as being cheap and cheesy. 5 years ago when I was a sysadmin, I was replacing WD at probably 5 times the rate of anything else. It was probably strongly influenced by the fact that more people were using them than the others but I don't think I've ever had an IBM or seagate 3.5" die (have a teetering stack of working ones that just got too small to bother with). I figured that since WD was still in business (tho seeming always on the verge of going bust) their quality must have improved, but maybe not. Happy to be contradicted. If there's a 'Consumer's Reports' for HD reliability on the web, I couldn't find it. Maybe it's time to set one up - with SMART disk records, it's possible to get more reliable runtime records and data than just the usual single point or small data set that individuals can provide (and you can bet that most vendors aren't going to make their failures public..). Perhaps something like the CDDB, where a you can run an applet on a cronjob that will upload your disk's SMART data cache to a remote DB on a regular basis, so the millions of disks out there can be profiled. Hmmm - sounds like a good undergrad project. Did someone say Summer of Code? (http://code.google.com/summerofcode.html) hjm On Wednesday 22 June 2005 12:33 pm, Mike Hardy wrote: > Not that I wish bad luck on anyone, but I always enjoy reading about > problems others have had and how they solved them, so I don't have to > learn the hard way. Hopefully your situation gets sorted out. -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 20:16 ` Harry Mangalam @ 2005-06-22 20:38 ` Jure Pecar 2005-06-22 21:33 ` Harry Mangalam 2005-06-22 21:09 ` Brad Dameron 2005-06-22 23:54 ` Jon Lewis 2 siblings, 1 reply; 34+ messages in thread From: Jure Pecar @ 2005-06-22 20:38 UTC (permalink / raw) To: linux-raid On Wed, 22 Jun 2005 13:16:33 -0700 Harry Mangalam <hjm@tacgi.com> wrote: > Perhaps something like the CDDB, where a you can run an applet on a > cronjob that will upload your disk's SMART data cache to a remote DB on a > regular basis, so the millions of disks out there can be profiled. Hmmm > - sounds like a good undergrad project. Did someone say Summer of Code? > (http://code.google.com/summerofcode.html) Actually if you look at the storagereview.com, they already have a crude form of this online under "reliability survey". You have to login, then you can enter disk model that you have expirience with and what were those expiriences. There's quite some info already available ... -- Jure Pečar http://jure.pecar.org/ - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 20:38 ` Jure Pecar @ 2005-06-22 21:33 ` Harry Mangalam 2005-06-22 23:15 ` SMART, was " Konstantin Olchanski 0 siblings, 1 reply; 34+ messages in thread From: Harry Mangalam @ 2005-06-22 21:33 UTC (permalink / raw) To: linux-raid Is this too OT? Let me know.. The Storage Review DB would be a decent 1st approximation (includes data on ~35K drives), but relies on entering the info on only drives that the DB knows about, and also relies on info that the person entering the data remembers - how many hours in use, etc. It's also available only to registered users who might be entering spurious data to gain access to the DB. The SMART data (granted, it would only be possible to collect data from SMART drives) would tell you considerably more info: (including that I was mistaken when I previously said I only bought IBM's and Seagates) :) This is the kind of info that wold be quite useful to have available in a CDDB-like DB. It would also point out which drives are susceptible to failure under continuous-on conditions versus frequent power-cycling, etc. an example from my home system: 1062 $ smartctl -a /dev/hda smartctl version 5.32 Copyright (C) 2002-4 Bruce Allen Home page is http://smartmontools.sourceforge.net/ === START OF INFORMATION SECTION === Device Model: WDC WD800JB-00CRA1 Serial Number: WD-WMA8E4088773 Firmware Version: 17.07W17 Device is: In smartctl database [for details use: -P show] ATA Version is: 5 ATA Standard is: Exact ATA specification draft version not indicated Local Time is: Wed Jun 22 14:24:56 2005 PDT SMART support is: Available - device has SMART capability. SMART support is: Enabled === START OF READ SMART DATA SECTION === SMART overall-health self-assessment test result: PASSED General SMART Values: Offline data collection status: (0x84) Offline data collection activity was suspended by an interrupting command from host. Auto Offline Data Collection: Enabled. Self-test execution status: ( 0) The previous self-test routine completed without error or no self-test has ever been run. Total time to complete Offline data collection: (3120) seconds. Offline data collection capabilities: (0x3b) SMART execute Offline immediate. Auto Offline data collection on/off support. Suspend Offline collection upon new command. Offline surface scan supported. Self-test supported. Conveyance Self-test supported. No Selective Self-test supported. SMART capabilities: (0x0003) Saves SMART data before entering power-saving mode. Supports SMART auto save timer. Error logging capability: (0x01) Error logging supported. No General Purpose Logging support. Short self-test routine recommended polling time: ( 2) minutes. Extended self-test routine recommended polling time: ( 58) minutes. Conveyance self-test routine recommended polling time: ( 5) minutes. SMART Attributes Data Structure revision number: 16 Vendor Specific SMART Attributes with Thresholds: ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE 1 Raw_Read_Error_Rate 0x000b 200 200 051 Pre-fail Always - 0 3 Spin_Up_Time 0x0007 099 095 021 Pre-fail Always - 4141 4 Start_Stop_Count 0x0032 100 100 040 Old_age Always - 122 5 Reallocated_Sector_Ct 0x0033 199 199 140 Pre-fail Always - 2 7 Seek_Error_Rate 0x000b 200 200 051 Pre-fail Always - 0 9 Power_On_Hours 0x0032 075 075 000 Old_age Always - 18940 10 Spin_Retry_Count 0x0013 100 100 051 Pre-fail Always - 0 11 Calibration_Retry_Count 0x0013 100 100 051 Pre-fail Always - 0 12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always - 120 196 Reallocated_Event_Count 0x0032 198 198 000 Old_age Always - 2 197 Current_Pending_Sector 0x0012 200 200 000 Old_age Always - 0 198 Offline_Uncorrectable 0x0012 200 200 000 Old_age Always - 0 199 UDMA_CRC_Error_Count 0x000a 200 253 000 Old_age Always - 0 200 Multi_Zone_Error_Rate 0x0009 200 200 051 Pre-fail Offline - 0 SMART Error Log Version: 1 No Errors Logged On Wednesday 22 June 2005 1:38 pm, Jure Pecar wrote: > On Wed, 22 Jun 2005 13:16:33 -0700 > > Harry Mangalam <hjm@tacgi.com> wrote: > > Perhaps something like the CDDB, where a you can run an applet on a > > cronjob that will upload your disk's SMART data cache to a remote DB on > > a regular basis, so the millions of disks out there can be profiled. > > Hmmm - sounds like a good undergrad project. Did someone say Summer of > > Code? (http://code.google.com/summerofcode.html) > > Actually if you look at the storagereview.com, they already have a crude > form of this online under "reliability survey". You have to login, then you > can enter disk model that you have expirience with and what were those > expiriences. There's quite some info already available ... -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> ^ permalink raw reply [flat|nested] 34+ messages in thread
* SMART, was Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 21:33 ` Harry Mangalam @ 2005-06-22 23:15 ` Konstantin Olchanski 2005-06-22 23:32 ` Harry Mangalam 2005-06-22 23:35 ` Mike Hardy 0 siblings, 2 replies; 34+ messages in thread From: Konstantin Olchanski @ 2005-06-22 23:15 UTC (permalink / raw) To: Harry Mangalam; +Cc: linux-raid On Wed, Jun 22, 2005 at 02:33:00PM -0700, Harry Mangalam wrote: > The SMART data ... would tell you considerably more info: > an example from my home system: > 1062 $ smartctl -a /dev/hda ... But does anybody know what all these SMART parameters *mean* ?!? What is this "Raw_Read_Error_Rate"? What is this "Seek_Error_Rate"? AFAIK, only a handful of parameters ("temperature", "reallocated sectors", and "start stop count") seem to measure something. All the other stuff appears to be complete gibberish. Then for the SMART "status", it is completely useless- I had SMART complain about perfectly good disks ("this disk will fail soon!!!"- a few 120GB WDs that sometimes stop seeking, probably through overheating) and I had completely unusable disks (i.e. develops about 10 unreadable sectors per day) that SMART was perfectly happy with. K.O. > Vendor Specific SMART Attributes with Thresholds: > ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED > WHEN_FAILED RAW_VALUE > 1 Raw_Read_Error_Rate 0x000b 200 200 051 Pre-fail Always > - 0 > 3 Spin_Up_Time 0x0007 099 095 021 Pre-fail Always > - 4141 > 4 Start_Stop_Count 0x0032 100 100 040 Old_age Always > - 122 > 5 Reallocated_Sector_Ct 0x0033 199 199 140 Pre-fail Always > - 2 > 7 Seek_Error_Rate 0x000b 200 200 051 Pre-fail Always > - 0 > 9 Power_On_Hours 0x0032 075 075 000 Old_age Always > - 18940 > 10 Spin_Retry_Count 0x0013 100 100 051 Pre-fail Always > - 0 > 11 Calibration_Retry_Count 0x0013 100 100 051 Pre-fail Always > - 0 > 12 Power_Cycle_Count 0x0032 100 100 000 Old_age Always > - 120 > 196 Reallocated_Event_Count 0x0032 198 198 000 Old_age Always > - 2 > 197 Current_Pending_Sector 0x0012 200 200 000 Old_age Always > - 0 > 198 Offline_Uncorrectable 0x0012 200 200 000 Old_age Always > - 0 > 199 UDMA_CRC_Error_Count 0x000a 200 253 000 Old_age Always > - 0 > 200 Multi_Zone_Error_Rate 0x0009 200 200 051 Pre-fail Offline > - 0 > > SMART Error Log Version: 1 > No Errors Logged > > > On Wednesday 22 June 2005 1:38 pm, Jure Pecar wrote: > > On Wed, 22 Jun 2005 13:16:33 -0700 > > > > Harry Mangalam <hjm@tacgi.com> wrote: > > > Perhaps something like the CDDB, where a you can run an applet on a > > > cronjob that will upload your disk's SMART data cache to a remote DB on > > > a regular basis, so the millions of disks out there can be profiled. > > > Hmmm - sounds like a good undergrad project. Did someone say Summer of > > > Code? (http://code.google.com/summerofcode.html) > > > > Actually if you look at the storagereview.com, they already have a crude > > form of this online under "reliability survey". You have to login, then you > > can enter disk model that you have expirience with and what were those > > expiriences. There's quite some info already available ... > > -- > Cheers, Harry > Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com > <<plain text preferred>> > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Konstantin Olchanski Data Acquisition Systems: The Bytes Must Flow! Email: olchansk-at-triumf-dot-ca Snail mail: 4004 Wesbrook Mall, TRIUMF, Vancouver, B.C., V6T 2A3, Canada ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: SMART, was Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 23:15 ` SMART, was " Konstantin Olchanski @ 2005-06-22 23:32 ` Harry Mangalam 2005-06-22 23:35 ` Mike Hardy 1 sibling, 0 replies; 34+ messages in thread From: Harry Mangalam @ 2005-06-22 23:32 UTC (permalink / raw) To: Konstantin Olchanski; +Cc: linux-raid On Wednesday 22 June 2005 4:15 pm, Konstantin Olchanski wrote: > But does anybody know what all these SMART parameters *mean* ?!? Yes, apparently, it's all written down: http://smartmontools.sourceforge.net/#references > AFAIK, only a handful of parameters ("temperature", "reallocated sectors", > and "start stop count") seem to measure something. All the other > stuff appears to be complete gibberish. Yup, I agree. But the information that isn't useful doesn't negate that which is. And the gibberish may indeed be useful once it is interpreted. > Then for the SMART "status", it is completely useless- I had SMART > complain about perfectly good disks ("this disk will fail soon!!!"- a few > 120GB WDs that sometimes stop seeking, probably through overheating) > and I had completely unusable disks (i.e. develops about 10 unreadable > sectors per day) that SMART was perfectly happy with. This I can't address - I've not had this experience - others can speak up about this. > K.O. (and thanks again for your help on my original problem!) -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: SMART, was Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 23:15 ` SMART, was " Konstantin Olchanski 2005-06-22 23:32 ` Harry Mangalam @ 2005-06-22 23:35 ` Mike Hardy 1 sibling, 0 replies; 34+ messages in thread From: Mike Hardy @ 2005-06-22 23:35 UTC (permalink / raw) To: Konstantin Olchanski; +Cc: Harry Mangalam, linux-raid Konstantin Olchanski wrote: > But does anybody know what all these SMART parameters *mean* ?!? > > What is this "Raw_Read_Error_Rate"? > What is this "Seek_Error_Rate"? Most of them seem to make sense. I'd guess raw read error rate is the number of read errors corrected by the firmware in some way (recalibration, ecc, or something). Seek error rate is probably a calibration thing. The last time I googled one, I got a pretty good answer. This was the site I saw: http://www.ariolic.com/activesmart/smart-attributes/soft-read-error-rate.html > Then for the SMART "status", it is completely useless- I had SMART > complain about perfectly good disks ("this disk will fail soon!!!"- a few > 120GB WDs that sometimes stop seeking, probably through overheating) > and I had completely unusable disks (i.e. develops about 10 unreadable > sectors per day) that SMART was perfectly happy with. Well, I guess anecdotal evidence is just that. At least for the mostly Maxtor drives I run (which do fail from time to time, alas), the SMART monitoring has been dead on. Every time. If there's a bad block, SMART sees it. Right before the drive goes belly-up mysteriously, I'll get lots of soft read error rate triggers or TA increase count triggers etc. Its gotten to where if its not a bad block, I should just about hook it up with an automated dealtime.com query to just buy a drive. I imagine most folks are somewhere in between. I'm never going to run a drive without smartd watching it again though, that's for sure. -Mike ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 20:16 ` Harry Mangalam 2005-06-22 20:38 ` Jure Pecar @ 2005-06-22 21:09 ` Brad Dameron 2005-06-22 21:43 ` Harry Mangalam 2005-06-22 23:54 ` Jon Lewis 2 siblings, 1 reply; 34+ messages in thread From: Brad Dameron @ 2005-06-22 21:09 UTC (permalink / raw) To: linux-raid On Wed, 2005-06-22 at 13:16 -0700, Harry Mangalam wrote: > Thanks for the note - I'd also be interested in what drives people people use > that seem to be robust and which are having problems. The 3ware tech that I > spoke to recommended the WD-SD series, but a local senior sysadmin has warned > his entire dept off WD drives for the reasons I've mentioned. He rec'ed > Seagate (5 year warr) and Hitachi. So far the only good thing that I can say > about WD is that their return policy has been relatively smooth (tho they > smoothly keep returning bad drives :( ) > > I personally buy IBM (now Hitachi I think) and Seagate and have always avoided > WD as being cheap and cheesy. 5 years ago when I was a sysadmin, I was > replacing WD at probably 5 times the rate of anything else. It was probably > strongly influenced by the fact that more people were using them than the > others but I don't think I've ever had an IBM or seagate 3.5" die (have a > teetering stack of working ones that just got too small to bother with). I > figured that since WD was still in business (tho seeming always on the verge > of going bust) their quality must have improved, but maybe not. > > Happy to be contradicted. If there's a 'Consumer's Reports' for HD > reliability on the web, I couldn't find it. Maybe it's time to set one up - > with SMART disk records, it's possible to get more reliable runtime records > and data than just the usual single point or small data set that individuals > can provide (and you can bet that most vendors aren't going to make their > failures public..). > > Perhaps something like the CDDB, where a you can run an applet on a cronjob > that will upload your disk's SMART data cache to a remote DB on a regular > basis, so the millions of disks out there can be profiled. Hmmm - sounds > like a good undergrad project. Did someone say Summer of Code? > (http://code.google.com/summerofcode.html) > > hjm Everyone has had issues with one brand or another. Sometimes the shipping can make all the difference. As I have had 4-5 out of 16 drives shipped come up bad out of the box. And these were Seagate 73GB SCSI drives. Not some cheap SATA/PATA drives. Now when you come to comparing drives you have to remember that not all manufacturers make their own drives. It is know that WD has sold IBM drives as theirs and vice-versa. I currently have a 12 drive 3ware 9500S-12 with 73GB WD 10k raptor drives in it that has been online for a year and a month without issues. And the controller has the very first firmware and linux drivers out. I also have several hundred Seagate SCSI 10k drives from 36gb to 146gb. And a few dozen Fujitsu 15k 73gb SCSI drives. And out of all these drives Seagate is the only ones I have to keep sending in for RMA. Now mind you I do have about 5 times as many drives. I used to not be partial to Maxtor either due to a issue several years ago with their PATA drives dieing on me in bulk. But I am currently purchasing 16 of their Maxline III SATA 300GB drives for a large storage array. Why am I going with them? Because they have the best reported performance for this type of drive. Not because they have a record of lasting longer. It is too hard to tell in a production/semi-production environment which brand will last the longest. So go with what you think is the best. And have spares/backups and monitoring programs to be prepared for when one does die. The odds of multiple failures at the same time are pretty high. And just for the record, look at the Areca SATA RAID cards. They have 16 and even 24 port cards if you want very large arrays. The cards greatly out perform the 3ware cards in READ/WRITE performance. They even offer RAID6. Just look at the processor on the ARC-12xx series. Just my 2 cents. Hope it helps. Brad Dameron SeaTab Software www.seatab.com ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 21:09 ` Brad Dameron @ 2005-06-22 21:43 ` Harry Mangalam 2005-06-22 22:00 ` Ming Zhang 0 siblings, 1 reply; 34+ messages in thread From: Harry Mangalam @ 2005-06-22 21:43 UTC (permalink / raw) To: linux-raid Good points and good info; I'll look into the areca products a bit more. Just yesterday I looked at the 16 tray Chenbro/Areca combo: http://www.chenbro.com.tw/Chenbro_Newsletter/Vol21_rm321.htm I'll read up on the Areca controllers. Thanks hjm On Wednesday 22 June 2005 2:09 pm, Brad Dameron wrote: > And just for the record, look at the Areca SATA RAID cards. They have 16 > and even 24 port cards if you want very large arrays. The cards greatly > out perform the 3ware cards in READ/WRITE performance. They even offer > RAID6. Just look at the processor on the ARC-12xx series. > > Just my 2 cents. Hope it helps. > > Brad Dameron > SeaTab Software > www.seatab.com > > > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 21:43 ` Harry Mangalam @ 2005-06-22 22:00 ` Ming Zhang 2005-06-22 22:11 ` John Madden ` (2 more replies) 0 siblings, 3 replies; 34+ messages in thread From: Ming Zhang @ 2005-06-22 22:00 UTC (permalink / raw) To: Harry Mangalam; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 1074 bytes --] will this 24 port card itself will be a bottleneck? ming On Wed, 2005-06-22 at 14:43 -0700, Harry Mangalam wrote: > Good points and good info; I'll look into the areca products a bit more. Just > yesterday I looked at the 16 tray Chenbro/Areca combo: > http://www.chenbro.com.tw/Chenbro_Newsletter/Vol21_rm321.htm > I'll read up on the Areca controllers. > Thanks > hjm > > On Wednesday 22 June 2005 2:09 pm, Brad Dameron wrote: > > And just for the record, look at the Areca SATA RAID cards. They have 16 > > and even 24 port cards if you want very large arrays. The cards greatly > > out perform the 3ware cards in READ/WRITE performance. They even offer > > RAID6. Just look at the processor on the ARC-12xx series. > > > > Just my 2 cents. Hope it helps. > > > > Brad Dameron > > SeaTab Software > > www.seatab.com > > > > > > - > > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > > the body of a message to majordomo@vger.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html > [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:00 ` Ming Zhang @ 2005-06-22 22:11 ` John Madden 2005-06-22 22:26 ` Ming Zhang 2005-06-23 0:20 ` bdameron 2005-06-22 22:45 ` Harry Mangalam 2005-06-23 0:14 ` bdameron 2 siblings, 2 replies; 34+ messages in thread From: John Madden @ 2005-06-22 22:11 UTC (permalink / raw) Cc: Harry Mangalam, linux-raid, Ming Zhang > will this 24 port card itself will be a bottleneck? Yes. The U320 SCSI bus on which it resides will be if nothing else. John -- John Madden UNIX Systems Engineer Ivy Tech Community College of Indiana jmadden@ivytech.edu ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:11 ` John Madden @ 2005-06-22 22:26 ` Ming Zhang 2005-06-23 0:20 ` bdameron 1 sibling, 0 replies; 34+ messages in thread From: Ming Zhang @ 2005-06-22 22:26 UTC (permalink / raw) To: John Madden; +Cc: Harry Mangalam, linux-raid [-- Attachment #1: Type: text/plain, Size: 270 bytes --] it is a sata card right? why u320 here? ming On Wed, 2005-06-22 at 17:11 -0500, John Madden wrote: > > will this 24 port card itself will be a bottleneck? > > Yes. The U320 SCSI bus on which it resides will be if nothing else. > > John > > > > [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:11 ` John Madden 2005-06-22 22:26 ` Ming Zhang @ 2005-06-23 0:20 ` bdameron 1 sibling, 0 replies; 34+ messages in thread From: bdameron @ 2005-06-23 0:20 UTC (permalink / raw) To: linux-raid There is no U320 SCSI bus. This is a SATA card. Each drive has it's own 150MBps port. The PCI-X speed of course limits it to 1.06GB per second. I doubt the drives will come close to that. Brad Dameron SeaTab Software www.seatab.com Quoting John Madden <jmadden@ivytech.edu>: > > will this 24 port card itself will be a bottleneck? > > Yes. The U320 SCSI bus on which it resides will be if nothing else. > > John > ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:00 ` Ming Zhang 2005-06-22 22:11 ` John Madden @ 2005-06-22 22:45 ` Harry Mangalam 2005-06-22 23:05 ` Ming Zhang 2005-06-23 0:25 ` bdameron 2005-06-23 0:14 ` bdameron 2 siblings, 2 replies; 34+ messages in thread From: Harry Mangalam @ 2005-06-22 22:45 UTC (permalink / raw) To: mingz; +Cc: linux-raid Also a good point - I was just discussing this in another thread. Areca makes a number of SATA RAID controllers - some are SCSI-SATA, some are PCI-X-SATA, some are PCIe-SATA. Areca supplied this URL to a relatively long review of recent SATA and SCSI controllers - it's worth a read. And surprisingly, the SATA controllers do pretty well against the SCSI controllers. http://www.tweakers.net/reviews/557/1 hjm On Wednesday 22 June 2005 3:00 pm, Ming Zhang wrote: > will this 24 port card itself will be a bottleneck? > > ming > -- Cheers, Harry Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com <<plain text preferred>> ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:45 ` Harry Mangalam @ 2005-06-22 23:05 ` Ming Zhang 2005-06-23 0:25 ` bdameron 1 sibling, 0 replies; 34+ messages in thread From: Ming Zhang @ 2005-06-22 23:05 UTC (permalink / raw) To: Harry Mangalam; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 764 bytes --] On Wed, 2005-06-22 at 15:45 -0700, Harry Mangalam wrote: > Also a good point - I was just discussing this in another thread. > Areca makes a number of SATA RAID controllers - some are SCSI-SATA, some are > PCI-X-SATA, some are PCIe-SATA. 24 port with 40MB/s per disk is 960MB/s, that is a huge number. :P > > Areca supplied this URL to a relatively long review of recent SATA and SCSI > controllers - it's worth a read. And surprisingly, the SATA controllers do > pretty well against the SCSI controllers. > > http://www.tweakers.net/reviews/557/1 thanks for this very good link! > > hjm > > On Wednesday 22 June 2005 3:00 pm, Ming Zhang wrote: > > will this 24 port card itself will be a bottleneck? > > > > ming > > > [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:45 ` Harry Mangalam 2005-06-22 23:05 ` Ming Zhang @ 2005-06-23 0:25 ` bdameron 1 sibling, 0 replies; 34+ messages in thread From: bdameron @ 2005-06-23 0:25 UTC (permalink / raw) To: linux-raid I read this article some time ago. Note that they were testing the older Areca card with it's first firmware version. The new PCIe card has twice the CPU power and has worked out their initial bugs. The performance is enourmously improved. However the price is quit a lot higher than most. Brad Dameron SeaTab Software www.seatab.com Quoting Harry Mangalam <hjm@tacgi.com>: > Also a good point - I was just discussing this in another thread. > Areca makes a number of SATA RAID controllers - some are SCSI-SATA, some are > > PCI-X-SATA, some are PCIe-SATA. > > Areca supplied this URL to a relatively long review of recent SATA and SCSI > controllers - it's worth a read. And surprisingly, the SATA controllers do > pretty well against the SCSI controllers. > > http://www.tweakers.net/reviews/557/1 > > hjm > > On Wednesday 22 June 2005 3:00 pm, Ming Zhang wrote: > > will this 24 port card itself will be a bottleneck? > > > > ming > > > > -- > Cheers, Harry > Harry J Mangalam - 949 856 2847 (vox; email for fax) - hjm@tacgi.com > <<plain text preferred>> > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 22:00 ` Ming Zhang 2005-06-22 22:11 ` John Madden 2005-06-22 22:45 ` Harry Mangalam @ 2005-06-23 0:14 ` bdameron 2005-06-23 0:49 ` Ming Zhang 2 siblings, 1 reply; 34+ messages in thread From: bdameron @ 2005-06-23 0:14 UTC (permalink / raw) To: linux-raid Quoting Ming Zhang <mingz@ele.uri.edu>: > will this 24 port card itself will be a bottleneck? > > ming Since the card is PCI-X the only bottleneck on it might be the Processor since it is shared with all 24 ports. But I do not know for sure without testing it. I personally am going to stick with the new 16 port version. Which is a PCI- Express card and has twice the CPU power. Since there are so many spindles it should be pretty darn fast. And remember that even tho the drives are 150MBps they realistically only do about 25-30MBps. Brad Dameron SeaTab Software www.seatab.com ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-23 0:14 ` bdameron @ 2005-06-23 0:49 ` Ming Zhang 2005-06-23 3:05 ` Guy 0 siblings, 1 reply; 34+ messages in thread From: Ming Zhang @ 2005-06-23 0:49 UTC (permalink / raw) To: bdameron; +Cc: linux-raid [-- Attachment #1: Type: text/plain, Size: 1147 bytes --] On Wed, 2005-06-22 at 17:14 -0700, bdameron@pivotlink.com wrote: > Quoting Ming Zhang <mingz@ele.uri.edu>: > > > will this 24 port card itself will be a bottleneck? > > > > ming > > Since the card is PCI-X the only bottleneck on it might be the Processor since > it is shared with all 24 ports. But I do not know for sure without testing it. > I personally am going to stick with the new 16 port version. Which is a PCI- > Express card and has twice the CPU power. Since there are so many spindles it > should be pretty darn fast. And remember that even tho the drives are 150MBps > they realistically only do about 25-30MBps. the problem here is taht each HD can stably deliver 25-30MBps while the PCI-x will not arrive that high if have 16 or 24 ports. i do not have a chance to try out though. those bus at most arrive 70-80% the claimed peak # :P > > Brad Dameron > SeaTab Software > www.seatab.com > - > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* RE: More tales of horror from the linux (HW) raid crypt 2005-06-23 0:49 ` Ming Zhang @ 2005-06-23 3:05 ` Guy 2005-06-23 12:31 ` Ming Zhang 0 siblings, 1 reply; 34+ messages in thread From: Guy @ 2005-06-23 3:05 UTC (permalink / raw) To: mingz, bdameron; +Cc: linux-raid > -----Original Message----- > From: linux-raid-owner@vger.kernel.org [mailto:linux-raid- > owner@vger.kernel.org] On Behalf Of Ming Zhang > Sent: Wednesday, June 22, 2005 8:49 PM > To: bdameron@pivotlink.com > Cc: linux-raid@vger.kernel.org > Subject: Re: More tales of horror from the linux (HW) raid crypt > > On Wed, 2005-06-22 at 17:14 -0700, bdameron@pivotlink.com wrote: > > Quoting Ming Zhang <mingz@ele.uri.edu>: > > > > > will this 24 port card itself will be a bottleneck? > > > > > > ming > > > > Since the card is PCI-X the only bottleneck on it might be the Processor > since > > it is shared with all 24 ports. But I do not know for sure without > testing it. > > I personally am going to stick with the new 16 port version. Which is a > PCI- > > Express card and has twice the CPU power. Since there are so many > spindles it > > should be pretty darn fast. And remember that even tho the drives are > 150MBps > > they realistically only do about 25-30MBps. > > the problem here is taht each HD can stably deliver 25-30MBps while the > PCI-x will not arrive that high if have 16 or 24 ports. i do not have a > chance to try out though. those bus at most arrive 70-80% the claimed > peak # :P Maybe my math is wrong... But 24 disks at 30 MB/s is 720 MB/s, that is about 68.2% of the PCI-X bandwidth of 1056 MB/s. Also, 30 MB/s assumes sequential disk access. That does not occur in the real world. Only during testing. IMO Guy > > > > > > > Brad Dameron > > SeaTab Software > > www.seatab.com > > - > > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > > the body of a message to majordomo@vger.kernel.org > > More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 34+ messages in thread
* RE: More tales of horror from the linux (HW) raid crypt 2005-06-23 3:05 ` Guy @ 2005-06-23 12:31 ` Ming Zhang 2005-06-23 13:03 ` Guy 0 siblings, 1 reply; 34+ messages in thread From: Ming Zhang @ 2005-06-23 12:31 UTC (permalink / raw) To: Guy; +Cc: bdameron, linux-raid [-- Attachment #1: Type: text/plain, Size: 1457 bytes --] On Wed, 2005-06-22 at 23:05 -0400, Guy wrote: > > -----Original Message----- > > > > > > > will this 24 port card itself will be a bottleneck? > > > > > > > > ming > > > > > > Since the card is PCI-X the only bottleneck on it might be the Processor > > since > > > it is shared with all 24 ports. But I do not know for sure without > > testing it. > > > I personally am going to stick with the new 16 port version. Which is a > > PCI- > > > Express card and has twice the CPU power. Since there are so many > > spindles it > > > should be pretty darn fast. And remember that even tho the drives are > > 150MBps > > > they realistically only do about 25-30MBps. > > > > the problem here is taht each HD can stably deliver 25-30MBps while the > > PCI-x will not arrive that high if have 16 or 24 ports. i do not have a > > chance to try out though. those bus at most arrive 70-80% the claimed > > peak # :P > > Maybe my math is wrong... > But 24 disks at 30 MB/s is 720 MB/s, that is about 68.2% of the PCI-X > bandwidth of 1056 MB/s. yes, u math is better. > > Also, 30 MB/s assumes sequential disk access. That does not occur in the > real world. Only during testing. IMO yes, only during test. but what if people build raid5 base on it, this is probably what people do. and then a disk fail? then a full disk sequential access becomes normal. and disk fail in 24 disk is not so uncommon. > > Guy > > > [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* RE: More tales of horror from the linux (HW) raid crypt 2005-06-23 12:31 ` Ming Zhang @ 2005-06-23 13:03 ` Guy 2005-06-23 13:17 ` Andy Smith 2005-06-23 13:19 ` Ming Zhang 0 siblings, 2 replies; 34+ messages in thread From: Guy @ 2005-06-23 13:03 UTC (permalink / raw) To: mingz; +Cc: bdameron, linux-raid > -----Original Message----- > From: Ming Zhang [mailto:mingz@ele.uri.edu] > Sent: Thursday, June 23, 2005 8:32 AM > To: Guy > Cc: bdameron@pivotlink.com; linux-raid@vger.kernel.org > Subject: RE: More tales of horror from the linux (HW) raid crypt > > On Wed, 2005-06-22 at 23:05 -0400, Guy wrote: > > > -----Original Message----- > > > > > > > > > will this 24 port card itself will be a bottleneck? > > > > > > > > > > ming > > > > > > > > Since the card is PCI-X the only bottleneck on it might be the > Processor > > > since > > > > it is shared with all 24 ports. But I do not know for sure without > > > testing it. > > > > I personally am going to stick with the new 16 port version. Which > is a > > > PCI- > > > > Express card and has twice the CPU power. Since there are so many > > > spindles it > > > > should be pretty darn fast. And remember that even tho the drives > are > > > 150MBps > > > > they realistically only do about 25-30MBps. > > > > > > the problem here is taht each HD can stably deliver 25-30MBps while > the > > > PCI-x will not arrive that high if have 16 or 24 ports. i do not have > a > > > chance to try out though. those bus at most arrive 70-80% the claimed > > > peak # :P > > > > Maybe my math is wrong... > > But 24 disks at 30 MB/s is 720 MB/s, that is about 68.2% of the PCI-X > > bandwidth of 1056 MB/s. > yes, u math is better. > > > > > Also, 30 MB/s assumes sequential disk access. That does not occur in > the > > real world. Only during testing. IMO > yes, only during test. but what if people build raid5 base on it, this > is probably what people do. and then a disk fail? then a full disk > sequential access becomes normal. and disk fail in 24 disk is not so > uncommon. But, this is hardware RAID. A re-sync would not affect the PCI bus. All disk i/o related to re-building the array would be internal to the card. However, even if it were a software RAID card, the PCI-X would be at 68.2% load, so it should not be a problem. If my math is correct! :) Also, a single RAID5 array on 24 disks would be high risk of a double failure. I think I would build 2 RAID5 arrays of 12 disks each. Or 2 RAID5 arrays of 11 disks each, with 2 spares. > > > > > Guy > > > > > ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-23 13:03 ` Guy @ 2005-06-23 13:17 ` Andy Smith 2005-06-23 13:19 ` Ming Zhang 1 sibling, 0 replies; 34+ messages in thread From: Andy Smith @ 2005-06-23 13:17 UTC (permalink / raw) To: linux-raid [-- Attachment #1: Type: text/plain, Size: 352 bytes --] On Thu, Jun 23, 2005 at 09:03:35AM -0400, Guy wrote: > Also, a single RAID5 array on 24 disks would be high risk of a double > failure. I think I would build 2 RAID5 arrays of 12 disks each. Or 2 RAID5 > arrays of 11 disks each, with 2 spares. Just out of interest, why wouldn't you do RAID 10 of 12 pairs? Because of getting only 50% capacity, or? [-- Attachment #2: Digital signature --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* RE: More tales of horror from the linux (HW) raid crypt 2005-06-23 13:03 ` Guy 2005-06-23 13:17 ` Andy Smith @ 2005-06-23 13:19 ` Ming Zhang 1 sibling, 0 replies; 34+ messages in thread From: Ming Zhang @ 2005-06-23 13:19 UTC (permalink / raw) To: Guy; +Cc: bdameron, linux-raid [-- Attachment #1: Type: text/plain, Size: 2683 bytes --] On Thu, 2005-06-23 at 09:03 -0400, Guy wrote: > > > -----Original Message----- > > From: Ming Zhang [mailto:mingz@ele.uri.edu] > > Sent: Thursday, June 23, 2005 8:32 AM > > To: Guy > > Cc: bdameron@pivotlink.com; linux-raid@vger.kernel.org > > Subject: RE: More tales of horror from the linux (HW) raid crypt > > > > On Wed, 2005-06-22 at 23:05 -0400, Guy wrote: > > > > -----Original Message----- > > > > > > > > > > > will this 24 port card itself will be a bottleneck? > > > > > > > > > > > > ming > > > > > > > > > > Since the card is PCI-X the only bottleneck on it might be the > > Processor > > > > since > > > > > it is shared with all 24 ports. But I do not know for sure without > > > > testing it. > > > > > I personally am going to stick with the new 16 port version. Which > > is a > > > > PCI- > > > > > Express card and has twice the CPU power. Since there are so many > > > > spindles it > > > > > should be pretty darn fast. And remember that even tho the drives > > are > > > > 150MBps > > > > > they realistically only do about 25-30MBps. > > > > > > > > the problem here is taht each HD can stably deliver 25-30MBps while > > the > > > > PCI-x will not arrive that high if have 16 or 24 ports. i do not have > > a > > > > chance to try out though. those bus at most arrive 70-80% the claimed > > > > peak # :P > > > > > > Maybe my math is wrong... > > > But 24 disks at 30 MB/s is 720 MB/s, that is about 68.2% of the PCI-X > > > bandwidth of 1056 MB/s. > > yes, u math is better. > > > > > > > > Also, 30 MB/s assumes sequential disk access. That does not occur in > > the > > > real world. Only during testing. IMO > > yes, only during test. but what if people build raid5 base on it, this > > is probably what people do. and then a disk fail? then a full disk > > sequential access becomes normal. and disk fail in 24 disk is not so > > uncommon. > But, this is hardware RAID. A re-sync would not affect the PCI bus. All > disk i/o related to re-building the array would be internal to the card. > However, even if it were a software RAID card, the PCI-X would be at 68.2% > load, so it should not be a problem. If my math is correct! :) > > Also, a single RAID5 array on 24 disks would be high risk of a double > failure. I think I would build 2 RAID5 arrays of 12 disks each. Or 2 RAID5 > arrays of 11 disks each, with 2 spares. > or build with raid6. but anyway, once u a disk, u will need to have a rsync. with MD, u might try the fr5. no money, otherwise buy one and try. :P i think i need a 4U unit, dual amd64... ming > > > > > > > > Guy > > > > > > > > > [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 20:16 ` Harry Mangalam 2005-06-22 20:38 ` Jure Pecar 2005-06-22 21:09 ` Brad Dameron @ 2005-06-22 23:54 ` Jon Lewis 2 siblings, 0 replies; 34+ messages in thread From: Jon Lewis @ 2005-06-22 23:54 UTC (permalink / raw) To: Harry Mangalam; +Cc: linux-raid On Wed, 22 Jun 2005, Harry Mangalam wrote: > I personally buy IBM (now Hitachi I think) and Seagate and have always avoided > WD as being cheap and cheesy. 5 years ago when I was a sysadmin, I was > replacing WD at probably 5 times the rate of anything else. It was probably Everyone has their own personal HD brand horror stories. I've seen multiple IBM DeathStar drives suddenly self destruct. I've had multiple defective brand new Maxtor SATA drives and Maxtor send an RMA replacement (for a PATA) drive that after being stored for a few months turned out to be dead when I tried to put it in service. I try to base my drive purchases on warranty length now, on the assumption that if they don't want to stand behind their product (<3 year warranty), I don't want to rely on it either. ---------------------------------------------------------------------- Jon Lewis | I route Senior Network Engineer | therefore you are Atlantic Net | _________ http://www.lewis.org/~jlewis/pgp for PGP public key_________ ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 19:33 ` Mike Hardy 2005-06-22 20:16 ` Harry Mangalam @ 2005-06-22 20:54 ` Dan Stromberg 2005-06-22 21:15 ` Brad Dameron 1 sibling, 1 reply; 34+ messages in thread From: Dan Stromberg @ 2005-06-22 20:54 UTC (permalink / raw) To: Mike Hardy; +Cc: Harry Mangalam, linux-raid, strombrg [-- Attachment #1: Type: text/plain, Size: 1146 bytes --] On Wed, 2005-06-22 at 12:33 -0700, Mike Hardy wrote: > I'm very interested in the relative SW raid / HW raid performance. I > have both in service (two raid 5 sets are actually the same size with > the same number of components) and see roughly the same as you mention. > One difference that I see is that HW raid should generate fewer > interrupts and lower bus traffic. In the early days of RAID, people always used to say that for speed, you had to get a hardware RAID controller instead of doing software RAID. However, I saw an interesting comment on a local linux user group mailing list recently. That comment was to the effect that hardware RAID controllers tend to have no where near the horsepower of a modern desktop CPU - so the claim was (I've not verified this!) that using software RAID would be much faster. I'm thinking that if you can do software RAID with a dedicated box that's doing nothing but RAID and a little network, then yes, that likely would be faster than some scaled-down hardware RAID controller - but it would also likely cost more, be less likely to have good hot swap, and so on. [-- Attachment #2: This is a digitally signed message part --] [-- Type: application/pgp-signature, Size: 189 bytes --] ^ permalink raw reply [flat|nested] 34+ messages in thread
* Re: More tales of horror from the linux (HW) raid crypt 2005-06-22 20:54 ` Dan Stromberg @ 2005-06-22 21:15 ` Brad Dameron 0 siblings, 0 replies; 34+ messages in thread From: Brad Dameron @ 2005-06-22 21:15 UTC (permalink / raw) To: linux-raid On Wed, 2005-06-22 at 13:54 -0700, Dan Stromberg wrote: > On Wed, 2005-06-22 at 12:33 -0700, Mike Hardy wrote: > > > I'm very interested in the relative SW raid / HW raid performance. I > > have both in service (two raid 5 sets are actually the same size with > > the same number of components) and see roughly the same as you mention. > > One difference that I see is that HW raid should generate fewer > > interrupts and lower bus traffic. > > In the early days of RAID, people always used to say that for speed, you > had to get a hardware RAID controller instead of doing software RAID. > > However, I saw an interesting comment on a local linux user group > mailing list recently. That comment was to the effect that hardware > RAID controllers tend to have no where near the horsepower of a modern > desktop CPU - so the claim was (I've not verified this!) that using > software RAID would be much faster. > > I'm thinking that if you can do software RAID with a dedicated box > that's doing nothing but RAID and a little network, then yes, that > likely would be faster than some scaled-down hardware RAID controller - > but it would also likely cost more, be less likely to have good hot > swap, and so on. This is very true. Especially with 64bit machines. The software RAID is much faster and efficient. However with servers that are using a lot of CPU resources it can become a bottleneck. Our inhouse application is proven faster with hardware RAID than with software RAID. Mainly due to the 100% CPU utilization. It being able to offload the IO to the MegaRAID card helps. I can't wait to test it with the new 320-2E. The CPU is twice as fast as the 320-2X we use now. Brad Dameron SeaTab Software www.seatab.com ^ permalink raw reply [flat|nested] 34+ messages in thread
end of thread, other threads:[~2005-06-23 13:19 UTC | newest] Thread overview: 34+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2005-06-18 11:47 when does it become faulty disk Raz Ben-Jehuda(caro) 2005-06-19 19:10 ` Molle Bestefich 2005-06-20 6:43 ` raz ben jehuda 2005-06-20 7:55 ` Molle Bestefich 2005-06-20 10:09 ` raz ben jehuda 2005-06-20 13:45 ` Michael Tokarev 2005-06-20 15:35 ` raz ben jehuda 2005-06-21 1:53 ` More tales of horror from the linux (HW) raid crypt Harry Mangalam 2005-06-22 19:33 ` Mike Hardy 2005-06-22 20:16 ` Harry Mangalam 2005-06-22 20:38 ` Jure Pecar 2005-06-22 21:33 ` Harry Mangalam 2005-06-22 23:15 ` SMART, was " Konstantin Olchanski 2005-06-22 23:32 ` Harry Mangalam 2005-06-22 23:35 ` Mike Hardy 2005-06-22 21:09 ` Brad Dameron 2005-06-22 21:43 ` Harry Mangalam 2005-06-22 22:00 ` Ming Zhang 2005-06-22 22:11 ` John Madden 2005-06-22 22:26 ` Ming Zhang 2005-06-23 0:20 ` bdameron 2005-06-22 22:45 ` Harry Mangalam 2005-06-22 23:05 ` Ming Zhang 2005-06-23 0:25 ` bdameron 2005-06-23 0:14 ` bdameron 2005-06-23 0:49 ` Ming Zhang 2005-06-23 3:05 ` Guy 2005-06-23 12:31 ` Ming Zhang 2005-06-23 13:03 ` Guy 2005-06-23 13:17 ` Andy Smith 2005-06-23 13:19 ` Ming Zhang 2005-06-22 23:54 ` Jon Lewis 2005-06-22 20:54 ` Dan Stromberg 2005-06-22 21:15 ` Brad Dameron
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).