* raid5 code ok with 2TB + ? @ 2004-11-28 10:48 Stephan van Hienen 2004-11-28 10:57 ` Brad Campbell 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-28 10:48 UTC (permalink / raw) To: linux-raid in the past the raid5 code was 32bits internal and the limit was 2TB is this fixed in the 2.6.x ? or is the 2TB limit still there ? ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? 2004-11-28 10:48 raid5 code ok with 2TB + ? Stephan van Hienen @ 2004-11-28 10:57 ` Brad Campbell 2004-11-28 11:19 ` Stephan van Hienen 2004-11-28 18:19 ` Guy 0 siblings, 2 replies; 25+ messages in thread From: Brad Campbell @ 2004-11-28 10:57 UTC (permalink / raw) To: Stephan van Hienen; +Cc: linux-raid Stephan van Hienen wrote: > in the past the raid5 code was 32bits internal > and the limit was 2TB > > is this fixed in the 2.6.x ? I hope so, I have been using it on a 2.2TB raid-5 for months now. (I have completely filled the partition and e2fsck it monthly so I would hope any glitches should have surfaced by now) brad@srv:~$ cat /proc/mdstat Personalities : [raid0] [raid5] [raid6] md2 : active raid5 sdl[0] sdm[2] sdk[1] 488396800 blocks level 5, 128k chunk, algorithm 2 [3/3] [UUU] md0 : active raid5 sda1[0] sdj1[9] sdi1[8] sdh1[7] sdg1[6] sdf1[5] sde1[4] sdd1[3] sdc1[2] sdb1[1] 2206003968 blocks level 5, 128k chunk, algorithm 0 [10/10] [UUUUUUUUUU] -- Brad /"\ Save the Forests \ / ASCII RIBBON CAMPAIGN Burn a Greenie. X AGAINST HTML MAIL / \ ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? 2004-11-28 10:57 ` Brad Campbell @ 2004-11-28 11:19 ` Stephan van Hienen 2004-11-28 11:28 ` Brad Campbell 2004-11-28 18:19 ` Guy 1 sibling, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-28 11:19 UTC (permalink / raw) To: Brad Campbell; +Cc: linux-raid On Sun, 28 Nov 2004, Brad Campbell wrote: > I hope so, I have been using it on a 2.2TB raid-5 for months now. (I have > completely filled the partition and e2fsck it monthly so I would hope any > glitches should have surfaced by now) ok looks ok when i tried 2.3TB last year it failed within a few days (data written to the 2TB+ was written to the first part (0-0.3TB) btw which filesystem are you using ? (and which stripesize?) ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? 2004-11-28 11:19 ` Stephan van Hienen @ 2004-11-28 11:28 ` Brad Campbell 0 siblings, 0 replies; 25+ messages in thread From: Brad Campbell @ 2004-11-28 11:28 UTC (permalink / raw) To: Stephan van Hienen; +Cc: linux-raid Stephan van Hienen wrote: > On Sun, 28 Nov 2004, Brad Campbell wrote: > >> I hope so, I have been using it on a 2.2TB raid-5 for months now. (I >> have completely filled the partition and e2fsck it monthly so I would >> hope any glitches should have surfaced by now) > > > ok looks ok > when i tried 2.3TB last year it failed within a few days (data written > to the 2TB+ was written to the first part (0-0.3TB) > > btw which filesystem are you using ? > (and which stripesize?) Ext3 and ... brad@srv:~$ sudo mdadm --misc --detail /dev/md0 /dev/md0: Version : 00.90.01 Creation Time : Sun May 2 18:02:14 2004 Raid Level : raid5 Array Size : 2206003968 (2103.81 GiB 2258.95 GB) Device Size : 245111552 (233.76 GiB 250.99 GB) Raid Devices : 10 Total Devices : 10 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Sun Nov 28 06:44:04 2004 State : clean Active Devices : 10 Working Devices : 10 Failed Devices : 0 Spare Devices : 0 Layout : left-asymmetric Chunk Size : 128K UUID : 05cc3f43:de1ecfa4:83a51293:78015f1e Events : 0.1031147 Number Major Minor RaidDevice State 0 8 1 0 active sync /dev/devfs/scsi/host0/bus0/target0/lun0/part1 1 8 17 1 active sync /dev/devfs/scsi/host1/bus0/target0/lun0/part1 2 8 33 2 active sync /dev/devfs/scsi/host2/bus0/target0/lun0/part1 3 8 49 3 active sync /dev/devfs/scsi/host3/bus0/target0/lun0/part1 4 8 65 4 active sync /dev/devfs/scsi/host4/bus0/target0/lun0/part1 5 8 81 5 active sync /dev/devfs/scsi/host5/bus0/target0/lun0/part1 6 8 97 6 active sync /dev/devfs/scsi/host6/bus0/target0/lun0/part1 7 8 113 7 active sync /dev/devfs/scsi/host7/bus0/target0/lun0/part1 8 8 129 8 active sync /dev/sdi1 9 8 145 9 active sync /dev/sdj1 -- Brad /"\ Save the Forests \ / ASCII RIBBON CAMPAIGN Burn a Greenie. X AGAINST HTML MAIL / \ ^ permalink raw reply [flat|nested] 25+ messages in thread
* RE: raid5 code ok with 2TB + ? 2004-11-28 10:57 ` Brad Campbell 2004-11-28 11:19 ` Stephan van Hienen @ 2004-11-28 18:19 ` Guy 2004-11-28 19:20 ` Brad Campbell 1 sibling, 1 reply; 25+ messages in thread From: Guy @ 2004-11-28 18:19 UTC (permalink / raw) To: 'Brad Campbell', 'Stephan van Hienen'; +Cc: linux-raid Your 2.2T array is not as big as you think! 1TB = 2^40 not 1*10^12 Maybe depending on if you are buying disk drives, or selling them! :) But when related to the 2TB limit it is 2^40. 2206003968 blocks Divide by 1024 gives you 2154300.75 meg Divide by 1024 gives you 2103.8093 Gig Divide by 1024 gives you 2.0545 TB So you are just over 2TB. by 58520320 blocks or 55.8 Gig. The only reason I am being exact is that you have not tested disk I/O beyond 2TB as much as you think. Once, someone else made a similar 2T claim, after the math he was really below 2T. Guy -----Original Message----- From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Brad Campbell Sent: Sunday, November 28, 2004 5:58 AM To: Stephan van Hienen Cc: linux-raid@vger.kernel.org Subject: Re: raid5 code ok with 2TB + ? Stephan van Hienen wrote: > in the past the raid5 code was 32bits internal > and the limit was 2TB > > is this fixed in the 2.6.x ? I hope so, I have been using it on a 2.2TB raid-5 for months now. (I have completely filled the partition and e2fsck it monthly so I would hope any glitches should have surfaced by now) brad@srv:~$ cat /proc/mdstat Personalities : [raid0] [raid5] [raid6] md2 : active raid5 sdl[0] sdm[2] sdk[1] 488396800 blocks level 5, 128k chunk, algorithm 2 [3/3] [UUU] md0 : active raid5 sda1[0] sdj1[9] sdi1[8] sdh1[7] sdg1[6] sdf1[5] sde1[4] sdd1[3] sdc1[2] sdb1[1] 2206003968 blocks level 5, 128k chunk, algorithm 0 [10/10] [UUUUUUUUUU] -- Brad /"\ Save the Forests \ / ASCII RIBBON CAMPAIGN Burn a Greenie. X AGAINST HTML MAIL / \ - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? 2004-11-28 18:19 ` Guy @ 2004-11-28 19:20 ` Brad Campbell 2004-11-28 20:17 ` raid5 code ok with 2TB + ? NEGATIVE :( Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Brad Campbell @ 2004-11-28 19:20 UTC (permalink / raw) To: Guy; +Cc: 'Stephan van Hienen', linux-raid Guy wrote: > Your 2.2T array is not as big as you think! > > 1TB = 2^40 not 1*10^12 > Maybe depending on if you are buying disk drives, or selling them! :) > But when related to the 2TB limit it is 2^40. > > 2206003968 blocks > Divide by 1024 gives you 2154300.75 meg > Divide by 1024 gives you 2103.8093 Gig > Divide by 1024 gives you 2.0545 TB > So you are just over 2TB. by 58520320 blocks or 55.8 Gig. > > The only reason I am being exact is that you have not tested disk I/O beyond > 2TB as much as you think. Once, someone else made a similar 2T claim, after > the math he was really below 2T. Fair call. Having said that, if the code wrapped at 2TB then I would have blown away the 1st 55.8 Gig of my partition, which would be enough to prove the code faulty :p) -- Brad /"\ Save the Forests \ / ASCII RIBBON CAMPAIGN Burn a Greenie. X AGAINST HTML MAIL / \ ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? NEGATIVE :( 2004-11-28 19:20 ` Brad Campbell @ 2004-11-28 20:17 ` Stephan van Hienen 2004-11-28 20:27 ` Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-28 20:17 UTC (permalink / raw) To: Brad Campbell; +Cc: Guy, linux-raid just tried creating a raid5 it was supposed to be 2180 GB but mdadm -D /dev/md0 reports 131.81 GiB looks like it's still not supported ? (kernel 2.6.9-ac11 and mdadm 1.8.0) ]# mdadm --create /dev/md0 -c 256 -l 5 -n 14 /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1 /dev/sdg1 /dev/sdh1 /dev/sdi1 /dev/sdj1 /dev/sdk1 /dev/sdl1 /dev/sdm1 /dev/sdn1 /dev/sdo1 <..> Continue creating array? y mdadm: array /dev/md0 started. [root@storage etc]# cat /proc/mdstat Personalities : [raid0] [raid5] md0 : active raid5 sdo1[14] sdn1[12] sdm1[11] sdl1[10] sdk1[9] sdj1[8] sdi1[7] sdh1[6] sdg1[5] sdf1[4] sde1[3] sdd1[2] sdc1[1] sdb1[0] 2285700352 blocks level 5, 256k chunk, algorithm 2 [14/13] [UUUUUUUUUUUUU_] [>....................] recovery = 0.0% (56064/175823104) finish=156.6min speed=18688K/sec unused devices: <none> [root@storage etc]# mdadm -D /dev/md0 /dev/md0: Version : 00.90.01 Creation Time : Sun Nov 28 21:15:32 2004 Raid Level : raid5 Array Size : 138216704 (131.81 GiB 141.53 GB) Device Size : 175823104 (167.68 GiB 180.04 GB) Raid Devices : 14 Total Devices : 14 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Sun Nov 28 21:15:32 2004 State : clean, degraded, recovering Active Devices : 13 Working Devices : 14 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 256K Rebuild Status : 0% complete UUID : 70c4df2b:2dc20e84:e555fa94:a57be873 Events : 0.4 Number Major Minor RaidDevice State 0 8 17 0 active sync /dev/sdb1 1 8 33 1 active sync /dev/sdc1 2 8 49 2 active sync /dev/sdd1 3 8 65 3 active sync /dev/sde1 4 8 81 4 active sync /dev/sdf1 5 8 97 5 active sync /dev/sdg1 6 8 113 6 active sync /dev/sdh1 7 8 129 7 active sync /dev/sdi1 8 8 145 8 active sync /dev/sdj1 9 8 161 9 active sync /dev/sdk1 10 8 177 10 active sync /dev/sdl1 11 8 193 11 active sync /dev/sdm1 12 8 209 12 active sync /dev/sdn1 13 0 0 - removed 14 8 225 13 spare rebuilding /dev/sdo1 ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? NEGATIVE :( 2004-11-28 20:17 ` raid5 code ok with 2TB + ? NEGATIVE :( Stephan van Hienen @ 2004-11-28 20:27 ` Stephan van Hienen 2004-11-28 20:45 ` Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-28 20:27 UTC (permalink / raw) To: Brad Campbell; +Cc: Guy, linux-raid btw 'Large block devices' is enabled in kernel config On Sun, 28 Nov 2004, Stephan van Hienen wrote: > just tried creating a raid5 > it was supposed to be 2180 GB > but mdadm -D /dev/md0 reports 131.81 GiB > looks like it's still not supported ? > (kernel 2.6.9-ac11 and mdadm 1.8.0) > kk ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? NEGATIVE :( 2004-11-28 20:27 ` Stephan van Hienen @ 2004-11-28 20:45 ` Stephan van Hienen 2004-11-28 23:34 ` raid5 code ok with 2TB + ? Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-28 20:45 UTC (permalink / raw) To: Brad Campbell; +Cc: Guy, linux-raid On Sun, 28 Nov 2004, Stephan van Hienen wrote: > btw 'Large block devices' is enabled in kernel config argh and i copyed it to the wrong place so i was still running a non LBD kernel ]# mdadm -D /dev/md0 /dev/md0: Version : 00.90.01 Creation Time : Sun Nov 28 21:44:33 2004 Raid Level : raid5 Array Size : 2285700352 (2179.81 GiB 2340.56 GB) Device Size : 175823104 (167.68 GiB 180.04 GB) now running a resync after that i'll run badblocks on it to test if i can read/write to the complete array ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 code ok with 2TB + ? 2004-11-28 20:45 ` Stephan van Hienen @ 2004-11-28 23:34 ` Stephan van Hienen 2004-11-29 4:49 ` Guy 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-28 23:34 UTC (permalink / raw) To: linux-raid On Sun, 28 Nov 2004, Stephan van Hienen wrote: > now running a resync > after that i'll run badblocks on it to test if i can read/write to the > complete array resync finished without any problems i wanted to run badblocks on the /dev/md0 to check but it looks like badblocks doesn't support 2TB + : ]# badblocks -sw -p99 -c8192 /dev/md0 Writing pattern 0xaaaaaaaa: 126464/-2009266944 any other hints for testing the raid ? ^ permalink raw reply [flat|nested] 25+ messages in thread
* RE: raid5 code ok with 2TB + ? 2004-11-28 23:34 ` raid5 code ok with 2TB + ? Stephan van Hienen @ 2004-11-29 4:49 ` Guy 2004-11-29 8:33 ` Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Guy @ 2004-11-29 4:49 UTC (permalink / raw) To: 'Stephan van Hienen', linux-raid I use dd to test disks, and arrays. Give this a test. It may stop at the 2TB limit, not sure. My monster 14 disk array is only about 240G. time dd if=/dev/md0 of=/dev/null bs=1024k I don't think badblocks would detect bad block on an array. If one did occur, the drive would be failed and the array would continue. If your re-sync finishes, that really does test your disks since every block used by md is read or written. Guy -----Original Message----- From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Stephan van Hienen Sent: Sunday, November 28, 2004 6:35 PM To: linux-raid@vger.kernel.org Subject: Re: raid5 code ok with 2TB + ? On Sun, 28 Nov 2004, Stephan van Hienen wrote: > now running a resync > after that i'll run badblocks on it to test if i can read/write to the > complete array resync finished without any problems i wanted to run badblocks on the /dev/md0 to check but it looks like badblocks doesn't support 2TB + : ]# badblocks -sw -p99 -c8192 /dev/md0 Writing pattern 0xaaaaaaaa: 126464/-2009266944 any other hints for testing the raid ? - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 25+ messages in thread
* RE: raid5 code ok with 2TB + ? 2004-11-29 4:49 ` Guy @ 2004-11-29 8:33 ` Stephan van Hienen 2004-11-29 8:51 ` raid5 slow Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-29 8:33 UTC (permalink / raw) To: Guy; +Cc: linux-raid On Sun, 28 Nov 2004, Guy wrote: > I use dd to test disks, and arrays. Give this a test. It may stop at the > 2TB limit, not sure. My monster 14 disk array is only about 240G. > > time dd if=/dev/md0 of=/dev/null bs=1024k > ]# dd if=/dev/zero of=testfile bs=8192 dd: writing `testfile': No space left on device 285696059+0 records in 285696058+0 records out du doesn't undestand the size : ]# du -h 132G . ls understands a bit (total is incorrect, but filesize is ok) : ]# ls -la total 138084832 drwxr-xr-x 21 root root 4096 Nov 29 01:40 .. drwxr-xr-x 2 root root 21 Nov 29 02:07 . -rw-r--r-- 1 root root 2340422111232 Nov 29 07:39 testfile ]# ls -al -h total 132G drwxr-xr-x 2 root root 21 Nov 29 02:07 . drwxr-xr-x 21 root root 4.0K Nov 29 01:40 .. -rw-r--r-- 1 root root 2.1T Nov 29 07:39 testfile no errors in dmesg so looks like it's ok it's only a bit slow : Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 26828 99 123983 73 58525 57 27336 99 171371 96 583.9 4 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1023 12 +++++ +++ 920 11 1004 14 +++++ +++ 722 9 storage.a2000.nu,2G,26828,99,123983,73,58525,57,27336,99,171371,96,583.9,4,16,1023,12,+++++,+++,920,11,1004,14,+++++,+++,722,9 compared to the old raid5 with 13 disk (with redhat 8 and kernel 2.4) : met 13 disken : (rh 8 ext3) Version 1.02c ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 1G 17775 100 100912 99 70290 66 23648 99 257029 81 624.1 3 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1944 99 +++++ +++ +++++ +++ 2042 98 +++++ +++ 5689 97 storage.a2000.nu,1G,17775,100,100912,99,70290,66,23648,99,257029,81, 624.1,3,16,1944,99,+++++,+++,+++++,+++,2042,98,+++++,+++,5689,97 ^ permalink raw reply [flat|nested] 25+ messages in thread
* raid5 slow 2004-11-29 8:33 ` Stephan van Hienen @ 2004-11-29 8:51 ` Stephan van Hienen 2004-11-29 17:13 ` Guy 2004-11-30 0:30 ` raid5 slow Neil Brown 0 siblings, 2 replies; 25+ messages in thread From: Stephan van Hienen @ 2004-11-29 8:51 UTC (permalink / raw) To: Guy; +Cc: linux-raid > it's only a bit slow : > > Version 1.03 ------Sequential Output------ --Sequential Input- > --Random- > -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- > --Seeks-- > Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec > %CP > storage.a2000.nu 2G 26828 99 123983 73 58525 57 27336 99 171371 96 583.9 > 4 > ------Sequential Create------ --------Random > Create-------- > -Create-- --Read--- -Delete-- -Create-- --Read--- > -Delete-- > files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec > %CP > 16 1023 12 +++++ +++ 920 11 1004 14 +++++ +++ 722 9 > storage.a2000.nu,2G,26828,99,123983,73,58525,57,27336,99,171371,96,583.9,4,16,1023,12,+++++,+++,920,11,1004,14,+++++,+++,722,9 created a raid0 with the same disks : Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 28435 99 219329 73 102949 51 29704 99 275323 62 639.5 2 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 10108 84 +++++ +++ 8825 69 9877 88 +++++ +++ 7546 66 storage.a2000.nu,2G,28435,99,219329,73,102949,51,29704,99,275323,62,639.5,2,16,10108,84,+++++,+++,8825,69,9877,88,+++++,+++,7546,66 shouldn't the raid be about the same ? now i get 275MB/sec with raid0 (14 disks) 171MB/sec with raid5 (14 disks) while the old raid5 situation was 257MB/sec with raid5 (13 disks) ^ permalink raw reply [flat|nested] 25+ messages in thread
* RE: raid5 slow 2004-11-29 8:51 ` raid5 slow Stephan van Hienen @ 2004-11-29 17:13 ` Guy 2004-11-29 18:11 ` Stephan van Hienen 2004-11-30 0:30 ` raid5 slow Neil Brown 1 sibling, 1 reply; 25+ messages in thread From: Guy @ 2004-11-29 17:13 UTC (permalink / raw) To: 'Stephan van Hienen'; +Cc: linux-raid Very odd, RAID5 reading should be about the same as RAID0. And your old array confirms that. My guess is RAID5 needs some help. Neil? Any ideas? Is your stripe size the same on both? Did you allow your array to re-sync before the tests? Guy -----Original Message----- From: Stephan van Hienen [mailto:raid@a2000.nu] Sent: Monday, November 29, 2004 3:52 AM To: Guy Cc: linux-raid@vger.kernel.org Subject: raid5 slow > it's only a bit slow : > > Version 1.03 ------Sequential Output------ --Sequential Input- > --Random- > -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- > --Seeks-- > Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec > %CP > storage.a2000.nu 2G 26828 99 123983 73 58525 57 27336 99 171371 96 583.9 > 4 > ------Sequential Create------ --------Random > Create-------- > -Create-- --Read--- -Delete-- -Create-- --Read--- > -Delete-- > files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec > %CP > 16 1023 12 +++++ +++ 920 11 1004 14 +++++ +++ 722 9 > storage.a2000.nu,2G,26828,99,123983,73,58525,57,27336,99,171371,96,583.9,4,1 6,1023,12,+++++,+++,920,11,1004,14,+++++,+++,722,9 created a raid0 with the same disks : Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 28435 99 219329 73 102949 51 29704 99 275323 62 639.5 2 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 10108 84 +++++ +++ 8825 69 9877 88 +++++ +++ 7546 66 storage.a2000.nu,2G,28435,99,219329,73,102949,51,29704,99,275323,62,639.5,2, 16,10108,84,+++++,+++,8825,69,9877,88,+++++,+++,7546,66 shouldn't the raid be about the same ? now i get 275MB/sec with raid0 (14 disks) 171MB/sec with raid5 (14 disks) while the old raid5 situation was 257MB/sec with raid5 (13 disks) -- No virus found in this incoming message. Checked by AVG Anti-Virus. Version: 7.0.289 / Virus Database: 265.4.3 - Release Date: 11/26/2004 ^ permalink raw reply [flat|nested] 25+ messages in thread
* RE: raid5 slow 2004-11-29 17:13 ` Guy @ 2004-11-29 18:11 ` Stephan van Hienen 2004-11-30 0:06 ` raid5 slow (looks like 2.6 problem) Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-29 18:11 UTC (permalink / raw) To: Guy; +Cc: linux-raid On Mon, 29 Nov 2004, Guy wrote: > Very odd, RAID5 reading should be about the same as RAID0. > And your old array confirms that. > My guess is RAID5 needs some help. Neil? Any ideas? > > Is your stripe size the same on both? > Did you allow your array to re-sync before the tests? stripesize is the same (128k) and yes i did a resync ^ permalink raw reply [flat|nested] 25+ messages in thread
* RE: raid5 slow (looks like 2.6 problem) 2004-11-29 18:11 ` Stephan van Hienen @ 2004-11-30 0:06 ` Stephan van Hienen 0 siblings, 0 replies; 25+ messages in thread From: Stephan van Hienen @ 2004-11-30 0:06 UTC (permalink / raw) To: Guy; +Cc: linux-raid just rebooted my system with an old 2.4.20 kernel (got from backup) (2.4.20 13d r5 ext3) Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 21931 99 96429 99 79370 73 28970 99 280316 84 675.7 3 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1444 71 +++++ +++ +++++ +++ 2024 99 +++++ +++ 5563 98 storage.a2000.nu,2G,21931,99,96429,99,79370,73,28970,99,280316,84,675.7,3,16,1444,71,+++++,+++,+++++,+++,2024,99,+++++,+++,5563,98 again rebooted to 2.6.9-ac11 : (previous score was from xfs, so retested with ext3) (2.6.9-ac11 13d r5 ext3) Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 21711 99 99906 95 67559 79 24099 92 138763 72 606.4 3 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1470 79 +++++ +++ +++++ +++ 1447 78 +++++ +++ 4546 96 storage.a2000.nu,2G,21711,99,99906,95,67559,79,24099,92,138763,72,606.4,3,16,1470,79,+++++,+++,+++++,+++,1447,78,+++++,+++,4546,96 looks like something is wrong with the 2.6 code ? i'm now compiling normal 2.6.9 (and maybe 2.6.8 after this?) to check for results any ideas ? ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow 2004-11-29 8:51 ` raid5 slow Stephan van Hienen 2004-11-29 17:13 ` Guy @ 2004-11-30 0:30 ` Neil Brown 2004-11-30 0:50 ` Stephan van Hienen 1 sibling, 1 reply; 25+ messages in thread From: Neil Brown @ 2004-11-30 0:30 UTC (permalink / raw) To: Stephan van Hienen; +Cc: Guy, linux-raid On Monday November 29, raid@a2000.nu wrote: > > now i get 275MB/sec with raid0 (14 disks) > 171MB/sec with raid5 (14 disks) > > while the old raid5 situation was 257MB/sec with raid5 (13 disks) I presume you are only looking at 'read' speed here. There is a read optimisation that I haven't implemented in 2.6 yet that could account for about a 10% difference, but a 30% difference definitely surprises me. Normally raid5 will read into the stripe cache, and then copy data out of the stripe cache and into the request buffer for reads. The optimisation is to read directly into the request buffer in situations where no degraded stripes or pending write requests. This optimisation is implemented in 2.4 and gave me about a 10% speedup on sequential reads. Its substantially more complicated in 2.6... I might do a bit of testing myself, but more results from others would be helpful... NeilBrown ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow 2004-11-30 0:30 ` raid5 slow Neil Brown @ 2004-11-30 0:50 ` Stephan van Hienen 2004-11-30 19:37 ` Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-30 0:50 UTC (permalink / raw) To: Neil Brown; +Cc: Guy, linux-raid On Tue, 30 Nov 2004, Neil Brown wrote: > On Monday November 29, raid@a2000.nu wrote: >> >> now i get 275MB/sec with raid0 (14 disks) >> 171MB/sec with raid5 (14 disks) >> >> while the old raid5 situation was 257MB/sec with raid5 (13 disks) > > I presume you are only looking at 'read' speed here. yes i'm looking at the read spead just tested 2.6.10-rc2 142148 this is 50% of the performance i get with the 2.4 kernel (the 171MB/sec i got yesterday with 14 disks was also with xfs) (xfs on 14d r5 gives 171371 read/ext3 on 14 r5 gives 132845 read) (2.6.10-rc2 13d r5 ext3) Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 24730 99 108265 88 59460 59 25759 91 142148 69 586.8 3 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1432 78 +++++ +++ +++++ +++ 1456 77 +++++ +++ 4908 97 storage.a2000.nu,2G,24730,99,108265,88,59460,59,25759,91,142148,69,586.8,3,16,1432,78,+++++,+++,+++++,+++,1456,77,+++++,+++,4908,97 > > There is a read optimisation that I haven't implemented in 2.6 yet > that could account for about a 10% difference, but a 30% difference > definitely surprises me. > > Normally raid5 will read into the stripe cache, and then copy data out > of the stripe cache and into the request buffer for reads. > The optimisation is to read directly into the request buffer in > situations where no degraded stripes or pending write requests. This > optimisation is implemented in 2.4 and gave me about a 10% speedup on > sequential reads. Its substantially more complicated in 2.6... > > I might do a bit of testing myself, but more results from others would > be helpful... > > NeilBrown > ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow 2004-11-30 0:50 ` Stephan van Hienen @ 2004-11-30 19:37 ` Stephan van Hienen 2004-11-30 23:26 ` Neil Brown 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-30 19:37 UTC (permalink / raw) To: Neil Brown; +Cc: Guy, linux-raid today also runned a test with 500gb size : Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000. 500G 23385 97 79328 69 59189 58 25015 91 122493 58 67.7 11 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1404 76 +++++ +++ +++++ +++ 1347 75 +++++ +++ 4868 97 storage.a2000.nu,500G,23385,97,79328,69,59189,58,25015,91,122493,58,67.7,11,16,1404,76,+++++,+++,+++++,+++,1347,75,+++++,+++,4868,97 it's really slow... :( is there anything i can do ? i want to fill the raid asap for now i have 2 options : -use the 2.6 and hope someone is going to fix this ? -use the 2.4 kernel but max 13 devices (there is no LBD support in 2.4?) or is there a LBD patch so i can create a 2TB+ raid5 array ? ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow 2004-11-30 19:37 ` Stephan van Hienen @ 2004-11-30 23:26 ` Neil Brown 2004-11-30 23:39 ` Stephan van Hienen 2004-11-30 23:55 ` Stephan van Hienen 0 siblings, 2 replies; 25+ messages in thread From: Neil Brown @ 2004-11-30 23:26 UTC (permalink / raw) To: Stephan van Hienen; +Cc: Neil Brown, Guy, linux-raid On Tuesday November 30, raid@a2000.nu wrote: > storage.a2000.nu,500G,23385,97,79328,69,59189,58,25015,91,122493,58,67.7,11,16,1404,76,+++++,+++,+++++,+++,1347,75,+++++,+++,4868,97 > > it's really slow... :( > is there anything i can do ? > i want to fill the raid asap > > for now i have 2 options : > > -use the 2.6 and hope someone is going to fix this ? > > -use the 2.4 kernel but max 13 devices (there is no LBD support in 2.4?) > or is there a LBD patch so i can create a 2TB+ raid5 array ? > > I don't think there is LBD support for 2.4. I haven't managed to duplicate the degree of slowdown that you report - I only get about 10%. However I found an old Email which also reported slow raid5 in 2.6 and it noted that when the array is degraded, it goes faster .... I just tried it and got most of the speed back when the array is degraded!!! Would you be able to test that option? I will look into the code and see if I can figure out what is going on. NeilBrown ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow 2004-11-30 23:26 ` Neil Brown @ 2004-11-30 23:39 ` Stephan van Hienen 2004-11-30 23:55 ` Stephan van Hienen 1 sibling, 0 replies; 25+ messages in thread From: Stephan van Hienen @ 2004-11-30 23:39 UTC (permalink / raw) To: Neil Brown; +Cc: Guy, linux-raid > I haven't managed to duplicate the degree of slowdown that you report > - I only get about 10%. > > However I found an old Email which also reported slow raid5 in 2.6 and > it noted that when the array is degraded, it goes faster .... > > I just tried it and got most of the speed back when the array is > degraded!!! > Would you be able to test that option? 2 tests : (kernel 2.6.10-rc2 13d r5 ext2) Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 25238 99 116958 98 67987 73 25702 91 142010 70 608.2 3 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1354 78 +++++ +++ +++++ +++ 1262 74 +++++ +++ 4895 95 storage.a2000.nu,2G,25238,99,116958,98,67987,73,25702,91,142010,70,608.2,3,16,1354,78,+++++,+++,+++++,+++,1262,74,+++++,+++,4895,95 ]# mdadm --manage --fail /dev/md0 /dev/sdn1 mdadm: set /dev/sdn1 faulty in /dev/md0 ]# mdadm --manage --remove /dev/md0 /dev/sdn1 mdadm: hot removed /dev/sdn1 Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage.a2000.nu 2G 25458 99 107295 83 59079 60 26079 93 145775 77 490.2 3 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1423 83 +++++ +++ +++++ +++ 1453 82 +++++ +++ 4872 100 storage.a2000.nu,2G,25458,99,107295,83,59079,60,26079,93,145775,77,490.2,3,16,1423,83,+++++,+++,+++++,+++,1453,82,+++++,+++,4872,100 ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow 2004-11-30 23:26 ` Neil Brown 2004-11-30 23:39 ` Stephan van Hienen @ 2004-11-30 23:55 ` Stephan van Hienen 2004-12-01 19:23 ` raid5 slow (test on another system) Stephan van Hienen 1 sibling, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-11-30 23:55 UTC (permalink / raw) To: Neil Brown; +Cc: Guy, linux-raid On Wed, 1 Dec 2004, Neil Brown wrote: > I don't think there is LBD support for 2.4. found this : https://www.gelato.unsw.edu.au/archives/lbd/2004-February/000039.html http://www.gelato.unsw.edu.au/patches/ 'Theoretically raid0 and linear should work; I'm certain raid1, raid4 and raid5 will not work. In any case, RAID is limited to a maximum of 2TB minus one block per member. ' i don't think i want this on my raid5.... ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow (test on another system) 2004-11-30 23:55 ` Stephan van Hienen @ 2004-12-01 19:23 ` Stephan van Hienen 2004-12-01 22:21 ` Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-12-01 19:23 UTC (permalink / raw) To: Neil Brown; +Cc: linux-raid today runned a test on another machine (dual p3-733 with 4*9gb scsi) looks like on this system the difference is arround the 10% maybe you need more disks/faster disks to get a bigger difference results : 2.6.10-rc2 4d r5 ext3 Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP linuxtest.expl 496M 10364 99 39689 55 22184 36 11485 96 52559 46 632.0 4 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 512 99 +++++ +++ 32502 99 528 99 +++++ +++ 1938 97 linuxtest.explainerdc.local,496M,10364,99,39689,55,22184,36,11485,96,52559,46,632.0,4,16,512,99,+++++,+++,32502,99,528,99,+++++,+++,1938,97 2.4.20 4d r5 ext3 Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP linuxtest.expl 496M 10087 100 53021 63 20934 34 11746 98 58858 34 665.3 4 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 573 99 +++++ +++ 30886 99 592 99 +++++ +++ 2264 96 linuxtest.explainerdc.local,496M,10087,100,53021,63,20934,34,11746,98,58858,34,665.3,4,16,573,99,+++++,+++,30886,99,592,99,+++++,+++,2264,96 ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow (test on another system) 2004-12-01 19:23 ` raid5 slow (test on another system) Stephan van Hienen @ 2004-12-01 22:21 ` Stephan van Hienen 2004-12-02 22:01 ` Stephan van Hienen 0 siblings, 1 reply; 25+ messages in thread From: Stephan van Hienen @ 2004-12-01 22:21 UTC (permalink / raw) To: Neil Brown; +Cc: linux-raid and another system P4 2.4Ghz with 8*120gb on 3ware 7805 pci32 so the 2.4 looks like limited by the pci bus : (2.6.8 8d r5 ext3) Version 1.93c ------Sequential Output------ --Sequential Input- --Random- Concurrency 1 -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage 1G 507 98 64149 20 33240 11 984 97 95124 18 663.0 2 Latency 105ms 1039ms 879ms 29443us 63097us 870ms Version 1.93c ------Sequential Create------ --------Random Create-------- storage -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1849 64 +++++ +++ 10044 9 1570 54 +++++ +++ 4406 49 Latency 64064us 184us 188us 133ms 102us 293us 1.93c,1.93c,storage,1,1101940252,1G,,507,98,64149,20,33240,11,984,97,95124,18,663.0,2,16,,,,,1849,64,+++++,+++,10044,9,1570,54,+++++,+++,4406,49,105ms,1039ms,879ms,29443us,63097us,870ms,64064us,184us,188us,133ms,102us,293us (2.4.28 8d r5 ext3) Version 1.93c ------Sequential Output------ --Sequential Input- --Random- Concurrency 1 -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP storage 1G 98 99 64449 44 42805 21 742 98 113930 29 732.5 4 Latency 129ms 809ms 819ms 23221us 139ms 754ms Version 1.93c ------Sequential Create------ --------Random Create-------- storage -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 1526 61 +++++ +++ 8337 20 1940 78 +++++ +++ 2441 36 Latency 175ms 39us 83us 78763us 20us 77558us 1.93c,1.93c,storage,1,1101940608,1G,,98,99,64449,44,42805,21,742,98,113930,29,732.5,4,16,,,,,1526,61,+++++,+++,8337,20,1940,78,+++++,+++,2441,36,129ms,809ms,819ms,23221us,139ms,754ms,175ms,39us,83us,78763us,20us,77558us ^ permalink raw reply [flat|nested] 25+ messages in thread
* Re: raid5 slow (test on another system) 2004-12-01 22:21 ` Stephan van Hienen @ 2004-12-02 22:01 ` Stephan van Hienen 0 siblings, 0 replies; 25+ messages in thread From: Stephan van Hienen @ 2004-12-02 22:01 UTC (permalink / raw) To: Neil Brown; +Cc: linux-raid and another system p4 2.8Ghz with 2*300gb (normal ide ports) i created a degraded raid5 25% difference in read performance : 2.4.20 : Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP gx270.a2000.nu 1G 29724 79 41447 15 17575 6 33146 80 55494 12 267.6 0 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 2539 98 +++++ +++ +++++ +++ 2523 98 +++++ +++ 6307 96 gx270.a2000.nu,1G,29724,79,41447,15,17575,6,33146,80,55494,12,267.6,0,16,2539,98 ,+++++,+++,+++++,+++,2523,98,+++++,+++,6307,96 2.6.9-ac11 : Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP gx270.a2000.nu 1G 21190 52 37804 8 10092 2 34726 80 68587 9 263.1 0 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 3305 99 +++++ +++ +++++ +++ 3372 100 +++++ +++ 10428 98 gx270.a2000.nu,1G,21190,52,37804,8,10092,2,34726,80,68587,9,263.1,0,16,3305,99,+ ++++,+++,+++++,+++,3372,100,+++++,+++,10428,98 ^ permalink raw reply [flat|nested] 25+ messages in thread
end of thread, other threads:[~2004-12-02 22:01 UTC | newest] Thread overview: 25+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2004-11-28 10:48 raid5 code ok with 2TB + ? Stephan van Hienen 2004-11-28 10:57 ` Brad Campbell 2004-11-28 11:19 ` Stephan van Hienen 2004-11-28 11:28 ` Brad Campbell 2004-11-28 18:19 ` Guy 2004-11-28 19:20 ` Brad Campbell 2004-11-28 20:17 ` raid5 code ok with 2TB + ? NEGATIVE :( Stephan van Hienen 2004-11-28 20:27 ` Stephan van Hienen 2004-11-28 20:45 ` Stephan van Hienen 2004-11-28 23:34 ` raid5 code ok with 2TB + ? Stephan van Hienen 2004-11-29 4:49 ` Guy 2004-11-29 8:33 ` Stephan van Hienen 2004-11-29 8:51 ` raid5 slow Stephan van Hienen 2004-11-29 17:13 ` Guy 2004-11-29 18:11 ` Stephan van Hienen 2004-11-30 0:06 ` raid5 slow (looks like 2.6 problem) Stephan van Hienen 2004-11-30 0:30 ` raid5 slow Neil Brown 2004-11-30 0:50 ` Stephan van Hienen 2004-11-30 19:37 ` Stephan van Hienen 2004-11-30 23:26 ` Neil Brown 2004-11-30 23:39 ` Stephan van Hienen 2004-11-30 23:55 ` Stephan van Hienen 2004-12-01 19:23 ` raid5 slow (test on another system) Stephan van Hienen 2004-12-01 22:21 ` Stephan van Hienen 2004-12-02 22:01 ` Stephan van Hienen
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).