* Stress testing system? @ 2004-10-08 21:32 Robin Bowes 2004-10-08 22:00 ` Mike Hardy 2004-10-08 22:02 ` Gordon Henderson 0 siblings, 2 replies; 15+ messages in thread From: Robin Bowes @ 2004-10-08 21:32 UTC (permalink / raw) To: linux-raid Hi, I've got six 250GB Maxtor drives connected to 2 Promise SATA controllers configured as follows: Each disk has two partitions: 1.5G and 248.5G. /dev/sda1 & /dev/sdd1 are mirrored and form the root filesystem. /dev/sd[abcdef]2 are configured as a RAID5 array with one hot spare. I use lvm to create a 10G /usr partition, a 5G /var partition, and the rest of the array (994G) in /home. The system in which I installed these drives was rock-solid before I added the RAID storage (it had a single 120G drive). However, since adding the 6 disks I have experienced the system simply powering down and requiring filesystem recovering when it restarted. I suspected this was down to an inadequate power supply (it was 400W) so I've upgrade to an OCZ 520W PSU. I'd like to stress test the system to see if the new PSU has sorted the problem, i.e. really work the disks. What's the best way to do get all six drives working as hard as possible? Thanks, R. -- http://robinbowes.com ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-08 21:32 Stress testing system? Robin Bowes @ 2004-10-08 22:00 ` Mike Hardy 2004-10-08 22:07 ` Gordon Henderson 2004-10-08 23:49 ` Guy 2004-10-08 22:02 ` Gordon Henderson 1 sibling, 2 replies; 15+ messages in thread From: Mike Hardy @ 2004-10-08 22:00 UTC (permalink / raw) To: 'linux-raid@vger.kernel.org' This is a little off topic, but I stress systems with four loops loop one unpacks the kernel source, moves it to a new name, unpacks the kernel source again, and diffs the two, deletes and repeats (tests memory, disk caching) loop two unpacks the kernel source, does a make allconfig and a make -j5 bzImage modules, then a make clean and repeats. should get the cpu burning loop three should run bonnie++ on the array loop four should work with another machine. each machine should wget some very large file (1GBish) with output to /dev/null so that the NIC has to serve interrupts at max If that doesn't cook your machine in 48 hours or so, I can't think of anything that will. This catches out every machine I try it on for some reason or another, but after a couple tweaks its usually solid. Slightly more on-topic, one thing that I have do to frequently is boot with noapic or acpi=off due to interrupt handling problems with various motherboards. Additionally, I think there have been reports of problems with raid and LVM, and there have also been problems with SATA and possibly with Maxtor drives, so you have may have some tweaking to do. Mentioning versions of things (distribution, kernel, hardware parts and part numbers etc) would help I'm interested to hear what other people do to burn their machines in though... Good luck. -Mike Robin Bowes wrote: > Hi, > > I've got six 250GB Maxtor drives connected to 2 Promise SATA controllers > configured as follows: > > Each disk has two partitions: 1.5G and 248.5G. > > /dev/sda1 & /dev/sdd1 are mirrored and form the root filesystem. > > /dev/sd[abcdef]2 are configured as a RAID5 array with one hot spare. > > I use lvm to create a 10G /usr partition, a 5G /var partition, and the > rest of the array (994G) in /home. > > The system in which I installed these drives was rock-solid before I > added the RAID storage (it had a single 120G drive). However, since > adding the 6 disks I have experienced the system simply powering down > and requiring filesystem recovering when it restarted. > > I suspected this was down to an inadequate power supply (it was 400W) so > I've upgrade to an OCZ 520W PSU. > > I'd like to stress test the system to see if the new PSU has sorted the > problem, i.e. really work the disks. > > What's the best way to do get all six drives working as hard as possible? > > Thanks, > > R. ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-08 22:00 ` Mike Hardy @ 2004-10-08 22:07 ` Gordon Henderson 2004-10-08 23:49 ` Guy 1 sibling, 0 replies; 15+ messages in thread From: Gordon Henderson @ 2004-10-08 22:07 UTC (permalink / raw) To: Mike Hardy; +Cc: 'linux-raid@vger.kernel.org' On Fri, 8 Oct 2004, Mike Hardy wrote: > Slightly more on-topic, one thing that I have do to frequently is boot > with noapic or acpi=off due to interrupt handling problems with various > motherboards. I've seen this (and made mention of it here in the past) it basically boils down to cheap motherboards with buggy chipsets )-: Some BIOSes let you select PIC or APIC - on these buggy ones, I stick to PIC. > Additionally, I think there have been reports of problems with raid and > LVM, and there have also been problems with SATA and possibly with > Maxtor drives, so you have may have some tweaking to do. Mentioning > versions of things (distribution, kernel, hardware parts and part > numbers etc) would help I had problems some 2 years ago when I first dabbled with LVM. Quickly found problems with performance and random crashes and not having time to persue it further, left it alone and haven't ventured back since! > I'm interested to hear what other people do to burn their machines in > though... Looks like we do someting similar... So that can't be bad! I'm sure there is plenty cpu left over for me to do stuff like compiles, etc. Many years ago I used to write diagnostics for compter systems - memory, cpu, IO, etc. there were cases where the systems would pass all the diags faultlessly for days on end, then crash 2 minutes into the application, so theres no substitute for real life testing... Gordon ^ permalink raw reply [flat|nested] 15+ messages in thread
* RE: Stress testing system? 2004-10-08 22:00 ` Mike Hardy 2004-10-08 22:07 ` Gordon Henderson @ 2004-10-08 23:49 ` Guy 1 sibling, 0 replies; 15+ messages in thread From: Guy @ 2004-10-08 23:49 UTC (permalink / raw) To: linux-raid When using wget, run it on both systems (full duplex). This will transfer about twice as much data per second. Assuming you have a full duplex network! If not, switches are very low cost today. I think this is what you intended, but your note was vague IMO. Guy -----Original Message----- From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Mike Hardy Sent: Friday, October 08, 2004 6:00 PM To: 'linux-raid@vger.kernel.org' Subject: Re: Stress testing system? This is a little off topic, but I stress systems with four loops loop one unpacks the kernel source, moves it to a new name, unpacks the kernel source again, and diffs the two, deletes and repeats (tests memory, disk caching) loop two unpacks the kernel source, does a make allconfig and a make -j5 bzImage modules, then a make clean and repeats. should get the cpu burning loop three should run bonnie++ on the array loop four should work with another machine. each machine should wget some very large file (1GBish) with output to /dev/null so that the NIC has to serve interrupts at max If that doesn't cook your machine in 48 hours or so, I can't think of anything that will. This catches out every machine I try it on for some reason or another, but after a couple tweaks its usually solid. Slightly more on-topic, one thing that I have do to frequently is boot with noapic or acpi=off due to interrupt handling problems with various motherboards. Additionally, I think there have been reports of problems with raid and LVM, and there have also been problems with SATA and possibly with Maxtor drives, so you have may have some tweaking to do. Mentioning versions of things (distribution, kernel, hardware parts and part numbers etc) would help I'm interested to hear what other people do to burn their machines in though... Good luck. -Mike Robin Bowes wrote: > Hi, > > I've got six 250GB Maxtor drives connected to 2 Promise SATA controllers > configured as follows: > > Each disk has two partitions: 1.5G and 248.5G. > > /dev/sda1 & /dev/sdd1 are mirrored and form the root filesystem. > > /dev/sd[abcdef]2 are configured as a RAID5 array with one hot spare. > > I use lvm to create a 10G /usr partition, a 5G /var partition, and the > rest of the array (994G) in /home. > > The system in which I installed these drives was rock-solid before I > added the RAID storage (it had a single 120G drive). However, since > adding the 6 disks I have experienced the system simply powering down > and requiring filesystem recovering when it restarted. > > I suspected this was down to an inadequate power supply (it was 400W) so > I've upgrade to an OCZ 520W PSU. > > I'd like to stress test the system to see if the new PSU has sorted the > problem, i.e. really work the disks. > > What's the best way to do get all six drives working as hard as possible? > > Thanks, > > R. - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-08 21:32 Stress testing system? Robin Bowes 2004-10-08 22:00 ` Mike Hardy @ 2004-10-08 22:02 ` Gordon Henderson 2004-10-08 23:44 ` Robin Bowes 1 sibling, 1 reply; 15+ messages in thread From: Gordon Henderson @ 2004-10-08 22:02 UTC (permalink / raw) To: Robin Bowes; +Cc: linux-raid On Fri, 8 Oct 2004, Robin Bowes wrote: > What's the best way to do get all six drives working as hard as possible? I always run 'bonnie' on each partition (sometimes 2 to a partition) when soak-testing a new server. Try to leave it running for as long as possible. (ie. days) That seems to get the disk heads moving (start the bonnies off at 1-2 minute interfals, but they desynchronise over time anyway, so some end up reading, some writing, and some seeking) and I suspect head movement is whats going to cause a disk drive to comsume the most current after startup. If I have a handy PC next to it, then I'll also run some scripted network FTPs (via wget) of 1GB binary files copying to /dev/null on the recipient side. I certinaly see the motherboard temperature rise when I do this (lm-sensors is your friend, but can be a PITA to get going) Gordon ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-08 22:02 ` Gordon Henderson @ 2004-10-08 23:44 ` Robin Bowes 2004-10-08 23:48 ` Guy 2004-10-10 20:36 ` Gordon Henderson 0 siblings, 2 replies; 15+ messages in thread From: Robin Bowes @ 2004-10-08 23:44 UTC (permalink / raw) To: Gordon Henderson; +Cc: linux-raid Gordon Henderson wrote: > On Fri, 8 Oct 2004, Robin Bowes wrote: > > >>What's the best way to do get all six drives working as hard as possible? > > > I always run 'bonnie' on each partition (sometimes 2 to a partition) when > soak-testing a new server. Try to leave it running for as long as > possible. (ie. days) Hi Gordon, I tried this - just a simple command to start with: # bonnie++ -d /home -s10 -r4 -u0 This gave the following results: Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP dude.robinbowes 10M 11482 92 +++++ +++ +++++ +++ 15370 100 +++++ +++ 13406 124 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 347 88 +++++ +++ 19794 91 332 86 +++++ +++ 1106 93 dude.robinbowes.com,10M,11482,92,+++++,+++,+++++,+++,15370,100,+++++,+++,13406.0 ,124,16,347,88,+++++,+++,19794,91,332,86,+++++,+++,1106,93 I then noticed that my raid array was using a lot of CPU: top - 00:41:28 up 33 min, 2 users, load average: 1.80, 1.78, 1.57 Tasks: 89 total, 1 running, 88 sleeping, 0 stopped, 0 zombie Cpu(s): 4.1% us, 32.9% sy, 0.0% ni, 59.8% id, 0.0% wa, 0.5% hi, 2.6% si Mem: 1554288k total, 368212k used, 1186076k free, 70520k buffers Swap: 0k total, 0k used, 0k free, 200140k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 239 root 15 0 0 0 0 S 61.0 0.0 20:08.37 md5_raid5 1414 slimserv 15 0 43644 38m 5772 S 9.0 2.5 2:38.99 slimserver.pl 241 root 15 0 0 0 0 D 6.3 0.0 2:05.45 md5_resync 1861 root 16 0 2888 908 1620 R 1.0 0.1 0:00.28 top 1826 root 16 0 9332 2180 4232 S 0.3 0.1 0:00.28 sshd So I checked the array: [root@dude root]# mdadm --detail /dev/md5 /dev/md5: Version : 00.90.01 Creation Time : Thu Jul 29 21:41:38 2004 Raid Level : raid5 Array Size : 974566400 (929.42 GiB 997.96 GB) Device Size : 243641600 (232.35 GiB 249.49 GB) Raid Devices : 5 Total Devices : 6 Preferred Minor : 5 Persistence : Superblock is persistent Update Time : Sat Oct 9 00:08:22 2004 State : dirty, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 128K Rebuild Status : 12% complete UUID : a4bbcd09:5e178c5b:3bf8bd45:8c31d2a1 Events : 0.1410301 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 1 8 18 1 active sync /dev/sdb2 2 8 34 2 active sync /dev/sdc2 3 8 50 3 active sync /dev/sdd2 4 8 66 4 active sync /dev/sde2 5 8 82 - spare /dev/sdf2 Is this normal? Should running bonnie++ result in the array being dirty and requiring resyncing? R. -- http://robinbowes.com ^ permalink raw reply [flat|nested] 15+ messages in thread
* RE: Stress testing system? 2004-10-08 23:44 ` Robin Bowes @ 2004-10-08 23:48 ` Guy 2004-10-09 9:52 ` Robin Bowes 2004-10-10 20:36 ` Gordon Henderson 1 sibling, 1 reply; 15+ messages in thread From: Guy @ 2004-10-08 23:48 UTC (permalink / raw) To: 'Robin Bowes', 'Gordon Henderson'; +Cc: linux-raid Not normal. Was it ever synced? Wait for it to re-sync, then reboot and check the status. Some people have problems after a reboot. Cat /proc/mdstat for the ETA and status. Guy -----Original Message----- From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Robin Bowes Sent: Friday, October 08, 2004 7:44 PM To: Gordon Henderson Cc: linux-raid@vger.kernel.org Subject: Re: Stress testing system? Gordon Henderson wrote: > On Fri, 8 Oct 2004, Robin Bowes wrote: > > >>What's the best way to do get all six drives working as hard as possible? > > > I always run 'bonnie' on each partition (sometimes 2 to a partition) when > soak-testing a new server. Try to leave it running for as long as > possible. (ie. days) Hi Gordon, I tried this - just a simple command to start with: # bonnie++ -d /home -s10 -r4 -u0 This gave the following results: Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP dude.robinbowes 10M 11482 92 +++++ +++ +++++ +++ 15370 100 +++++ +++ 13406 124 ------Sequential Create------ --------Random Create-------- -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete-- files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP 16 347 88 +++++ +++ 19794 91 332 86 +++++ +++ 1106 93 dude.robinbowes.com,10M,11482,92,+++++,+++,+++++,+++,15370,100,+++++,+++,134 06.0 ,124,16,347,88,+++++,+++,19794,91,332,86,+++++,+++,1106,93 I then noticed that my raid array was using a lot of CPU: top - 00:41:28 up 33 min, 2 users, load average: 1.80, 1.78, 1.57 Tasks: 89 total, 1 running, 88 sleeping, 0 stopped, 0 zombie Cpu(s): 4.1% us, 32.9% sy, 0.0% ni, 59.8% id, 0.0% wa, 0.5% hi, 2.6% si Mem: 1554288k total, 368212k used, 1186076k free, 70520k buffers Swap: 0k total, 0k used, 0k free, 200140k cached PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND 239 root 15 0 0 0 0 S 61.0 0.0 20:08.37 md5_raid5 1414 slimserv 15 0 43644 38m 5772 S 9.0 2.5 2:38.99 slimserver.pl 241 root 15 0 0 0 0 D 6.3 0.0 2:05.45 md5_resync 1861 root 16 0 2888 908 1620 R 1.0 0.1 0:00.28 top 1826 root 16 0 9332 2180 4232 S 0.3 0.1 0:00.28 sshd So I checked the array: [root@dude root]# mdadm --detail /dev/md5 /dev/md5: Version : 00.90.01 Creation Time : Thu Jul 29 21:41:38 2004 Raid Level : raid5 Array Size : 974566400 (929.42 GiB 997.96 GB) Device Size : 243641600 (232.35 GiB 249.49 GB) Raid Devices : 5 Total Devices : 6 Preferred Minor : 5 Persistence : Superblock is persistent Update Time : Sat Oct 9 00:08:22 2004 State : dirty, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 128K Rebuild Status : 12% complete UUID : a4bbcd09:5e178c5b:3bf8bd45:8c31d2a1 Events : 0.1410301 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 1 8 18 1 active sync /dev/sdb2 2 8 34 2 active sync /dev/sdc2 3 8 50 3 active sync /dev/sdd2 4 8 66 4 active sync /dev/sde2 5 8 82 - spare /dev/sdf2 Is this normal? Should running bonnie++ result in the array being dirty and requiring resyncing? R. -- http://robinbowes.com - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-08 23:48 ` Guy @ 2004-10-09 9:52 ` Robin Bowes 2004-10-09 16:58 ` Guy 0 siblings, 1 reply; 15+ messages in thread From: Robin Bowes @ 2004-10-09 9:52 UTC (permalink / raw) To: Guy; +Cc: 'Gordon Henderson', linux-raid Guy wrote: > Not normal. Was it ever synced? > Wait for it to re-sync, then reboot and check the status. > Some people have problems after a reboot. > > Cat /proc/mdstat for the ETA and status. Well, the array re-synced overnight and I re-ran bonnie++ again a couple of times this morning with no apparent ill-effect. To be honest, I'd just had some problems with a loose power cable so I suspect that the array was already re-syncing before I ran bonnie. Thanks, R. -- http://robinbowes.com ^ permalink raw reply [flat|nested] 15+ messages in thread
* RE: Stress testing system? 2004-10-09 9:52 ` Robin Bowes @ 2004-10-09 16:58 ` Guy 2004-10-09 17:19 ` Robin Bowes 0 siblings, 1 reply; 15+ messages in thread From: Guy @ 2004-10-09 16:58 UTC (permalink / raw) To: 'Robin Bowes'; +Cc: linux-raid Once a drive fails, md will not re-sync it automatically. It will just sit there in a failed state. If you had a spare, then it would re-sync automatically. I am not 100% sure but I think...if you were to reboot, after the reboot md will resync. If you reboot before the re-sync is done, the re-sync will start over, at lease with my version. Guy -----Original Message----- From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Robin Bowes Sent: Saturday, October 09, 2004 5:52 AM To: Guy Cc: 'Gordon Henderson'; linux-raid@vger.kernel.org Subject: Re: Stress testing system? Guy wrote: > Not normal. Was it ever synced? > Wait for it to re-sync, then reboot and check the status. > Some people have problems after a reboot. > > Cat /proc/mdstat for the ETA and status. Well, the array re-synced overnight and I re-ran bonnie++ again a couple of times this morning with no apparent ill-effect. To be honest, I'd just had some problems with a loose power cable so I suspect that the array was already re-syncing before I ran bonnie. Thanks, R. -- http://robinbowes.com - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-09 16:58 ` Guy @ 2004-10-09 17:19 ` Robin Bowes 0 siblings, 0 replies; 15+ messages in thread From: Robin Bowes @ 2004-10-09 17:19 UTC (permalink / raw) To: Guy; +Cc: linux-raid Guy wrote: > Once a drive fails, md will not re-sync it automatically. It will just sit > there in a failed state. If you had a spare, then it would re-sync > automatically. I am not 100% sure but I think...if you were to reboot, > after the reboot md will resync. > > If you reboot before the re-sync is done, the re-sync will start over, at > lease with my version. I think that's what must have happened. I replaced the power supply and brought the box back up. It then froze on me, so I powered down again and investigated, bring the box up and down a few times in the process. I then noticed the loose power connection, fixed it, and brought the box back up. I reckon the array must have been re-syncing from that point and it had nothing to do with bonnie++. Incidentally, I do have a spare drive: [root@dude home]# mdadm --detail /dev/md5 /dev/md5: Version : 00.90.01 Creation Time : Thu Jul 29 21:41:38 2004 Raid Level : raid5 Array Size : 974566400 (929.42 GiB 997.96 GB) Device Size : 243641600 (232.35 GiB 249.49 GB) Raid Devices : 5 Total Devices : 6 Preferred Minor : 5 Persistence : Superblock is persistent Update Time : Sat Oct 9 18:18:54 2004 State : clean Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 128K UUID : a4bbcd09:5e178c5b:3bf8bd45:8c31d2a1 Events : 0.1410301 Number Major Minor RaidDevice State 0 8 2 0 active sync /dev/sda2 1 8 18 1 active sync /dev/sdb2 2 8 34 2 active sync /dev/sdc2 3 8 50 3 active sync /dev/sdd2 4 8 66 4 active sync /dev/sde2 5 8 82 - spare /dev/sdf2 Cheers, R. -- http://robinbowes.com ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-08 23:44 ` Robin Bowes 2004-10-08 23:48 ` Guy @ 2004-10-10 20:36 ` Gordon Henderson 2004-10-10 21:35 ` Robin Bowes 1 sibling, 1 reply; 15+ messages in thread From: Gordon Henderson @ 2004-10-10 20:36 UTC (permalink / raw) To: Robin Bowes; +Cc: linux-raid On Sat, 9 Oct 2004, Robin Bowes wrote: > I tried this - just a simple command to start with: > > # bonnie++ -d /home -s10 -r4 -u0 > > This gave the following results: > > Version 1.03 ------Sequential Output------ --Sequential Input- --Random- > -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- > Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP > dude.robinbowes 10M 11482 92 +++++ +++ +++++ +++ 15370 100 +++++ +++ 13406 124 Why are you removing the speed - is it something to be embarased about? If it's very slow, are you sure the devices are operating in a DMA mode? POI mode will use a lot of CPU and geneally make things really clunky.. > Is this normal? Should running bonnie++ result in the array being dirty > and requiring resyncing? No - but reading some of the later replies it seems it might not have been fully synced to start with? Have you let it sync now and run the tests again? Ah right - I've just run that bonnie myself - it's +++'d out the times as 10MB is really too small a file to do anything accurate with and you've told it you only have 4MB of RAM. It'll all end up in memory cache. I got similar results with that command. Don't bother with the -n option, and do get it to use a filesize of double your RAM size. You really just want to move data into & out of the disks, who cares (at this point) about actual file, seek, etc. IO. I use the following scripts when testing: /usr/local/bin/doBon: #!/bin/csh @ n = 1 while (1) echo Pass number $n bonnie -u0 -g0 -n0 -s 1024 @ n = $n + 1 end /usr/local/bin/doBon2: #!/bin/csh doBon & sleep 120 doBon and usually run a "doBon2" on each partition. Memory size here is 512MB. Gordon Ps. stop this with killall doBon2 ; killall doBon ; killall Bonnie ... then rm the Bonnie* files... ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-10 20:36 ` Gordon Henderson @ 2004-10-10 21:35 ` Robin Bowes 2004-10-10 22:38 ` Guy 2004-10-11 8:38 ` Gordon Henderson 0 siblings, 2 replies; 15+ messages in thread From: Robin Bowes @ 2004-10-10 21:35 UTC (permalink / raw) To: Gordon Henderson; +Cc: linux-raid Gordon Henderson wrote: > On Sat, 9 Oct 2004, Robin Bowes wrote: >>Version 1.03 ------Sequential Output------ --Sequential Input- --Random- >> -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- >>Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP >>dude.robinbowes 10M 11482 92 +++++ +++ +++++ +++ 15370 100 +++++ +++ 13406 124 > > > Why are you removing the speed - is it something to be embarased about? As you found out, bonnie does this without any carbon-based intervention! >>Is this normal? Should running bonnie++ result in the array being dirty >>and requiring resyncing? > > > No - but reading some of the later replies it seems it might not have been > fully synced to start with? On reflection, I'm pretty sure it wasn't. It is now. > Have you let it sync now and run the tests again? Yes. It was faster when the array had re-synced :) > Ah right - I've just run that bonnie myself - it's +++'d out the times as > 10MB is really too small a file to do anything accurate with and you've > told it you only have 4MB of RAM. It'll all end up in memory cache. I got > similar results with that command. > > Don't bother with the -n option, and do get it to use a filesize of double > your RAM size. You really just want to move data into & out of the disks, > who cares (at this point) about actual file, seek, etc. IO. I use the > following scripts when testing: > > /usr/local/bin/doBon: > > #!/bin/csh > @ n = 1 > while (1) > echo Pass number $n > bonnie -u0 -g0 -n0 -s 1024 > @ n = $n + 1 > end > > /usr/local/bin/doBon2: > > #!/bin/csh > doBon & sleep 120 > doBon > > and usually run a "doBon2" on each partition. Memory size here is 512MB. OK, I've tried: bonnie++ -d /home -u0 -g0 -n0 -s 3096 (I've got 1.5G of RAM here - RAM's so cheap it's daft not to!) This gave the following results: Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP dude.robinbow 3096M 13081 95 34159 75 12617 21 15311 92 40429 30 436.1 3 dude.robinbowes.com,3096M,13081,95,34159,75,12617,21,15311,92,40429,30,436.1,3,,,,,,,,,,,,, I don't actually know what the figures mean - is this fast?? R. -- http://robinbowes.com ^ permalink raw reply [flat|nested] 15+ messages in thread
* RE: Stress testing system? 2004-10-10 21:35 ` Robin Bowes @ 2004-10-10 22:38 ` Guy 2004-10-11 8:38 ` Gordon Henderson 1 sibling, 0 replies; 15+ messages in thread From: Guy @ 2004-10-10 22:38 UTC (permalink / raw) To: 'Robin Bowes', 'Gordon Henderson'; +Cc: linux-raid My system is a 500 Mhz P3 with 2 CPUs and 512 Meg ram. My array is faster! Hehe :) I have a 14 disk raid5 array. 18 Gig SCSI disks, 3 SCSI buses. bonnie++ -u0 -g0 -n0 -s 1024 Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP watkins-ho 1G 3293 97 34587 89 22306 53 3549 99 67492 63 500.1 9 dude.ro 3096M 13081 95 34159 75 12617 21 15311 92 40429 30 436.1 3 2 at the same time. This used both CPUs. Over 100 Meg per second reads! watkins-ho 1G 3091 85 18593 44 9733 24 3443 97 59895 60 249.6 6 watkins-ho 1G 2980 87 21176 54 10167 23 3478 99 44525 44 384.2 9 ---- ----- ----- ---- ------ ----- Total 6071 39769 19900 6921 104420 633.8 You win on "Per Chr", this is CPU bound since it reads only 1 byte at a time. This is more of a CPU speed test than a disk speed test, IMHO. During the "Per Chr" test, only 1 CPU had a load, it was at about 100%. My guess is you have a real computer! Maybe 1.5 Ghz. In the other tests your CPU usage was lower, which is good for you. Ramdom seeks... My guess is having 14 moving heads helps me a lot on this one! Since my disks are old. But they are 10,000 RPM. The bottom line: I don't know if my array is considered fast. I bet my array is slow. Today disks are so much faster than what I have. But I have more of them which helps performance. Guy -----Original Message----- From: linux-raid-owner@vger.kernel.org [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Robin Bowes Sent: Sunday, October 10, 2004 5:35 PM To: Gordon Henderson Cc: linux-raid@vger.kernel.org Subject: Re: Stress testing system? Gordon Henderson wrote: > On Sat, 9 Oct 2004, Robin Bowes wrote: >>Version 1.03 ------Sequential Output------ --Sequential Input- --Random- >> -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- >>Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP >>dude.robinbowes 10M 11482 92 +++++ +++ +++++ +++ 15370 100 +++++ +++ 13406 124 > > > Why are you removing the speed - is it something to be embarased about? As you found out, bonnie does this without any carbon-based intervention! >>Is this normal? Should running bonnie++ result in the array being dirty >>and requiring resyncing? > > > No - but reading some of the later replies it seems it might not have been > fully synced to start with? On reflection, I'm pretty sure it wasn't. It is now. > Have you let it sync now and run the tests again? Yes. It was faster when the array had re-synced :) > Ah right - I've just run that bonnie myself - it's +++'d out the times as > 10MB is really too small a file to do anything accurate with and you've > told it you only have 4MB of RAM. It'll all end up in memory cache. I got > similar results with that command. > > Don't bother with the -n option, and do get it to use a filesize of double > your RAM size. You really just want to move data into & out of the disks, > who cares (at this point) about actual file, seek, etc. IO. I use the > following scripts when testing: > > /usr/local/bin/doBon: > > #!/bin/csh > @ n = 1 > while (1) > echo Pass number $n > bonnie -u0 -g0 -n0 -s 1024 > @ n = $n + 1 > end > > /usr/local/bin/doBon2: > > #!/bin/csh > doBon & sleep 120 > doBon > > and usually run a "doBon2" on each partition. Memory size here is 512MB. OK, I've tried: bonnie++ -d /home -u0 -g0 -n0 -s 3096 (I've got 1.5G of RAM here - RAM's so cheap it's daft not to!) This gave the following results: Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP dude.robinbow 3096M 13081 95 34159 75 12617 21 15311 92 40429 30 436.1 3 dude.robinbowes.com,3096M,13081,95,34159,75,12617,21,15311,92,40429,30,436.1 ,3,,,,,,,,,,,,, I don't actually know what the figures mean - is this fast?? R. -- http://robinbowes.com - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-10 21:35 ` Robin Bowes 2004-10-10 22:38 ` Guy @ 2004-10-11 8:38 ` Gordon Henderson 2004-10-11 9:01 ` Brad Campbell 1 sibling, 1 reply; 15+ messages in thread From: Gordon Henderson @ 2004-10-11 8:38 UTC (permalink / raw) To: Robin Bowes; +Cc: linux-raid On Sun, 10 Oct 2004, Robin Bowes wrote: > This gave the following results: > > Version 1.03 ------Sequential Output------ --Sequential Input- --Random- > -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- > Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP > dude.robinbow 3096M 13081 95 34159 75 12617 21 15311 92 40429 30 436.1 3 > dude.robinbowes.com,3096M,13081,95,34159,75,12617,21,15311,92,40429,30,436.1,3,,,,,,,,,,,,, > > I don't actually know what the figures mean - is this fast?? It's not brilliant, but reasonable for a RAID5 on IDE drives. Disk head bandwidth for comodity 7200 RPM drives is about 55MB/sec peak - although I haven't been able to get that with the SATA controllers I've used so-far, but I suspect thats because they are running in PATA mode. This is a Dell with 4 SCSI drivers split over 2 controllers: Version 1.02b ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP pixel 2000M 20492 72 54865 12 28260 6 25413 86 108190 13 332.9 0 This is a dual-Athlon with 5 IDE drives (4+hot spare) Version 1.02b ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP red 1023M 14216 99 26663 25 17697 13 13767 95 72211 34 239.5 2 This is anothe dual athlon with just 4 IDE drives: Version 1.02b ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP blue 1023M 12003 99 40147 44 24048 19 12078 98 87887 46 232.6 2 Your block input seems a shade low, but this is what I experienced on a server with SATA drives which look like /dev/hdX drives. I suspect the drivers have a bit more development to go through though. And in any-case, depending on what you are using it for, it's probably fast enough anyway... 100Mb Ethernet can only chuck files out at 10MB/sec anyway, but it's always nice to have bandwidth in-hand! Gordon ^ permalink raw reply [flat|nested] 15+ messages in thread
* Re: Stress testing system? 2004-10-11 8:38 ` Gordon Henderson @ 2004-10-11 9:01 ` Brad Campbell 0 siblings, 0 replies; 15+ messages in thread From: Brad Campbell @ 2004-10-11 9:01 UTC (permalink / raw) To: Gordon Henderson; +Cc: Robin Bowes, linux-raid Gordon Henderson wrote: > On Sun, 10 Oct 2004, Robin Bowes wrote: > > >>This gave the following results: >> >>Version 1.03 ------Sequential Output------ --Sequential Input- --Random- >> -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- >>Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP >>dude.robinbow 3096M 13081 95 34159 75 12617 21 15311 92 40429 30 436.1 3 >>dude.robinbowes.com,3096M,13081,95,34159,75,12617,21,15311,92,40429,30,436.1,3,,,,,,,,,,,,, >> >>I don't actually know what the figures mean - is this fast?? > > > It's not brilliant, but reasonable for a RAID5 on IDE drives. Disk head > bandwidth for comodity 7200 RPM drives is about 55MB/sec peak - although I > haven't been able to get that with the SATA controllers I've used so-far, > but I suspect thats because they are running in PATA mode. > Just a point of reference. This is on a single Athlon 2600+ with 10 7200 RPM Maxtor drives on 3 Promise SATA150TX4 controllers Version 1.03 ------Sequential Output------ --Sequential Input- --Random- -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks-- Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP srv 1G 14502 45 23251 12 17102 9 23390 67 67688 24 638.4 1 srv,1G,14502,45,23251,12,17102,9,23390,67,67688,24,638.4,1,,,,,,,,,,,,, Brad ^ permalink raw reply [flat|nested] 15+ messages in thread
end of thread, other threads:[~2004-10-11 9:01 UTC | newest] Thread overview: 15+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2004-10-08 21:32 Stress testing system? Robin Bowes 2004-10-08 22:00 ` Mike Hardy 2004-10-08 22:07 ` Gordon Henderson 2004-10-08 23:49 ` Guy 2004-10-08 22:02 ` Gordon Henderson 2004-10-08 23:44 ` Robin Bowes 2004-10-08 23:48 ` Guy 2004-10-09 9:52 ` Robin Bowes 2004-10-09 16:58 ` Guy 2004-10-09 17:19 ` Robin Bowes 2004-10-10 20:36 ` Gordon Henderson 2004-10-10 21:35 ` Robin Bowes 2004-10-10 22:38 ` Guy 2004-10-11 8:38 ` Gordon Henderson 2004-10-11 9:01 ` Brad Campbell
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).