* bad performance on RAID 5
@ 2007-01-17 21:33 Sevrin Robstad
2007-01-17 22:38 ` Justin Piszcz
` (2 more replies)
0 siblings, 3 replies; 16+ messages in thread
From: Sevrin Robstad @ 2007-01-17 21:33 UTC (permalink / raw)
To: linux-raid
I'm suffering from bad performance on my RAID5.
a "echo check >/sys/block/md0/md/sync_action"
gives a speed at only about 5000K/sec , and HIGH load average :
# uptime
20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
kernel is 2.6.18.1.2257.fc5
mdadm is v2.5.5
the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA PCI
cards with a total of 6 250gb S-ATA drives connected.
[root@compaq ~]# mdadm --detail /dev/md0
/dev/md0:
Version : 00.90.03
Creation Time : Tue Dec 5 00:33:01 2006
Raid Level : raid5
Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
Device Size : 243786240 (232.49 GiB 249.64 GB)
Raid Devices : 6
Total Devices : 6
Preferred Minor : 0
Persistence : Superblock is persistent
Update Time : Wed Jan 17 23:14:39 2007
State : clean
Active Devices : 6
Working Devices : 6
Failed Devices : 0
Spare Devices : 0
Layout : left-symmetric
Chunk Size : 256K
UUID : 27dce477:6f45d11b:77377d08:732fa0e6
Events : 0.58
Number Major Minor RaidDevice State
0 8 1 0 active sync /dev/sda1
1 8 17 1 active sync /dev/sdb1
2 8 33 2 active sync /dev/sdc1
3 8 49 3 active sync /dev/sdd1
4 8 65 4 active sync /dev/sde1
5 8 81 5 active sync /dev/sdf1
[root@compaq ~]#
Sevrin
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-17 21:33 bad performance on RAID 5 Sevrin Robstad
@ 2007-01-17 22:38 ` Justin Piszcz
2007-01-18 9:06 ` Raz Ben-Jehuda(caro)
2007-01-18 18:19 ` Sevrin Robstad
2007-01-18 18:29 ` Steve Cousins
2007-01-19 1:33 ` dean gaudet
2 siblings, 2 replies; 16+ messages in thread
From: Justin Piszcz @ 2007-01-17 22:38 UTC (permalink / raw)
To: Sevrin Robstad; +Cc: linux-raid
Sevrin Robstad wrote:
> I'm suffering from bad performance on my RAID5.
>
> a "echo check >/sys/block/md0/md/sync_action"
>
> gives a speed at only about 5000K/sec , and HIGH load average :
>
> # uptime
> 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
>
> kernel is 2.6.18.1.2257.fc5
> mdadm is v2.5.5
>
> the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
> PCI cards with a total of 6 250gb S-ATA drives connected.
>
> [root@compaq ~]# mdadm --detail /dev/md0
> /dev/md0:
> Version : 00.90.03
> Creation Time : Tue Dec 5 00:33:01 2006
> Raid Level : raid5
> Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
> Device Size : 243786240 (232.49 GiB 249.64 GB)
> Raid Devices : 6
> Total Devices : 6
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> Update Time : Wed Jan 17 23:14:39 2007
> State : clean
> Active Devices : 6
> Working Devices : 6
> Failed Devices : 0
> Spare Devices : 0
>
> Layout : left-symmetric
> Chunk Size : 256K
>
> UUID : 27dce477:6f45d11b:77377d08:732fa0e6
> Events : 0.58
>
> Number Major Minor RaidDevice State
> 0 8 1 0 active sync /dev/sda1
> 1 8 17 1 active sync /dev/sdb1
> 2 8 33 2 active sync /dev/sdc1
> 3 8 49 3 active sync /dev/sdd1
> 4 8 65 4 active sync /dev/sde1
> 5 8 81 5 active sync /dev/sdf1
> [root@compaq ~]#
>
>
> Sevrin
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
If they are on the PCI bus, that is about right, you probably should be
getting 10-15MB/s, but it is about right. If you had each drive on its
own PCI-e controller, then you would get much faster speeds.
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-17 22:38 ` Justin Piszcz
@ 2007-01-18 9:06 ` Raz Ben-Jehuda(caro)
2007-01-18 18:20 ` Sevrin Robstad
2007-01-18 18:19 ` Sevrin Robstad
1 sibling, 1 reply; 16+ messages in thread
From: Raz Ben-Jehuda(caro) @ 2007-01-18 9:06 UTC (permalink / raw)
To: Justin Piszcz; +Cc: Sevrin Robstad, linux-raid
did u increase the stripe cache size ?
On 1/18/07, Justin Piszcz <jpiszcz@lucidpixels.com> wrote:
> Sevrin Robstad wrote:
> > I'm suffering from bad performance on my RAID5.
> >
> > a "echo check >/sys/block/md0/md/sync_action"
> >
> > gives a speed at only about 5000K/sec , and HIGH load average :
> >
> > # uptime
> > 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
> >
> > kernel is 2.6.18.1.2257.fc5
> > mdadm is v2.5.5
> >
> > the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
> > PCI cards with a total of 6 250gb S-ATA drives connected.
> >
> > [root@compaq ~]# mdadm --detail /dev/md0
> > /dev/md0:
> > Version : 00.90.03
> > Creation Time : Tue Dec 5 00:33:01 2006
> > Raid Level : raid5
> > Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
> > Device Size : 243786240 (232.49 GiB 249.64 GB)
> > Raid Devices : 6
> > Total Devices : 6
> > Preferred Minor : 0
> > Persistence : Superblock is persistent
> >
> > Update Time : Wed Jan 17 23:14:39 2007
> > State : clean
> > Active Devices : 6
> > Working Devices : 6
> > Failed Devices : 0
> > Spare Devices : 0
> >
> > Layout : left-symmetric
> > Chunk Size : 256K
> >
> > UUID : 27dce477:6f45d11b:77377d08:732fa0e6
> > Events : 0.58
> >
> > Number Major Minor RaidDevice State
> > 0 8 1 0 active sync /dev/sda1
> > 1 8 17 1 active sync /dev/sdb1
> > 2 8 33 2 active sync /dev/sdc1
> > 3 8 49 3 active sync /dev/sdd1
> > 4 8 65 4 active sync /dev/sde1
> > 5 8 81 5 active sync /dev/sdf1
> > [root@compaq ~]#
> >
> >
> > Sevrin
> > -
> > To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> > the body of a message to majordomo@vger.kernel.org
> > More majordomo info at http://vger.kernel.org/majordomo-info.html
>
> If they are on the PCI bus, that is about right, you probably should be
> getting 10-15MB/s, but it is about right. If you had each drive on its
> own PCI-e controller, then you would get much faster speeds.
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
--
Raz
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-17 22:38 ` Justin Piszcz
2007-01-18 9:06 ` Raz Ben-Jehuda(caro)
@ 2007-01-18 18:19 ` Sevrin Robstad
2007-01-18 19:02 ` Mark Hahn
1 sibling, 1 reply; 16+ messages in thread
From: Sevrin Robstad @ 2007-01-18 18:19 UTC (permalink / raw)
To: Justin Piszcz; +Cc: linux-raid
Justin Piszcz wrote:
>> I'm suffering from bad performance on my RAID5.
>>
>> a "echo check >/sys/block/md0/md/sync_action"
>>
>> gives a speed at only about 5000K/sec , and HIGH load average :
>>
>> # uptime
>> 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
>>
>> kernel is 2.6.18.1.2257.fc5
>> mdadm is v2.5.5
>>
>> the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
>> PCI cards with a total of 6 250gb S-ATA drives connected.
>>
>> [root@compaq ~]# mdadm --detail /dev/md0
>> /dev/md0:
>> Version : 00.90.03
>> Creation Time : Tue Dec 5 00:33:01 2006
>> Raid Level : raid5
>> Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
>> Device Size : 243786240 (232.49 GiB 249.64 GB)
>> Raid Devices : 6
>> Total Devices : 6
>> Preferred Minor : 0
>> Persistence : Superblock is persistent
>>
>> Update Time : Wed Jan 17 23:14:39 2007
>> State : clean
>> Active Devices : 6
>> Working Devices : 6
>> Failed Devices : 0
>> Spare Devices : 0
>>
>> Layout : left-symmetric
>> Chunk Size : 256K
>>
>> UUID : 27dce477:6f45d11b:77377d08:732fa0e6
>> Events : 0.58
>>
>> Number Major Minor RaidDevice State
>> 0 8 1 0 active sync /dev/sda1
>> 1 8 17 1 active sync /dev/sdb1
>> 2 8 33 2 active sync /dev/sdc1
>> 3 8 49 3 active sync /dev/sdd1
>> 4 8 65 4 active sync /dev/sde1
>> 5 8 81 5 active sync /dev/sdf1
>> [root@compaq ~]#
>>
>>
>> Sevrin
>> -
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
> If they are on the PCI bus, that is about right, you probably should
> be getting 10-15MB/s, but it is about right. If you had each drive on
> its own PCI-e controller, then you would get much faster speeds.
>
>
There are two controllers, 4 disks connected to one controller on the
PCI-bus and 2 disks connected to the other controller.
As you say I should have had 10-15MB/s, but I'm having 5MB *and* with
really high load average.
Sevrin
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 9:06 ` Raz Ben-Jehuda(caro)
@ 2007-01-18 18:20 ` Sevrin Robstad
2007-01-18 19:56 ` Raz Ben-Jehuda(caro)
2007-01-18 23:01 ` Bill Davidsen
0 siblings, 2 replies; 16+ messages in thread
From: Sevrin Robstad @ 2007-01-18 18:20 UTC (permalink / raw)
To: Raz Ben-Jehuda(caro); +Cc: linux-raid
I've tried to increase the cache size - I can't measure any difference.....
Raz Ben-Jehuda(caro) wrote:
> did u increase the stripe cache size ?
>
>
> On 1/18/07, Justin Piszcz <jpiszcz@lucidpixels.com> wrote:
>> Sevrin Robstad wrote:
>> > I'm suffering from bad performance on my RAID5.
>> >
>> > a "echo check >/sys/block/md0/md/sync_action"
>> >
>> > gives a speed at only about 5000K/sec , and HIGH load average :
>> >
>> > # uptime
>> > 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
>> >
>> > kernel is 2.6.18.1.2257.fc5
>> > mdadm is v2.5.5
>> >
>> > the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
>> > PCI cards with a total of 6 250gb S-ATA drives connected.
>> >
>> > [root@compaq ~]# mdadm --detail /dev/md0
>> > /dev/md0:
>> > Version : 00.90.03
>> > Creation Time : Tue Dec 5 00:33:01 2006
>> > Raid Level : raid5
>> > Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
>> > Device Size : 243786240 (232.49 GiB 249.64 GB)
>> > Raid Devices : 6
>> > Total Devices : 6
>> > Preferred Minor : 0
>> > Persistence : Superblock is persistent
>> >
>> > Update Time : Wed Jan 17 23:14:39 2007
>> > State : clean
>> > Active Devices : 6
>> > Working Devices : 6
>> > Failed Devices : 0
>> > Spare Devices : 0
>> >
>> > Layout : left-symmetric
>> > Chunk Size : 256K
>> >
>> > UUID : 27dce477:6f45d11b:77377d08:732fa0e6
>> > Events : 0.58
>> >
>> > Number Major Minor RaidDevice State
>> > 0 8 1 0 active sync /dev/sda1
>> > 1 8 17 1 active sync /dev/sdb1
>> > 2 8 33 2 active sync /dev/sdc1
>> > 3 8 49 3 active sync /dev/sdd1
>> > 4 8 65 4 active sync /dev/sde1
>> > 5 8 81 5 active sync /dev/sdf1
>> > [root@compaq ~]#
>> >
>> >
>> > Sevrin
>> > -
>> > To unsubscribe from this list: send the line "unsubscribe
>> linux-raid" in
>> > the body of a message to majordomo@vger.kernel.org
>> > More majordomo info at http://vger.kernel.org/majordomo-info.html
>>
>> If they are on the PCI bus, that is about right, you probably should be
>> getting 10-15MB/s, but it is about right. If you had each drive on its
>> own PCI-e controller, then you would get much faster speeds.
>>
>> -
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>>
>
>
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-17 21:33 bad performance on RAID 5 Sevrin Robstad
2007-01-17 22:38 ` Justin Piszcz
@ 2007-01-18 18:29 ` Steve Cousins
2007-01-18 20:43 ` Sevrin Robstad
2007-01-18 23:10 ` Bill Davidsen
2007-01-19 1:33 ` dean gaudet
2 siblings, 2 replies; 16+ messages in thread
From: Steve Cousins @ 2007-01-18 18:29 UTC (permalink / raw)
To: Sevrin Robstad; +Cc: linux-raid
Sevrin Robstad wrote:
> I'm suffering from bad performance on my RAID5.
>
> a "echo check >/sys/block/md0/md/sync_action"
>
> gives a speed at only about 5000K/sec , and HIGH load average :
What do you get when you try something like:
time dd if=/dev/zero of=/mount-point/test.dat bs=1024k count=1024
where /mount-point is where /dev/md0 is mounted.
This will create a 1 GiB file and it will tell you how long it takes to
create it. Also, I'd try running Bonnie++ on it to see what the
different performance values are.
I don't know a lot about the md sync process but I remember having my
sync action stuck at a low value at one point and it didn't have
anything to do with the performance of the RAID array in general.
Steve
> # uptime
> 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
>
> kernel is 2.6.18.1.2257.fc5
> mdadm is v2.5.5
>
> the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA PCI
> cards with a total of 6 250gb S-ATA drives connected.
>
> [root@compaq ~]# mdadm --detail /dev/md0
> /dev/md0:
> Version : 00.90.03
> Creation Time : Tue Dec 5 00:33:01 2006
> Raid Level : raid5
> Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
> Device Size : 243786240 (232.49 GiB 249.64 GB)
> Raid Devices : 6
> Total Devices : 6
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> Update Time : Wed Jan 17 23:14:39 2007
> State : clean
> Active Devices : 6
> Working Devices : 6
> Failed Devices : 0
> Spare Devices : 0
>
> Layout : left-symmetric
> Chunk Size : 256K
>
> UUID : 27dce477:6f45d11b:77377d08:732fa0e6
> Events : 0.58
>
> Number Major Minor RaidDevice State
> 0 8 1 0 active sync /dev/sda1
> 1 8 17 1 active sync /dev/sdb1
> 2 8 33 2 active sync /dev/sdc1
> 3 8 49 3 active sync /dev/sdd1
> 4 8 65 4 active sync /dev/sde1
> 5 8 81 5 active sync /dev/sdf1
> [root@compaq ~]#
>
>
> Sevrin
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
--
______________________________________________________________________
Steve Cousins, Ocean Modeling Group Email: cousins@umit.maine.edu
Marine Sciences, 452 Aubert Hall http://rocky.umeoce.maine.edu
Univ. of Maine, Orono, ME 04469 Phone: (207) 581-4302
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 18:19 ` Sevrin Robstad
@ 2007-01-18 19:02 ` Mark Hahn
2007-01-18 20:47 ` Sevrin Robstad
0 siblings, 1 reply; 16+ messages in thread
From: Mark Hahn @ 2007-01-18 19:02 UTC (permalink / raw)
To: Sevrin Robstad; +Cc: Justin Piszcz, linux-raid
>>> gives a speed at only about 5000K/sec , and HIGH load average :
>>> # uptime
>>> 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
loadav is a bit misleading - it doesn't mean you had >11 runnable jobs.
you might just have more jobs waiting on IO, being starved by the
IO done by resync.
>>> Chunk Size : 256K
well, that's pretty big. it means 6*256K is necessary to do a
whole-stripe update; your stripe cache may be too small to be effective.
>> If they are on the PCI bus, that is about right, you probably should be
>> getting 10-15MB/s, but it is about right. If you had each drive on its own
>> PCI-e controller, then you would get much faster speeds.
10-15 seems bizarrely low - one can certainly achieve >100 MB/s
over the PCI bus, so where does the factor of 6-10 come in?
seems like a R6 resync would do 4 reads and 2 writes for every
4 chunks of throughput (so should achieve more like 50 MB/s if
the main limit is the bus at 100.)
> There are two controllers, 4 disks connected to one controller on the PCI-bus
> and 2 disks connected to the other controller.
well, you should probably look at the PCI topology ("lspci -v -t"),
and perhaps even the PCI settings (as well as stripe cache size,
perhaps nr_requests, etc)
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 18:20 ` Sevrin Robstad
@ 2007-01-18 19:56 ` Raz Ben-Jehuda(caro)
2007-01-18 23:01 ` Bill Davidsen
1 sibling, 0 replies; 16+ messages in thread
From: Raz Ben-Jehuda(caro) @ 2007-01-18 19:56 UTC (permalink / raw)
To: Sevrin Robstad; +Cc: linux-raid
iorder to understand what is going in your system you should:
1. determine the access pattern to the volume. meaning:
sequetial ? random access ?
sync io ? async io ?
mostly read ? mostly write ?
Are you using small buffers ? big buffers ?
2. you should test the controller capabilty.
meaning :
see if dd'in for each disk in the system seperately reduces the total
throughput.
On 1/18/07, Sevrin Robstad <quackyo@start.no> wrote:
> I've tried to increase the cache size - I can't measure any difference.....
>
> Raz Ben-Jehuda(caro) wrote:
> > did u increase the stripe cache size ?
> >
> >
> > On 1/18/07, Justin Piszcz <jpiszcz@lucidpixels.com> wrote:
> >> Sevrin Robstad wrote:
> >> > I'm suffering from bad performance on my RAID5.
> >> >
> >> > a "echo check >/sys/block/md0/md/sync_action"
> >> >
> >> > gives a speed at only about 5000K/sec , and HIGH load average :
> >> >
> >> > # uptime
> >> > 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
> >> >
> >> > kernel is 2.6.18.1.2257.fc5
> >> > mdadm is v2.5.5
> >> >
> >> > the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
> >> > PCI cards with a total of 6 250gb S-ATA drives connected.
> >> >
> >> > [root@compaq ~]# mdadm --detail /dev/md0
> >> > /dev/md0:
> >> > Version : 00.90.03
> >> > Creation Time : Tue Dec 5 00:33:01 2006
> >> > Raid Level : raid5
> >> > Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
> >> > Device Size : 243786240 (232.49 GiB 249.64 GB)
> >> > Raid Devices : 6
> >> > Total Devices : 6
> >> > Preferred Minor : 0
> >> > Persistence : Superblock is persistent
> >> >
> >> > Update Time : Wed Jan 17 23:14:39 2007
> >> > State : clean
> >> > Active Devices : 6
> >> > Working Devices : 6
> >> > Failed Devices : 0
> >> > Spare Devices : 0
> >> >
> >> > Layout : left-symmetric
> >> > Chunk Size : 256K
> >> >
> >> > UUID : 27dce477:6f45d11b:77377d08:732fa0e6
> >> > Events : 0.58
> >> >
> >> > Number Major Minor RaidDevice State
> >> > 0 8 1 0 active sync /dev/sda1
> >> > 1 8 17 1 active sync /dev/sdb1
> >> > 2 8 33 2 active sync /dev/sdc1
> >> > 3 8 49 3 active sync /dev/sdd1
> >> > 4 8 65 4 active sync /dev/sde1
> >> > 5 8 81 5 active sync /dev/sdf1
> >> > [root@compaq ~]#
> >> >
> >> >
> >> > Sevrin
> >> > -
> >> > To unsubscribe from this list: send the line "unsubscribe
> >> linux-raid" in
> >> > the body of a message to majordomo@vger.kernel.org
> >> > More majordomo info at http://vger.kernel.org/majordomo-info.html
> >>
> >> If they are on the PCI bus, that is about right, you probably should be
> >> getting 10-15MB/s, but it is about right. If you had each drive on its
> >> own PCI-e controller, then you would get much faster speeds.
> >>
> >> -
> >> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> >> the body of a message to majordomo@vger.kernel.org
> >> More majordomo info at http://vger.kernel.org/majordomo-info.html
> >>
> >
> >
>
>
--
Raz
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 18:29 ` Steve Cousins
@ 2007-01-18 20:43 ` Sevrin Robstad
2007-01-18 23:10 ` Bill Davidsen
1 sibling, 0 replies; 16+ messages in thread
From: Sevrin Robstad @ 2007-01-18 20:43 UTC (permalink / raw)
To: Steve Cousins; +Cc: linux-raid
Steve Cousins wrote:
> Sevrin Robstad wrote:
>> I'm suffering from bad performance on my RAID5.
>>
>> a "echo check >/sys/block/md0/md/sync_action"
>>
>> gives a speed at only about 5000K/sec , and HIGH load average :
> What do you get when you try something like:
>
> time dd if=/dev/zero of=/mount-point/test.dat bs=1024k count=1024
>
> where /mount-point is where /dev/md0 is mounted.
>
> This will create a 1 GiB file and it will tell you how long it takes
> to create it. Also, I'd try running Bonnie++ on it to see what the
> different performance values are.
>
> I don't know a lot about the md sync process but I remember having my
> sync action stuck at a low value at one point and it didn't have
> anything to do with the performance of the RAID array in general.
>
I ran the a couple of times, and got either about 28MB/s or about
34MB/s. strange.
When I run the same test on a single disk connected to another
controller I get about 60MB/s.
[root@compaq ~]# time dd if=/dev/zero of=/mnt/gigaraid/1gb.testfile
bs=1024k cou nt=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB) copied, 38.1873 seconds, 28.1 MB/s
real 0m38.321s
user 0m0.009s
sys 0m8.602s
[root@compaq ~]# time dd if=/dev/zero of=/mnt/gigaraid/1gb.testfile
bs=1024k cou nt=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB) copied, 31.4377 seconds, 34.2 MB/s
real 0m31.949s
user 0m0.016s
sys 0m8.988s
[root@compaq ~]# time dd if=/dev/zero of=/mnt/gigaraid/1gb.testfile
bs=1024k cou nt=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB) copied, 37.599 seconds, 28.6 MB/s
real 0m38.151s
user 0m0.011s
sys 0m9.291s
[root@compaq ~]# time dd if=/dev/zero of=/mnt/gigaraid/1gb.testfile
bs=1024k count=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB) copied, 31.2569 seconds, 34.4 MB/s
real 0m31.765s
user 0m0.007s
sys 0m8.913s
[root@compaq ~]# time dd if=/dev/zero of=/mnt/gigaraid/1gb.testfile
bs=1024k count=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB) copied, 37.4778 seconds, 28.7 MB/s
real 0m37.923s
user 0m0.009s
sys 0m9.231s
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 19:02 ` Mark Hahn
@ 2007-01-18 20:47 ` Sevrin Robstad
0 siblings, 0 replies; 16+ messages in thread
From: Sevrin Robstad @ 2007-01-18 20:47 UTC (permalink / raw)
To: Mark Hahn; +Cc: linux-raid
Mark Hahn wrote:
>>> Chunk Size : 256K
>
> well, that's pretty big. it means 6*256K is necessary to do a
> whole-stripe update; your stripe cache may be too small to be effective.
>
>>> If they are on the PCI bus, that is about right, you probably should
>>> be getting 10-15MB/s, but it is about right. If you had each drive
>>> on its own PCI-e controller, then you would get much faster speeds.
>
> 10-15 seems bizarrely low - one can certainly achieve >100 MB/s
> over the PCI bus, so where does the factor of 6-10 come in?
> seems like a R6 resync would do 4 reads and 2 writes for every
> 4 chunks of throughput (so should achieve more like 50 MB/s if the
> main limit is the bus at 100.)
>
>> There are two controllers, 4 disks connected to one controller on the
>> PCI-bus and 2 disks connected to the other controller.
>
> well, you should probably look at the PCI topology ("lspci -v -t"),
> and perhaps even the PCI settings (as well as stripe cache size,
> perhaps nr_requests, etc)
>
>
I've tried to increase the stripe_cache_size a lot, no noticable difference.
I set it to 8096....
[root@compaq md]# lspci -v -t
-[0000:00]-+-00.0 VIA Technologies, Inc. VT8366/A/7 [Apollo KT266/A/333]
+-01.0-[0000:01]----00.0 ATI Technologies Inc Rage 128 RF/SG AGP
+-05.0 Silicon Image, Inc. SiI 3114 [SATALink/SATARaid]
Serial ATA Controller
+-07.0 Silicon Image, Inc. SiI 3114 [SATALink/SATARaid]
Serial ATA Controller
+-08.0 Realtek Semiconductor Co., Ltd. RTL-8169 Gigabit Ethernet
+-11.0 VIA Technologies, Inc. VT8233 PCI to ISA Bridge
\-11.1 VIA Technologies, Inc.
VT82C586A/B/VT82C686/A/B/VT823x/A/C PIPC Bus Master IDE
[root@compaq md]#
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 18:20 ` Sevrin Robstad
2007-01-18 19:56 ` Raz Ben-Jehuda(caro)
@ 2007-01-18 23:01 ` Bill Davidsen
1 sibling, 0 replies; 16+ messages in thread
From: Bill Davidsen @ 2007-01-18 23:01 UTC (permalink / raw)
To: Sevrin Robstad; +Cc: Raz Ben-Jehuda(caro), linux-raid
Sevrin Robstad wrote:
> I've tried to increase the cache size - I can't measure any
> difference.....
You probably won't help small writes, but large writes will go faster
with a stripe cache of size num_disks*chunk_size*2 or larger.
>
> Raz Ben-Jehuda(caro) wrote:
>> did u increase the stripe cache size ?
>>
>>
>> On 1/18/07, Justin Piszcz <jpiszcz@lucidpixels.com> wrote:
>>> Sevrin Robstad wrote:
>>> > I'm suffering from bad performance on my RAID5.
>>> >
>>> > a "echo check >/sys/block/md0/md/sync_action"
>>> >
>>> > gives a speed at only about 5000K/sec , and HIGH load average :
>>> >
>>> > # uptime
>>> > 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
>>> >
>>> > kernel is 2.6.18.1.2257.fc5
>>> > mdadm is v2.5.5
>>> >
>>> > the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
>>> > PCI cards with a total of 6 250gb S-ATA drives connected.
>>> >
>>> > [root@compaq ~]# mdadm --detail /dev/md0
>>> > /dev/md0:
>>> > Version : 00.90.03
>>> > Creation Time : Tue Dec 5 00:33:01 2006
>>> > Raid Level : raid5
>>> > Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
>>> > Device Size : 243786240 (232.49 GiB 249.64 GB)
>>> > Raid Devices : 6
>>> > Total Devices : 6
>>> > Preferred Minor : 0
>>> > Persistence : Superblock is persistent
>>> >
>>> > Update Time : Wed Jan 17 23:14:39 2007
>>> > State : clean
>>> > Active Devices : 6
>>> > Working Devices : 6
>>> > Failed Devices : 0
>>> > Spare Devices : 0
>>> >
>>> > Layout : left-symmetric
>>> > Chunk Size : 256K
>>> >
>>> > UUID : 27dce477:6f45d11b:77377d08:732fa0e6
>>> > Events : 0.58
>>> >
>>> > Number Major Minor RaidDevice State
>>> > 0 8 1 0 active sync /dev/sda1
>>> > 1 8 17 1 active sync /dev/sdb1
>>> > 2 8 33 2 active sync /dev/sdc1
>>> > 3 8 49 3 active sync /dev/sdd1
>>> > 4 8 65 4 active sync /dev/sde1
>>> > 5 8 81 5 active sync /dev/sdf1
>>> > [root@compaq ~]#
>>> >
>>> >
>>> > Sevrin
>>> > -
>>> > To unsubscribe from this list: send the line "unsubscribe
>>> linux-raid" in
>>> > the body of a message to majordomo@vger.kernel.org
>>> > More majordomo info at http://vger.kernel.org/majordomo-info.html
>>>
>>> If they are on the PCI bus, that is about right, you probably should be
>>> getting 10-15MB/s, but it is about right. If you had each drive on its
>>> own PCI-e controller, then you would get much faster speeds.
--
bill davidsen <davidsen@tmr.com>
CTO TMR Associates, Inc
Doing interesting things with small computers since 1979
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 18:29 ` Steve Cousins
2007-01-18 20:43 ` Sevrin Robstad
@ 2007-01-18 23:10 ` Bill Davidsen
2007-01-21 17:08 ` Nix
1 sibling, 1 reply; 16+ messages in thread
From: Bill Davidsen @ 2007-01-18 23:10 UTC (permalink / raw)
To: Steve Cousins; +Cc: Sevrin Robstad, linux-raid
) Steve Cousins wrote:
> Sevrin Robstad wrote:
>> I'm suffering from bad performance on my RAID5.
>>
>> a "echo check >/sys/block/md0/md/sync_action"
>>
>> gives a speed at only about 5000K/sec , and HIGH load average :
>
> What do you get when you try something like:
>
> time dd if=/dev/zero of=/mount-point/test.dat bs=1024k count=1024
That doesn't give valid (repeatable) results due to caching issues. Go
back to the thread I started on RAID-5 write, and see my results. More
important, the way I got rid of the cache effects (beside an unloaded
systems) was:
sync; time bash -c "dd if=/dev/zero bs=1024k count=2048
of=/mnt/point/file; sync"
I empty the cache, then time the dd including the sync at the end.
Results are far more repeatable.
I actually was able to create custom arrays on unused devices to play
with array setting, once you use the array you can't tune as much.
>
> where /mount-point is where /dev/md0 is mounted.
>
> This will create a 1 GiB file and it will tell you how long it takes
> to create it. Also, I'd try running Bonnie++ on it to see what the
> different performance values are.
>
> I don't know a lot about the md sync process but I remember having my
> sync action stuck at a low value at one point and it didn't have
> anything to do with the performance of the RAID array in general.
>
> Steve
>
>> # uptime
>> 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
>>
>> kernel is 2.6.18.1.2257.fc5
>> mdadm is v2.5.5
>>
>> the system consist of an athlon XP1,2GHz and two Sil3114 4port S-ATA
>> PCI cards with a total of 6 250gb S-ATA drives connected.
>>
>> [root@compaq ~]# mdadm --detail /dev/md0
>> /dev/md0:
>> Version : 00.90.03
>> Creation Time : Tue Dec 5 00:33:01 2006
>> Raid Level : raid5
>> Array Size : 1218931200 (1162.46 GiB 1248.19 GB)
>> Device Size : 243786240 (232.49 GiB 249.64 GB)
>> Raid Devices : 6
>> Total Devices : 6
>> Preferred Minor : 0
>> Persistence : Superblock is persistent
>>
>> Update Time : Wed Jan 17 23:14:39 2007
>> State : clean
>> Active Devices : 6
>> Working Devices : 6
>> Failed Devices : 0
>> Spare Devices : 0
>>
>> Layout : left-symmetric
>> Chunk Size : 256K
>>
>> UUID : 27dce477:6f45d11b:77377d08:732fa0e6
>> Events : 0.58
>>
>> Number Major Minor RaidDevice State
>> 0 8 1 0 active sync /dev/sda1
>> 1 8 17 1 active sync /dev/sdb1
>> 2 8 33 2 active sync /dev/sdc1
>> 3 8 49 3 active sync /dev/sdd1
>> 4 8 65 4 active sync /dev/sde1
>> 5 8 81 5 active sync /dev/sdf1
>> [root@compaq ~]#
>>
>>
>> Sevrin
--
bill davidsen <davidsen@tmr.com>
CTO TMR Associates, Inc
Doing interesting things with small computers since 1979
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-17 21:33 bad performance on RAID 5 Sevrin Robstad
2007-01-17 22:38 ` Justin Piszcz
2007-01-18 18:29 ` Steve Cousins
@ 2007-01-19 1:33 ` dean gaudet
2007-01-20 10:18 ` Roger Lucas
2 siblings, 1 reply; 16+ messages in thread
From: dean gaudet @ 2007-01-19 1:33 UTC (permalink / raw)
To: Sevrin Robstad; +Cc: linux-raid
On Wed, 17 Jan 2007, Sevrin Robstad wrote:
> I'm suffering from bad performance on my RAID5.
>
> a "echo check >/sys/block/md0/md/sync_action"
>
> gives a speed at only about 5000K/sec , and HIGH load average :
>
> # uptime
> 20:03:55 up 8 days, 19:55, 1 user, load average: 11.70, 4.04, 1.52
iostat -kx /dev/sd? 10 ... and sum up the total IO...
also try increasing sync_speed_min/max....
and a loadavg jump like that suggests to me you have other things
competing for the disk at the same time as the "check".
-dean
^ permalink raw reply [flat|nested] 16+ messages in thread
* RE: bad performance on RAID 5
2007-01-19 1:33 ` dean gaudet
@ 2007-01-20 10:18 ` Roger Lucas
0 siblings, 0 replies; 16+ messages in thread
From: Roger Lucas @ 2007-01-20 10:18 UTC (permalink / raw)
To: 'Sevrin Robstad'; +Cc: linux-raid
Hi Sevrin,
Are you sure all the disks are working OK?
We saw a problem here with a 4-disk SATA array. The array was working
without errors but was _very_ slow. When we checked each disk individually,
we found one disk was reporting SMART errors and running slow. We removed
it and replaced it with a new disk and the RAID array ran at full speed
again. Further tests on the removed disk found that it was having a lot of
problems but just about hanging in there (although taking a long time for
each operation) - hence the RAID array didn't mark it as faulty.
I would check the SMART logs with smartctl to see if anything looks a bit
wrong and try benchmarking the disks individually too.
- Roger
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-18 23:10 ` Bill Davidsen
@ 2007-01-21 17:08 ` Nix
2007-01-21 19:09 ` Bill Davidsen
0 siblings, 1 reply; 16+ messages in thread
From: Nix @ 2007-01-21 17:08 UTC (permalink / raw)
To: Bill Davidsen; +Cc: Steve Cousins, Sevrin Robstad, linux-raid
On 18 Jan 2007, Bill Davidsen spake thusly:
> ) Steve Cousins wrote:
>> time dd if=/dev/zero of=/mount-point/test.dat bs=1024k count=1024
> That doesn't give valid (repeatable) results due to caching issues. Go
> back to the thread I started on RAID-5 write, and see my results. More
> important, the way I got rid of the cache effects (beside an unloaded
> systems) was:
> sync; time bash -c "dd if=/dev/zero bs=1024k count=2048 of=/mnt/point/file; sync"
> I empty the cache, then time the dd including the sync at the
> end. Results are far more repeatable.
Recent versions of dd have `oflag=direct' as well, to open the output
with O_DIRECT. (I'm not sure what the state of O_DIRECT on regular files
is though.)
--
`The serial comma, however, is correct and proper, and abandoning it will
surely lead to chaos, anarchy, rioting in the streets, the Terrorists
taking over, and possibly the complete collapse of Human Civilization.'
^ permalink raw reply [flat|nested] 16+ messages in thread
* Re: bad performance on RAID 5
2007-01-21 17:08 ` Nix
@ 2007-01-21 19:09 ` Bill Davidsen
0 siblings, 0 replies; 16+ messages in thread
From: Bill Davidsen @ 2007-01-21 19:09 UTC (permalink / raw)
To: Nix; +Cc: Steve Cousins, Sevrin Robstad, linux-raid
Nix wrote:
> On 18 Jan 2007, Bill Davidsen spake thusly:
>
>> ) Steve Cousins wrote:
>>
>>> time dd if=/dev/zero of=/mount-point/test.dat bs=1024k count=1024
>>>
>> That doesn't give valid (repeatable) results due to caching issues. Go
>> back to the thread I started on RAID-5 write, and see my results. More
>> important, the way I got rid of the cache effects (beside an unloaded
>> systems) was:
>> sync; time bash -c "dd if=/dev/zero bs=1024k count=2048 of=/mnt/point/file; sync"
>> I empty the cache, then time the dd including the sync at the
>> end. Results are far more repeatable.
>>
>
> Recent versions of dd have `oflag=direct' as well, to open the output
> with O_DIRECT. (I'm not sure what the state of O_DIRECT on regular files
> is though.)
>
>
Doing the write using page cache and then just a single sync at the end
gives a closer estimate of what can be written to the array in general.
By going to O_DIRECT every i/o take place unbuffered, which is more
typical of database use or similar. My original problem was capturing
real data at 75MB/s and not being able to write it to an array that
fast, taking advantage of page cache. By going to huge stripe cache,
about 20x larger than default, performance was boosted to a more useful
level.
I think both measurements are useful, but they aren't measuring the same
thing.
--
bill davidsen <davidsen@tmr.com>
CTO TMR Associates, Inc
Doing interesting things with small computers since 1979
^ permalink raw reply [flat|nested] 16+ messages in thread
end of thread, other threads:[~2007-01-21 19:09 UTC | newest]
Thread overview: 16+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2007-01-17 21:33 bad performance on RAID 5 Sevrin Robstad
2007-01-17 22:38 ` Justin Piszcz
2007-01-18 9:06 ` Raz Ben-Jehuda(caro)
2007-01-18 18:20 ` Sevrin Robstad
2007-01-18 19:56 ` Raz Ben-Jehuda(caro)
2007-01-18 23:01 ` Bill Davidsen
2007-01-18 18:19 ` Sevrin Robstad
2007-01-18 19:02 ` Mark Hahn
2007-01-18 20:47 ` Sevrin Robstad
2007-01-18 18:29 ` Steve Cousins
2007-01-18 20:43 ` Sevrin Robstad
2007-01-18 23:10 ` Bill Davidsen
2007-01-21 17:08 ` Nix
2007-01-21 19:09 ` Bill Davidsen
2007-01-19 1:33 ` dean gaudet
2007-01-20 10:18 ` Roger Lucas
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).