* sw raid0 read bottleneck
@ 2007-03-13 16:23 Tomka Gergely
2007-03-13 16:48 ` Justin Piszcz
0 siblings, 1 reply; 11+ messages in thread
From: Tomka Gergely @ 2007-03-13 16:23 UTC (permalink / raw)
To: linux-raid
Hi!
I am currently testing 3ware raid cards. Now i have 15 disks, and on these
a swraid0. The write speed seems good (700 MBps), but the read performance
only 350 MBps. Another problem when i try to read with two process, then
the _sum_ of the read speeds fall back to 200 MBps. So there is a
bottleneck, or something i need to know, but i dont have ideas.
The details:
/dev/md0:
Version : 00.90.03
Creation Time : Tue Mar 13 16:57:32 2007
Raid Level : raid0
Array Size : 7325797440 (6986.43 GiB 7501.62 GB)
Raid Devices : 15
Total Devices : 15
Preferred Minor : 0
Persistence : Superblock is persistent
Update Time : Tue Mar 13 16:57:32 2007
State : clean
Active Devices : 15
Working Devices : 15
Failed Devices : 0
Spare Devices : 0
Chunk Size : 64K
# uname -a
Linux ursula 2.6.18-4-686-bigmem #1 SMP Wed Feb 21 17:30:22 UTC 2007 i686
GNU/Linux
# xfs_info /mnt/
meta-data=/dev/md0 isize=256 agcount=32, agsize=57232784
blks
= sectsz=512 attr=0
data = bsize=4096 blocks=1831449088,
imaxpct=25
= sunit=16 swidth=240 blks, unwritten=1
naming =version 2 bsize=4096
log =internal bsize=4096 blocks=32768, version=1
= sectsz=512 sunit=0 blks
realtime =none extsz=983040 blocks=0, rtextents=0
(all software Debian Etch)
four Intel(R) Xeon(TM) CPU 3.00GHz
two 3ware 9590SE-8ML on PCIe
Intel Corporation 5000P Chipset
--
Tomka Gergely, gergely@tomka.hu
^ permalink raw reply [flat|nested] 11+ messages in thread* Re: sw raid0 read bottleneck
2007-03-13 16:23 sw raid0 read bottleneck Tomka Gergely
@ 2007-03-13 16:48 ` Justin Piszcz
2007-03-13 17:34 ` Tomka Gergely
0 siblings, 1 reply; 11+ messages in thread
From: Justin Piszcz @ 2007-03-13 16:48 UTC (permalink / raw)
To: Tomka Gergely; +Cc: linux-raid
On Tue, 13 Mar 2007, Tomka Gergely wrote:
> Hi!
>
> I am currently testing 3ware raid cards. Now i have 15 disks, and on these
> a swraid0. The write speed seems good (700 MBps), but the read performance
> only 350 MBps. Another problem when i try to read with two process, then
> the _sum_ of the read speeds fall back to 200 MBps. So there is a
> bottleneck, or something i need to know, but i dont have ideas.
>
> The details:
>
> /dev/md0:
> Version : 00.90.03
> Creation Time : Tue Mar 13 16:57:32 2007
> Raid Level : raid0
> Array Size : 7325797440 (6986.43 GiB 7501.62 GB)
> Raid Devices : 15
> Total Devices : 15
> Preferred Minor : 0
> Persistence : Superblock is persistent
>
> Update Time : Tue Mar 13 16:57:32 2007
> State : clean
> Active Devices : 15
> Working Devices : 15
> Failed Devices : 0
> Spare Devices : 0
>
> Chunk Size : 64K
>
> # uname -a
> Linux ursula 2.6.18-4-686-bigmem #1 SMP Wed Feb 21 17:30:22 UTC 2007 i686
> GNU/Linux
>
> # xfs_info /mnt/
> meta-data=/dev/md0 isize=256 agcount=32, agsize=57232784
> blks
> = sectsz=512 attr=0
> data = bsize=4096 blocks=1831449088,
> imaxpct=25
> = sunit=16 swidth=240 blks, unwritten=1
> naming =version 2 bsize=4096
> log =internal bsize=4096 blocks=32768, version=1
> = sectsz=512 sunit=0 blks
> realtime =none extsz=983040 blocks=0, rtextents=0
>
> (all software Debian Etch)
>
> four Intel(R) Xeon(TM) CPU 3.00GHz
>
> two 3ware 9590SE-8ML on PCIe
>
> Intel Corporation 5000P Chipset
>
>
> --
> Tomka Gergely, gergely@tomka.hu
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
Have you tried increasing your readahead values for the md device?
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 16:48 ` Justin Piszcz
@ 2007-03-13 17:34 ` Tomka Gergely
2007-03-13 18:15 ` Tomka Gergely
0 siblings, 1 reply; 11+ messages in thread
From: Tomka Gergely @ 2007-03-13 17:34 UTC (permalink / raw)
Cc: linux-raid
On Tue, 13 Mar 2007, Justin Piszcz wrote:
> Have you tried increasing your readahead values for the md device?
Yes. No real change. According to my humble mental image, readahead not a
too useful thing, when we read 1-4 thread with sdd. The io subsystem
already reading with the possible maximum speed, so don't have time to
read ahead. Correct me, if i wrong.
--
Tomka Gergely, gergely@tomka.hu
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 17:34 ` Tomka Gergely
@ 2007-03-13 18:15 ` Tomka Gergely
2007-03-13 18:42 ` Justin Piszcz
2007-03-13 20:51 ` Neil Brown
0 siblings, 2 replies; 11+ messages in thread
From: Tomka Gergely @ 2007-03-13 18:15 UTC (permalink / raw)
To: Tomka Gergely; +Cc: linux-raid
On Tue, 13 Mar 2007, Tomka Gergely wrote:
> On Tue, 13 Mar 2007, Justin Piszcz wrote:
>
> > Have you tried increasing your readahead values for the md device?
>
> Yes. No real change. According to my humble mental image, readahead not a
> too useful thing, when we read 1-4 thread with sdd. The io subsystem
> already reading with the possible maximum speed, so don't have time to
> read ahead. Correct me, if i wrong.
I was wrong, readahead can speed things up, to 450 MBps.
--
Tomka Gergely, gergely@tomka.hu
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 18:15 ` Tomka Gergely
@ 2007-03-13 18:42 ` Justin Piszcz
2007-03-13 20:51 ` Neil Brown
1 sibling, 0 replies; 11+ messages in thread
From: Justin Piszcz @ 2007-03-13 18:42 UTC (permalink / raw)
To: Tomka Gergely; +Cc: linux-raid
Nice.
On Tue, 13 Mar 2007, Tomka Gergely wrote:
> On Tue, 13 Mar 2007, Tomka Gergely wrote:
>
>> On Tue, 13 Mar 2007, Justin Piszcz wrote:
>>
>>> Have you tried increasing your readahead values for the md device?
>>
>> Yes. No real change. According to my humble mental image, readahead not a
>> too useful thing, when we read 1-4 thread with sdd. The io subsystem
>> already reading with the possible maximum speed, so don't have time to
>> read ahead. Correct me, if i wrong.
>
> I was wrong, readahead can speed things up, to 450 MBps.
>
> --
> Tomka Gergely, gergely@tomka.hu
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 18:15 ` Tomka Gergely
2007-03-13 18:42 ` Justin Piszcz
@ 2007-03-13 20:51 ` Neil Brown
2007-03-13 21:09 ` Tomka Gergely
2007-03-13 21:29 ` Mark Hahn
1 sibling, 2 replies; 11+ messages in thread
From: Neil Brown @ 2007-03-13 20:51 UTC (permalink / raw)
To: Tomka Gergely; +Cc: linux-raid
On Tuesday March 13, gergely@tomka.hu wrote:
> On Tue, 13 Mar 2007, Tomka Gergely wrote:
>
> > On Tue, 13 Mar 2007, Justin Piszcz wrote:
> >
> > > Have you tried increasing your readahead values for the md device?
> >
> > Yes. No real change. According to my humble mental image, readahead not a
> > too useful thing, when we read 1-4 thread with sdd. The io subsystem
> > already reading with the possible maximum speed, so don't have time to
> > read ahead. Correct me, if i wrong.
>
> I was wrong, readahead can speed things up, to 450 MBps.
Can you tell use what read-ahead size you needed?
15 drives and 64K chunks gives 960K per stripe.
The raid0 code should set the read-ahead to twice that: 1920K
which I would have thought would be enough, but apparently not.
Thanks,
NeilBrown
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 20:51 ` Neil Brown
@ 2007-03-13 21:09 ` Tomka Gergely
2007-03-13 21:29 ` Mark Hahn
1 sibling, 0 replies; 11+ messages in thread
From: Tomka Gergely @ 2007-03-13 21:09 UTC (permalink / raw)
To: Neil Brown; +Cc: linux-raid
On Wed, 14 Mar 2007, Neil Brown wrote:
> On Tuesday March 13, gergely@tomka.hu wrote:
> > On Tue, 13 Mar 2007, Tomka Gergely wrote:
> >
> > > On Tue, 13 Mar 2007, Justin Piszcz wrote:
> > >
> > > > Have you tried increasing your readahead values for the md device?
> > >
> > > Yes. No real change. According to my humble mental image, readahead not a
> > > too useful thing, when we read 1-4 thread with sdd. The io subsystem
> > > already reading with the possible maximum speed, so don't have time to
> > > read ahead. Correct me, if i wrong.
> >
> > I was wrong, readahead can speed things up, to 450 MBps.
>
> Can you tell use what read-ahead size you needed?
>
> 15 drives and 64K chunks gives 960K per stripe.
> The raid0 code should set the read-ahead to twice that: 1920K
> which I would have thought would be enough, but apparently not.
blockdev --setra 262144 /dev/md0 gives me 650+ MB/s with 4 threads
(paralell running sdd). Lower values give lower speeds, greater values not
giving higher speeds.
--
Tomka Gergely, gergely@tomka.hu
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 20:51 ` Neil Brown
2007-03-13 21:09 ` Tomka Gergely
@ 2007-03-13 21:29 ` Mark Hahn
2007-03-14 18:41 ` Bill Davidsen
1 sibling, 1 reply; 11+ messages in thread
From: Mark Hahn @ 2007-03-13 21:29 UTC (permalink / raw)
To: Neil Brown; +Cc: Tomka Gergely, linux-raid
> 15 drives and 64K chunks gives 960K per stripe.
> The raid0 code should set the read-ahead to twice that: 1920K
> which I would have thought would be enough, but apparently not.
choosing the RA size should depend in some way on speed, shouldn't it?
after all, the goal is to have enough reads queued to avoid a "stall"
while un-read-ahead-ed sectors pass under the head. for a typical
60 MB/s, 7200 rpm disk, a track is .5 MB. the heuristic above only reads
ahead .12 MB...
regards, mark hahn.
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-13 21:29 ` Mark Hahn
@ 2007-03-14 18:41 ` Bill Davidsen
2007-03-14 19:32 ` Mark Hahn
0 siblings, 1 reply; 11+ messages in thread
From: Bill Davidsen @ 2007-03-14 18:41 UTC (permalink / raw)
To: Mark Hahn; +Cc: Neil Brown, Tomka Gergely, linux-raid
Mark Hahn wrote:
>> 15 drives and 64K chunks gives 960K per stripe.
>> The raid0 code should set the read-ahead to twice that: 1920K
>> which I would have thought would be enough, but apparently not.
>
> choosing the RA size should depend in some way on speed, shouldn't it?
> after all, the goal is to have enough reads queued to avoid a "stall"
> while un-read-ahead-ed sectors pass under the head. for a typical
> 60 MB/s, 7200 rpm disk, a track is .5 MB. the heuristic above only reads
> ahead .12 MB...
While I'm sure you realize it, some reader won't, so I'll point out that
a cylinder is all the tracks which can be read without a seek, and
cylinder size is track size times number of data heads. If you want max
speed you can read that much off a drive at a time...
--
bill davidsen <davidsen@tmr.com>
CTO TMR Associates, Inc
Doing interesting things with small computers since 1979
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-14 18:41 ` Bill Davidsen
@ 2007-03-14 19:32 ` Mark Hahn
2007-03-14 21:53 ` Bill Davidsen
0 siblings, 1 reply; 11+ messages in thread
From: Mark Hahn @ 2007-03-14 19:32 UTC (permalink / raw)
To: Bill Davidsen; +Cc: Neil Brown, Tomka Gergely, linux-raid
>>> 15 drives and 64K chunks gives 960K per stripe.
>>> The raid0 code should set the read-ahead to twice that: 1920K
>>> which I would have thought would be enough, but apparently not.
>>
>> choosing the RA size should depend in some way on speed, shouldn't it?
>> after all, the goal is to have enough reads queued to avoid a "stall"
>> while un-read-ahead-ed sectors pass under the head. for a typical
>> 60 MB/s, 7200 rpm disk, a track is .5 MB. the heuristic above only reads
>> ahead .12 MB...
>
> While I'm sure you realize it, some reader won't, so I'll point out that a
> cylinder is all the tracks which can be read without a seek, and cylinder
> size is track size times number of data heads. If you want max speed you can
> read that much off a drive at a time...
hmm. I certainly agree with your terminology, but afaikt, disks only read
from a head at a time. the last time I saw a vendor quoting head-switch
times, they were within a factor of two one-track seek times, which explains
why a same-family disk with more heads/surfaces shows only slight speedup.
regards, mark hahn.
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: sw raid0 read bottleneck
2007-03-14 19:32 ` Mark Hahn
@ 2007-03-14 21:53 ` Bill Davidsen
0 siblings, 0 replies; 11+ messages in thread
From: Bill Davidsen @ 2007-03-14 21:53 UTC (permalink / raw)
To: Mark Hahn; +Cc: Neil Brown, Tomka Gergely, linux-raid
Mark Hahn wrote:
>>>> 15 drives and 64K chunks gives 960K per stripe.
>>>> The raid0 code should set the read-ahead to twice that: 1920K
>>>> which I would have thought would be enough, but apparently not.
>>>
>>> choosing the RA size should depend in some way on speed, shouldn't it?
>>> after all, the goal is to have enough reads queued to avoid a "stall"
>>> while un-read-ahead-ed sectors pass under the head. for a typical
>>> 60 MB/s, 7200 rpm disk, a track is .5 MB. the heuristic above only
>>> reads
>>> ahead .12 MB...
>>
>> While I'm sure you realize it, some reader won't, so I'll point out
>> that a cylinder is all the tracks which can be read without a seek,
>> and cylinder size is track size times number of data heads. If you
>> want max speed you can read that much off a drive at a time...
>
> hmm. I certainly agree with your terminology, but afaikt, disks only
> read
> from a head at a time. the last time I saw a vendor quoting head-switch
> times, they were within a factor of two one-track seek times, which
> explains
> why a same-family disk with more heads/surfaces shows only slight
> speedup.
Agreed. The only real gain is avoiding intervening seeks. Makes a case
for using cylinder size as chunk size? Maybe. Makes a case for being
sure your readahead is large enough, though, for some sensible value of
"enough."
--
bill davidsen <davidsen@tmr.com>
CTO TMR Associates, Inc
Doing interesting things with small computers since 1979
^ permalink raw reply [flat|nested] 11+ messages in thread
end of thread, other threads:[~2007-03-14 21:53 UTC | newest]
Thread overview: 11+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2007-03-13 16:23 sw raid0 read bottleneck Tomka Gergely
2007-03-13 16:48 ` Justin Piszcz
2007-03-13 17:34 ` Tomka Gergely
2007-03-13 18:15 ` Tomka Gergely
2007-03-13 18:42 ` Justin Piszcz
2007-03-13 20:51 ` Neil Brown
2007-03-13 21:09 ` Tomka Gergely
2007-03-13 21:29 ` Mark Hahn
2007-03-14 18:41 ` Bill Davidsen
2007-03-14 19:32 ` Mark Hahn
2007-03-14 21:53 ` Bill Davidsen
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).