* MD write performance issue
@ 2009-10-16 8:45 mark delfman
2009-10-16 10:42 ` Asdo
0 siblings, 1 reply; 3+ messages in thread
From: mark delfman @ 2009-10-16 8:45 UTC (permalink / raw)
To: Linux RAID Mailing List
After further work we are sure that there is a significant write
performance issue with either the Kernel+MD or with
Me_Being_Stupid+MD. I would be happy to find out it is the latter,
but given the possibility it may be Kernel+MD I thought I should raise
this again under a new thread with below info.
On previous threads I stated that on later kernels XFS > MD dropped up
to 50% performance... I now have tested several filesystems (it is not
XFS related) and several Kernel’s (it is related to Kernel)
Below is the basic summary of the various Kernel write performance
results... all on the same hardware and Open SUSE foundation.
1.1GBsec is the max this particular hardware will provide and in a
single stream XFS would normally match this. You can see that
depending on Kernel the write performance can drop to 50%.....
(note: we have tried this with a hardware raid and there is no
performance drop between direct or via a FS.)
I really am stuck as to where to go from here, I would like to use .30
kernels and I am sure that there is a solution to this but do not have
enough linux experience to think of any further steps, so very much
appreciate any advice.
Linux linux-tlfp 2.6.27.7-vanilla #1 SMP Thu Oct 15 19:52:43 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1 GB/s
XFS: 1.1 GB/s
(tests with simple DD’s, but I have also tried FIO etc)
2.6.30.8
RAW: 1.1 GB/s
XFS: 900 MB/s
Linux linux-tlfp 2.6.28.4-vanilla #1 SMP Thu Oct 15 21:36:51 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1 GB/s
XFS: 1.1 GB/s
Linux linux-tlfp 2.6.31.4-vanilla #1 SMP Thu Oct 15 22:11:16 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW 1,1 GB/s
XFS: 920 MB/s
Linux linux-tlfp 2.6.28.10-vanilla #1 SMP Thu Oct 15 22:44:26 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1
XFS: 684 MB/s
Linux linux-tlfp 2.6.27.37-vanilla #1 SMP Thu Oct 15 23:32:53 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1 GB/s
XFS 520 MB/s
Linux linux-tlfp 2.6.27.20-vanilla #1 SMP Thu Oct 15 23:59:32 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAw 1.1 GB/s
XFS: 487 MB/s
Linux linux-tlfp 2.6.27.14-vanilla #1 SMP Fri Oct 16 00:56:25 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1
XFS 1.1
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: MD write performance issue
2009-10-16 8:45 MD write performance issue mark delfman
@ 2009-10-16 10:42 ` Asdo
2009-10-16 15:46 ` mark delfman
0 siblings, 1 reply; 3+ messages in thread
From: Asdo @ 2009-10-16 10:42 UTC (permalink / raw)
To: mark delfman; +Cc: linux-raid
mark delfman wrote:
> After further work we are sure that there is a significant write
> performance issue with either the Kernel+MD or...
Hm!
Pretty strange repeated ups and downs of the speed with increasing
kernel versions.
Have you checked:
that compile options are the same (preferably by taking 2.6.31 compile
options and porting them down)
disk schedulers are the same
the test was long enough to level jitters, like 2-3 minutes
Also: looking at "iostat -x 1" during the transfer could show something...
Apart from this, I confirm I noticed in my 2.6.31-rc? earlier tests,
that performances on xfs writes were very inconsistent :
These were my benchmarks (I wrote them on file at that time):
Stripe_cache_size was 1024, 13 devices raid-5:
bs=1M -> 206MB/s
bs=256K -> 229MB/s
retrying soon after, identical settings:
bs=1M -> 129MB/s
bs=256K -> 140MB/s
Transfer speed was hence very unreliable, depending on something that is
not clearly user visible... maybe dirty page cache? I thought that
depending on the exact amount of data being pushed out by the pdflush at
the first round, that would cause a sequence of read-modify-write stuff
which would cause further read-modify-write and further instability
later on. But I was doing that with raid-5 while you Mark are using
raid-0 right? My theory doesn't hold on raid-0.
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: MD write performance issue
2009-10-16 10:42 ` Asdo
@ 2009-10-16 15:46 ` mark delfman
0 siblings, 0 replies; 3+ messages in thread
From: mark delfman @ 2009-10-16 15:46 UTC (permalink / raw)
To: Asdo; +Cc: linux-raid
Quick update on .30.x kernels... which are still showing MD reduced performance:
Linux linux-tlfp 2.6.30-vanilla #1 SMP Fri Oct 16 14:22:54 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1
XFS 870 MB/s
Linux linux-tlfp 2.6.31.3-vanilla #1 SMP Fri Oct 16 14:52:09 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1
XFS 920 MB/s
linux-tlfp:/ # uname -a
Linux linux-tlfp 2.6.31.2-vanilla #1 SMP Fri Oct 16 15:44:44 BST 2009
x86_64 x86_64 x86_64 GNU/Linux
RAW: 1.1
XFS: 935 MB/s
On Fri, Oct 16, 2009 at 11:42 AM, Asdo <asdo@shiftmail.org> wrote:
> mark delfman wrote:
>>
>> After further work we are sure that there is a significant write
>> performance issue with either the Kernel+MD or...
>
> Hm!
> Pretty strange repeated ups and downs of the speed with increasing kernel
> versions.
>
> Have you checked:
> that compile options are the same (preferably by taking 2.6.31 compile
> options and porting them down)
> disk schedulers are the same
> the test was long enough to level jitters, like 2-3 minutes
> Also: looking at "iostat -x 1" during the transfer could show something...
>
> Apart from this, I confirm I noticed in my 2.6.31-rc? earlier tests, that
> performances on xfs writes were very inconsistent :
> These were my benchmarks (I wrote them on file at that time):
>
> Stripe_cache_size was 1024, 13 devices raid-5:
>
> bs=1M -> 206MB/s
> bs=256K -> 229MB/s
>
> retrying soon after, identical settings:
>
> bs=1M -> 129MB/s
> bs=256K -> 140MB/s
>
>
> Transfer speed was hence very unreliable, depending on something that is not
> clearly user visible... maybe dirty page cache? I thought that depending on
> the exact amount of data being pushed out by the pdflush at the first round,
> that would cause a sequence of read-modify-write stuff which would cause
> further read-modify-write and further instability later on. But I was doing
> that with raid-5 while you Mark are using raid-0 right? My theory doesn't
> hold on raid-0.
>
>
>
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2009-10-16 15:46 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-10-16 8:45 MD write performance issue mark delfman
2009-10-16 10:42 ` Asdo
2009-10-16 15:46 ` mark delfman
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).