* XFS write speed drop @ 2006-05-19 9:34 Jan Engelhardt 2006-05-22 0:53 ` Nathan Scott 0 siblings, 1 reply; 12+ messages in thread From: Jan Engelhardt @ 2006-05-19 9:34 UTC (permalink / raw) To: linux-xfs; +Cc: Linux Kernel Mailing List Hello, I have noticed that after an upgrade from 2.6.16-rcX -> 2.6.17-rc4, writes to one (hdc) xfs filesystem have become significantly slower (factor 6 to 8), like if -o sync (or funky journal options on ext3) was on. Also, reads would stall until writes have completed. I would only expect such behavior when /proc/sys/vm/dirty_* are set to high values (like 95%, like I do on a notebook). hda remained fast. It eventually turned out that it are the log barriers; hda does not support barriers and XFS gave a nice kernel message hint that pointed me to try out a mount flag. -o nobarrier makes it fast again. Jan Engelhardt -- ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-19 9:34 XFS write speed drop Jan Engelhardt @ 2006-05-22 0:53 ` Nathan Scott 2006-05-22 12:21 ` Jan Engelhardt 0 siblings, 1 reply; 12+ messages in thread From: Nathan Scott @ 2006-05-22 0:53 UTC (permalink / raw) To: Jan Engelhardt; +Cc: linux-xfs, Linux Kernel Mailing List On Fri, May 19, 2006 at 11:34:28AM +0200, Jan Engelhardt wrote: > Hello, Hi Jan, > I have noticed that after an upgrade from 2.6.16-rcX -> 2.6.17-rc4, writes > to one (hdc) xfs filesystem have become significantly slower (factor 6 to Buffered writes? Direct writes? Sync writes? Log writes? You're a bit light on details here. Can you send the benchmark results themselves please? (as in, the test(s) you've run that lead you to see 6-8x, and the data those tests produced). Also, xfs_info output, and maybe list the device driver(s) involved here too. thanks! -- Nathan ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-22 0:53 ` Nathan Scott @ 2006-05-22 12:21 ` Jan Engelhardt 2006-05-22 22:43 ` Nathan Scott 2006-05-23 1:26 ` David Chinner 0 siblings, 2 replies; 12+ messages in thread From: Jan Engelhardt @ 2006-05-22 12:21 UTC (permalink / raw) To: Nathan Scott; +Cc: linux-xfs, Linux Kernel Mailing List Hi Nathan, >> I have noticed that after an upgrade from 2.6.16-rcX -> 2.6.17-rc4, writes >> to one (hdc) xfs filesystem have become significantly slower (factor 6 to > >Buffered writes? Direct writes? Sync writes? Log writes? How would I know? :) "The standard thing". I shuffled a bit in the source code and the word 'order' is quite often around 'barrier', so I thought the 'barrier' option makes the IO scheduler strictly ordered, that is, write log first, then data. (like ext3's data=ordered). Anyhow, just see below >You're a bit light on details here. > >Can you send the benchmark results themselves please? (as in, >the test(s) you've run that lead you to see 6-8x, and the data >those tests produced). Also, xfs_info output, and maybe list >the device driver(s) involved here too. Drivers ======= xfs.ko sis5513.ko ide_disk.ko ide_core.ko xfs_info ======== 14:20 shanghai:/root # xfs_info / meta-data=/dev/hda3 isize=256 agcount=19, agsize=462621 blks = sectsz=512 attr=1 data = bsize=4096 blocks=8586742, imaxpct=25 = sunit=0 swidth=0 blks, unwritten=1 naming =version 2 bsize=4096 log =internal bsize=4096 blocks=3614, version=1 = sectsz=512 sunit=0 blks realtime =none extsz=65536 blocks=0, rtextents=0 14:20 shanghai:/root # xfs_info /D meta-data=/dev/hdc2 isize=256 agcount=16, agsize=3821963 blks = sectsz=512 attr=1 data = bsize=4096 blocks=61151408, imaxpct=25 = sunit=0 swidth=0 blks, unwritten=1 naming =version 2 bsize=4096 log =internal bsize=4096 blocks=29859, version=1 = sectsz=512 sunit=0 blks realtime =none extsz=65536 blocks=0, rtextents=0 14:20 shanghai:/root # fdisk -l /dev/hda Disk /dev/hda: 40.0 GB, 40020664320 bytes 255 heads, 63 sectors/track, 4865 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/hda1 1 64 514048+ 82 Linux swap / Solaris /dev/hda2 * 65 588 4209030 c W95 FAT32 (LBA) /dev/hda3 590 4865 34346970 83 Linux 14:21 shanghai:/root # fdisk -l /dev/hdc Disk /dev/hdc: 251.0 GB, 251000193024 bytes 255 heads, 63 sectors/track, 30515 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/hdc1 1 63 506016 82 Linux swap / Solaris /dev/hdc2 64 30515 244605690 83 Linux So here are the results. I have run these tests from single-user mode (-b bootflag) so that no other userspace programs would interfere. Script started on Mon May 22 15:32:29 2006 15:32 (none):~ # mount /dev/hdc2 /D -o barrier 15:32 (none):~ # df -Th / /D Filesystem Type Size Used Avail Use% Mounted on /dev/hda3 xfs 33G 6.7G 27G 21% / /dev/hdc2 xfs 234G 142G 92G 61% /D 15:33 (none):~ # cat /proc/mounts rootfs / rootfs rw 0 0 udev /dev tmpfs rw 0 0 /dev/hda3 / xfs rw 0 0 devpts /dev/pts devpts rw 0 0 /dev/hdc2 /D xfs rw 0 0 proc /proc proc rw 0 0 A small nitpick BTW, the barrier flag is not listed in /proc/mounts. hda is always in nobarrier mode. <6>hda: cache flushes not supported <6>hdc: cache flushes supported <5>Filesystem "hda3": Disabling barriers, not supported by the underlying device hdc can do barriers if wanted. CASE 1: Copying from one disk to another ======================================== Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090 directories. 15:33 (none):~ # cd /tmp 15:33 (none):/tmp # time rsync -PHSav kernel-source/ /D/kernel-0/ >/dev/null real 7m31.776s user 0m7.030s sys 0m14.450s I rebooted here (only here - no more reboots below) to get a cold slate. 15:42 (none):~ # mount /dev/hdc2 /D -o nobarrier 15:42 (none):/tmp # rsync -PHSav kernel-source/ /D/kernel-1/ >/dev/null real 2m14.401s user 0m7.290s sys 0m14.900s 15:45 (none):/tmp # umount /D CASE 2: Removing ================ Remove the copies we created in case 1. 15:45 (none):/tmp # mount /dev/hdc2 /D -o barrier 15:45 (none):/tmp # cd /D 15:45 (none):/D # time rm -Rf kernel-0 real 3m31.901s user 0m0.050s sys 0m3.140s 15:49 (none):/D # cd / 15:49 (none):/ # umount /D 15:49 (none):/ # mount /dev/hdc2 /D -o nobarrier 15:49 (none):/ # cd /D 15:49 (none):/D # time rm -Rf kernel-1 real 0m53.471s user 0m0.070s sys 0m1.990s 15:50 (none):/D # cd / 15:50 (none):/ # umount /D CASE 3: Copying to the same disk ================================ Copy a file from hdc2 to hdc2. The result is _extremely_ interesting. So the 'barrier' thing is only relevant for large sets of files, it seems. 15:50 (none):/ # mount /dev/hdc2 /D -o barrier 15:50 (none):/ # cd /D/Video 15:50 (none):/D/Video # ls -l The* -rw-r--r-- 1 jengelh users 6861772800 Mar 17 23:02 TLK3.iso 15:50 (none):/D/Video # time cp TLK3.iso ../tlk-0.iso real 8m1.051s user 0m0.670s sys 0m41.120s 15:59 (none):/D/Video # cd / 15:59 (none):/ # umount /D 15:59 (none):/ # mount /dev/hdc2 /D -o nobarrier 15:59 (none):/ # Kcd /D/Video 15:59 (none):/D/Video # time cp TLK3.iso ../tlk-1.iso real 7m53.275s user 0m0.560s sys 0m40.010s 16:07 (none):/D/Video # cd / 16:08 (none):/ # umount /D 16:08 (none):/ # exit Script done on Mon May 22 16:08:09 2006 Jan Engelhardt -- ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-22 12:21 ` Jan Engelhardt @ 2006-05-22 22:43 ` Nathan Scott 2006-05-23 13:23 ` Jan Engelhardt 2006-05-23 1:26 ` David Chinner 1 sibling, 1 reply; 12+ messages in thread From: Nathan Scott @ 2006-05-22 22:43 UTC (permalink / raw) To: Jan Engelhardt; +Cc: linux-xfs, Linux Kernel Mailing List Hi Jan, On Mon, May 22, 2006 at 02:21:48PM +0200, Jan Engelhardt wrote: > I shuffled a bit in the source code and the word 'order' is quite often > around 'barrier', so I thought the 'barrier' option makes the IO > scheduler strictly ordered, that is, write log first, then data. (like Not quite, theres an explanation here: http://oss.sgi.com/projects/xfs/faq.html#wcache > >Can you send the benchmark results themselves please? (as in, > >the test(s) you've run that lead you to see 6-8x, and the data > >those tests produced). Also, xfs_info output, and maybe list > >the device driver(s) involved here too. Thanks! > xfs_info > ======== > 14:20 shanghai:/root # xfs_info / > meta-data=/dev/hda3 isize=256 agcount=19, agsize=462621 blks > = sectsz=512 attr=1 > data = bsize=4096 blocks=8586742, imaxpct=25 > = sunit=0 swidth=0 blks, unwritten=1 > naming =version 2 bsize=4096 > log =internal bsize=4096 blocks=3614, version=1 > = sectsz=512 sunit=0 blks > realtime =none extsz=65536 blocks=0, rtextents=0 > 14:20 shanghai:/root # xfs_info /D > meta-data=/dev/hdc2 isize=256 agcount=16, agsize=3821963 blks > = sectsz=512 attr=1 > data = bsize=4096 blocks=61151408, imaxpct=25 > = sunit=0 swidth=0 blks, unwritten=1 > naming =version 2 bsize=4096 > log =internal bsize=4096 blocks=29859, version=1 > = sectsz=512 sunit=0 blks > realtime =none extsz=65536 blocks=0, rtextents=0 > ... > Script started on Mon May 22 15:32:29 2006 > 15:32 (none):~ # mount /dev/hdc2 /D -o barrier > 15:32 (none):~ # df -Th / /D > Filesystem Type Size Used Avail Use% Mounted on > /dev/hda3 xfs 33G 6.7G 27G 21% / > /dev/hdc2 xfs 234G 142G 92G 61% /D > A small nitpick BTW, the barrier flag is not listed in /proc/mounts. Its the default, and so we don't list that. Quirk of the code I guess, we could add code to distinguish default from requested but it doesn't seem worth it really. > hda is always in nobarrier mode. > <6>hda: cache flushes not supported > <6>hdc: cache flushes supported > <5>Filesystem "hda3": Disabling barriers, not supported by the underlying > device > hdc can do barriers if wanted. > > CASE 1: Copying from one disk to another > ======================================== > Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090 > directories. OK, we can call this a metadata intensive workload - lots of small files, lots of creates. Barriers will hurt the most here, as we'd already have been log I/O bound most likely, and I'd expect barriers to only slow that further. > 15:33 (none):/tmp # time rsync -PHSav kernel-source/ /D/kernel-0/ >/dev/null > real 7m31.776s > user 0m7.030s > sys 0m14.450s versus: > 15:42 (none):/tmp # rsync -PHSav kernel-source/ /D/kernel-1/ >/dev/null > real 2m14.401s > user 0m7.290s > sys 0m14.900s Yep, note the user/sys shows no change, we're basically IO bound in both tests, and barriers are hurting (as expected). You may be able to combat this by switching to version 2 log format (see the xfs_db version command in the _very_ latest xfs_db, as of a few days ago, which allows this switch to be made on an unmounted fs). Then you will be able to use larger incore log buffers, and this should reduce the impact the barriers have (fewer, larger log IOs, so fewer barriers). > CASE 2: Removing > ================ > Remove the copies we created in case 1. > > 15:45 (none):/tmp # mount /dev/hdc2 /D -o barrier > 15:45 (none):/D # time rm -Rf kernel-0 > real 3m31.901s > user 0m0.050s > sys 0m3.140s versus: > 15:49 (none):/ # mount /dev/hdc2 /D -o nobarrier > 15:49 (none):/D # time rm -Rf kernel-1 > > real 0m53.471s > user 0m0.070s > sys 0m1.990s > 15:50 (none):/D # cd / > 15:50 (none):/ # umount /D Also metadata intensive, of course. All the same issues as above, and the same techniques should be used to address it. Odd that the system time jumped here. Roughly the same decrease in performance though (3-4x). > CASE 3: Copying to the same disk > ================================ > Copy a file from hdc2 to hdc2. The result is _extremely_ interesting. So the > 'barrier' thing is only relevant for large sets of files, it seems. Its consistent, and easily understood however... > 15:50 (none):/ # mount /dev/hdc2 /D -o barrier > 15:50 (none):/D/Video # ls -l The* > -rw-r--r-- 1 jengelh users 6861772800 Mar 17 23:02 TLK3.iso > 15:50 (none):/D/Video # time cp TLK3.iso ../tlk-0.iso > real 8m1.051s > user 0m0.670s > sys 0m41.120s versus: > 15:59 (none):/ # mount /dev/hdc2 /D -o nobarrier > 15:59 (none):/D/Video # time cp TLK3.iso ../tlk-1.iso > real 7m53.275s > user 0m0.560s > sys 0m40.010s This one is a file throughput workload, and so there are far fewer metadata updates involved here. This means far fewer barrier operations (far fewer log writes), so we should expect the results to be as they are - slightly worse with barriers enabled, nothing drastic though. So, I agree, you're seeing the cost of write barriers here, but I don't see anything unexpected (unfortunately for you, I guess). The FAQ entry above will explain why they're enabled by default. Try the v2 log change I suggested, hopefully that will mitigate the problem somewhat. Alternatively, buy some decent hardware if you're after performance. ;) cheers. -- Nathan ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-22 22:43 ` Nathan Scott @ 2006-05-23 13:23 ` Jan Engelhardt 2006-05-23 22:22 ` Nathan Scott 2006-05-26 7:20 ` Helge Hafting 0 siblings, 2 replies; 12+ messages in thread From: Jan Engelhardt @ 2006-05-23 13:23 UTC (permalink / raw) To: Nathan Scott; +Cc: linux-xfs, Linux Kernel Mailing List >> CASE 1: Copying from one disk to another >> ======================================== >> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090 >> directories. > >OK, we can call this a metadata intensive workload - lots of small >files, lots of creates. Barriers will hurt the most here, as we'd >already have been log I/O bound most likely, and I'd expect barriers >to only slow that further. > Yes and the most important thing is that someone made -o barrier the default and did not notice. Someone else? :-D >Yep, note the user/sys shows no change, we're basically IO bound in >both tests, and barriers are hurting (as expected). >> CASE 2: Removing >> ================ >> Remove the copies we created in case 1. >> >> 15:45 (none):/tmp # mount /dev/hdc2 /D -o barrier >> 15:45 (none):/D # time rm -Rf kernel-0 >> real 3m31.901s >> user 0m0.050s >> sys 0m3.140s > >versus: > >> 15:49 (none):/ # mount /dev/hdc2 /D -o nobarrier >> 15:49 (none):/D # time rm -Rf kernel-1 >> >> real 0m53.471s >> user 0m0.070s >> sys 0m1.990s >> 15:50 (none):/D # cd / >> 15:50 (none):/ # umount /D > >Also metadata intensive, of course. All the same issues as above, >and the same techniques should be used to address it. > >Odd that the system time jumped here. Roughly the same decrease in >performance though (3-4x). You may, or may not, take that serious. I only ran each test once (since 3m31s vs 53s shows "enough" of the issue). >So, I agree, you're seeing the cost of write barriers here, but I >don't see anything unexpected (unfortunately for you, I guess). > >The FAQ entry above will explain why they're enabled by default. >Try the v2 log change I suggested, hopefully that will mitigate >the problem somewhat. Alternatively, buy some decent hardware if >you're after performance. ;) > Well as mentioned, -o nobarrier solved it, and that's it. I do not actually need barriers (or an UPS, to poke on another thread), because power failures are rather rare in Germany. Jan Engelhardt -- ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-23 13:23 ` Jan Engelhardt @ 2006-05-23 22:22 ` Nathan Scott 2006-05-23 23:17 ` Nathan Scott 2006-05-26 7:20 ` Helge Hafting 1 sibling, 1 reply; 12+ messages in thread From: Nathan Scott @ 2006-05-23 22:22 UTC (permalink / raw) To: Jan Engelhardt; +Cc: xfs, linux-kernel On Tue, May 23, 2006 at 03:23:31PM +0200, Jan Engelhardt wrote: > >> CASE 1: Copying from one disk to another > >> ======================================== > >> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090 > >> directories. > > > >OK, we can call this a metadata intensive workload - lots of small > >files, lots of creates. Barriers will hurt the most here, as we'd > >already have been log I/O bound most likely, and I'd expect barriers > >to only slow that further. > > > Yes and the most important thing is that someone made -o barrier the > default and did not notice. Someone else? :-D Not sure what you're trying to say here. Yes, barriers are on by default now if the hardware supports them, yes, they will slow things down relative to write-cache-without-barriers, and yes we all knew that ... its not the case that someone "did not notice" or forgot about something. There is no doubt that this is the right thing to be doing by default - there's no way that I can tell from inside XFS in the kernel that you have a UPS. ;) cheers. -- Nathan ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-23 22:22 ` Nathan Scott @ 2006-05-23 23:17 ` Nathan Scott 2006-05-24 7:16 ` Jeremy Higdon 0 siblings, 1 reply; 12+ messages in thread From: Nathan Scott @ 2006-05-23 23:17 UTC (permalink / raw) To: Jan Engelhardt; +Cc: xfs, linux-kernel On Wed, May 24, 2006 at 08:22:19AM +1000, Nathan Scott wrote: > On Tue, May 23, 2006 at 03:23:31PM +0200, Jan Engelhardt wrote: > > >> CASE 1: Copying from one disk to another > > >> ======================================== > > >> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090 > > >> directories. > > > > > >OK, we can call this a metadata intensive workload - lots of small > > >files, lots of creates. Barriers will hurt the most here, as we'd > > >already have been log I/O bound most likely, and I'd expect barriers > > >to only slow that further. > > > > > Yes and the most important thing is that someone made -o barrier the > > default and did not notice. Someone else? :-D > > Not sure what you're trying to say here. Yes, barriers are on > by default now if the hardware supports them, yes, they will > slow things down relative to write-cache-without-barriers, and > yes we all knew that ... its not the case that someone "did not > notice" or forgot about something. There is no doubt that this > is the right thing to be doing by default - there's no way that > I can tell from inside XFS in the kernel that you have a UPS. ;) Oh, I realised I've slightly misread your mail, you said... | I do not actually need barriers (or an UPS, to poke on another thread), | because power failures are rather rare in Germany. Hmm, even harder for us to detect at runtime, in the kernel, that you're in Germany. :) Power failures aren't the only thing to cause crashes, however. cheers. -- Nathan ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-23 23:17 ` Nathan Scott @ 2006-05-24 7:16 ` Jeremy Higdon 0 siblings, 0 replies; 12+ messages in thread From: Jeremy Higdon @ 2006-05-24 7:16 UTC (permalink / raw) To: Nathan Scott; +Cc: Jan Engelhardt, xfs, linux-kernel On Wed, May 24, 2006 at 09:17:41AM +1000, Nathan Scott wrote: > On Wed, May 24, 2006 at 08:22:19AM +1000, Nathan Scott wrote: > > > > Not sure what you're trying to say here. Yes, barriers are on > > by default now if the hardware supports them, yes, they will > > slow things down relative to write-cache-without-barriers, and > > yes we all knew that ... its not the case that someone "did not > > notice" or forgot about something. There is no doubt that this > > is the right thing to be doing by default - there's no way that > > I can tell from inside XFS in the kernel that you have a UPS. ;) > > Oh, I realised I've slightly misread your mail, you said... > > | I do not actually need barriers (or an UPS, to poke on another thread), > | because power failures are rather rare in Germany. > > Hmm, even harder for us to detect at runtime, in the kernel, > that you're in Germany. :) > > Power failures aren't the only thing to cause crashes, however. There have been several examples of filesystem corruption without power failures too, though that is a common cause. It can even happen on a normal system shutdown if there is no synchronize operation to the disk at shutdown time, depending on how much data is in the cache and how long you have until the ATA chip is reset. jeremy ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-23 13:23 ` Jan Engelhardt 2006-05-23 22:22 ` Nathan Scott @ 2006-05-26 7:20 ` Helge Hafting 2006-05-26 8:51 ` Jan Engelhardt 1 sibling, 1 reply; 12+ messages in thread From: Helge Hafting @ 2006-05-26 7:20 UTC (permalink / raw) To: Jan Engelhardt; +Cc: Nathan Scott, linux-xfs, Linux Kernel Mailing List Jan Engelhardt wrote: > Well as mentioned, -o nobarrier solved it, and that's it. I do not actually > need barriers (or an UPS, to poke on another thread), because power > failures are rather rare in Germany. Then there are mistakes like someone stepping on the power cord, pulling out the wrong one, drilling holes in the wall, there are kernel crashes, there is lightning, there is the possibility that some component inside the box just breaks. Helge Hafting ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-26 7:20 ` Helge Hafting @ 2006-05-26 8:51 ` Jan Engelhardt 2006-05-26 21:06 ` Bernd Eckenfels 0 siblings, 1 reply; 12+ messages in thread From: Jan Engelhardt @ 2006-05-26 8:51 UTC (permalink / raw) To: Helge Hafting; +Cc: Nathan Scott, linux-xfs, Linux Kernel Mailing List >> Well as mentioned, -o nobarrier solved it, and that's it. I do not >> actually need barriers (or an UPS, to poke on another thread), because >> power failures are rather rare in Germany. > > Then there are mistakes like someone stepping on the power > cord, pulling out the wrong one, drilling holes in the wall, > there are kernel crashes, there is lightning, > there is the possibility that some > component inside the box just breaks. > A mathematician has fear of flying because there could possibly be a terrorist bomb inside the plane. What does he do? Taking an bomb with himself, because the probablity that there are two bombs inside a plane is lower than one. What I am wanting to say is that the factors you mentioned just do not happen here, at this one - home - box. Jan Engelhardt -- ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-26 8:51 ` Jan Engelhardt @ 2006-05-26 21:06 ` Bernd Eckenfels 0 siblings, 0 replies; 12+ messages in thread From: Bernd Eckenfels @ 2006-05-26 21:06 UTC (permalink / raw) To: linux-kernel Jan Engelhardt <jengelh@linux01.gwdg.de> wrote: > What I am wanting to say is that the factors you mentioned just do not happen > here, at this one - home - box. They will as soon as you run the box in unsafe mode and you have important data without backup :) Gruss Bernd ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop 2006-05-22 12:21 ` Jan Engelhardt 2006-05-22 22:43 ` Nathan Scott @ 2006-05-23 1:26 ` David Chinner 1 sibling, 0 replies; 12+ messages in thread From: David Chinner @ 2006-05-23 1:26 UTC (permalink / raw) To: Jan Engelhardt; +Cc: Nathan Scott, linux-xfs, Linux Kernel Mailing List On Mon, May 22, 2006 at 02:21:48PM +0200, Jan Engelhardt wrote: > >You're a bit light on details here. > > > >Can you send the benchmark results themselves please? (as in, > >the test(s) you've run that lead you to see 6-8x, and the data > >those tests produced). Also, xfs_info output, and maybe list > >the device driver(s) involved here too. Jan, just out of curiousity, can you also run the tests with barriers off and the drive write cache turned off? I would expect the cache+barriers to be faster... Cheers, Dave. -- Dave Chinner R&D Software Enginner SGI Australian Software Group ^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2006-05-26 21:06 UTC | newest] Thread overview: 12+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2006-05-19 9:34 XFS write speed drop Jan Engelhardt 2006-05-22 0:53 ` Nathan Scott 2006-05-22 12:21 ` Jan Engelhardt 2006-05-22 22:43 ` Nathan Scott 2006-05-23 13:23 ` Jan Engelhardt 2006-05-23 22:22 ` Nathan Scott 2006-05-23 23:17 ` Nathan Scott 2006-05-24 7:16 ` Jeremy Higdon 2006-05-26 7:20 ` Helge Hafting 2006-05-26 8:51 ` Jan Engelhardt 2006-05-26 21:06 ` Bernd Eckenfels 2006-05-23 1:26 ` David Chinner
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox