* XFS write speed drop
@ 2006-05-19 9:34 Jan Engelhardt
2006-05-22 0:53 ` Nathan Scott
0 siblings, 1 reply; 12+ messages in thread
From: Jan Engelhardt @ 2006-05-19 9:34 UTC (permalink / raw)
To: linux-xfs; +Cc: Linux Kernel Mailing List
Hello,
I have noticed that after an upgrade from 2.6.16-rcX -> 2.6.17-rc4, writes
to one (hdc) xfs filesystem have become significantly slower (factor 6 to
8), like if -o sync (or funky journal options on ext3) was on. Also, reads
would stall until writes have completed. I would only expect such behavior
when /proc/sys/vm/dirty_* are set to high values (like 95%, like I do on a
notebook). hda remained fast.
It eventually turned out that it are the log barriers; hda does not support
barriers and XFS gave a nice kernel message hint that pointed me to try out
a mount flag. -o nobarrier makes it fast again.
Jan Engelhardt
--
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-19 9:34 XFS write speed drop Jan Engelhardt
@ 2006-05-22 0:53 ` Nathan Scott
2006-05-22 12:21 ` Jan Engelhardt
0 siblings, 1 reply; 12+ messages in thread
From: Nathan Scott @ 2006-05-22 0:53 UTC (permalink / raw)
To: Jan Engelhardt; +Cc: linux-xfs, Linux Kernel Mailing List
On Fri, May 19, 2006 at 11:34:28AM +0200, Jan Engelhardt wrote:
> Hello,
Hi Jan,
> I have noticed that after an upgrade from 2.6.16-rcX -> 2.6.17-rc4, writes
> to one (hdc) xfs filesystem have become significantly slower (factor 6 to
Buffered writes? Direct writes? Sync writes? Log writes?
You're a bit light on details here.
Can you send the benchmark results themselves please? (as in,
the test(s) you've run that lead you to see 6-8x, and the data
those tests produced). Also, xfs_info output, and maybe list
the device driver(s) involved here too.
thanks!
--
Nathan
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-22 0:53 ` Nathan Scott
@ 2006-05-22 12:21 ` Jan Engelhardt
2006-05-22 22:43 ` Nathan Scott
2006-05-23 1:26 ` David Chinner
0 siblings, 2 replies; 12+ messages in thread
From: Jan Engelhardt @ 2006-05-22 12:21 UTC (permalink / raw)
To: Nathan Scott; +Cc: linux-xfs, Linux Kernel Mailing List
Hi Nathan,
>> I have noticed that after an upgrade from 2.6.16-rcX -> 2.6.17-rc4, writes
>> to one (hdc) xfs filesystem have become significantly slower (factor 6 to
>
>Buffered writes? Direct writes? Sync writes? Log writes?
How would I know? :) "The standard thing".
I shuffled a bit in the source code and the word 'order' is quite often
around 'barrier', so I thought the 'barrier' option makes the IO
scheduler strictly ordered, that is, write log first, then data. (like
ext3's data=ordered). Anyhow, just see below
>You're a bit light on details here.
>
>Can you send the benchmark results themselves please? (as in,
>the test(s) you've run that lead you to see 6-8x, and the data
>those tests produced). Also, xfs_info output, and maybe list
>the device driver(s) involved here too.
Drivers
=======
xfs.ko sis5513.ko ide_disk.ko ide_core.ko
xfs_info
========
14:20 shanghai:/root # xfs_info /
meta-data=/dev/hda3 isize=256 agcount=19, agsize=462621 blks
= sectsz=512 attr=1
data = bsize=4096 blocks=8586742, imaxpct=25
= sunit=0 swidth=0 blks, unwritten=1
naming =version 2 bsize=4096
log =internal bsize=4096 blocks=3614, version=1
= sectsz=512 sunit=0 blks
realtime =none extsz=65536 blocks=0, rtextents=0
14:20 shanghai:/root # xfs_info /D
meta-data=/dev/hdc2 isize=256 agcount=16, agsize=3821963 blks
= sectsz=512 attr=1
data = bsize=4096 blocks=61151408, imaxpct=25
= sunit=0 swidth=0 blks, unwritten=1
naming =version 2 bsize=4096
log =internal bsize=4096 blocks=29859, version=1
= sectsz=512 sunit=0 blks
realtime =none extsz=65536 blocks=0, rtextents=0
14:20 shanghai:/root # fdisk -l /dev/hda
Disk /dev/hda: 40.0 GB, 40020664320 bytes
255 heads, 63 sectors/track, 4865 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot Start End Blocks Id System
/dev/hda1 1 64 514048+ 82 Linux swap / Solaris
/dev/hda2 * 65 588 4209030 c W95 FAT32 (LBA)
/dev/hda3 590 4865 34346970 83 Linux
14:21 shanghai:/root # fdisk -l /dev/hdc
Disk /dev/hdc: 251.0 GB, 251000193024 bytes
255 heads, 63 sectors/track, 30515 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Device Boot Start End Blocks Id System
/dev/hdc1 1 63 506016 82 Linux swap / Solaris
/dev/hdc2 64 30515 244605690 83 Linux
So here are the results. I have run these tests from single-user mode (-b
bootflag) so that no other userspace programs would interfere.
Script started on Mon May 22 15:32:29 2006
15:32 (none):~ # mount /dev/hdc2 /D -o barrier
15:32 (none):~ # df -Th / /D
Filesystem Type Size Used Avail Use% Mounted on
/dev/hda3 xfs 33G 6.7G 27G 21% /
/dev/hdc2 xfs 234G 142G 92G 61% /D
15:33 (none):~ # cat /proc/mounts
rootfs / rootfs rw 0 0
udev /dev tmpfs rw 0 0
/dev/hda3 / xfs rw 0 0
devpts /dev/pts devpts rw 0 0
/dev/hdc2 /D xfs rw 0 0
proc /proc proc rw 0 0
A small nitpick BTW, the barrier flag is not listed in /proc/mounts.
hda is always in nobarrier mode.
<6>hda: cache flushes not supported
<6>hdc: cache flushes supported
<5>Filesystem "hda3": Disabling barriers, not supported by the underlying
device
hdc can do barriers if wanted.
CASE 1: Copying from one disk to another
========================================
Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090
directories.
15:33 (none):~ # cd /tmp
15:33 (none):/tmp # time rsync -PHSav kernel-source/ /D/kernel-0/ >/dev/null
real 7m31.776s
user 0m7.030s
sys 0m14.450s
I rebooted here (only here - no more reboots below) to get a cold slate.
15:42 (none):~ # mount /dev/hdc2 /D -o nobarrier
15:42 (none):/tmp # rsync -PHSav kernel-source/ /D/kernel-1/ >/dev/null
real 2m14.401s
user 0m7.290s
sys 0m14.900s
15:45 (none):/tmp # umount /D
CASE 2: Removing
================
Remove the copies we created in case 1.
15:45 (none):/tmp # mount /dev/hdc2 /D -o barrier
15:45 (none):/tmp # cd /D
15:45 (none):/D # time rm -Rf kernel-0
real 3m31.901s
user 0m0.050s
sys 0m3.140s
15:49 (none):/D # cd /
15:49 (none):/ # umount /D
15:49 (none):/ # mount /dev/hdc2 /D -o nobarrier
15:49 (none):/ # cd /D
15:49 (none):/D # time rm -Rf kernel-1
real 0m53.471s
user 0m0.070s
sys 0m1.990s
15:50 (none):/D # cd /
15:50 (none):/ # umount /D
CASE 3: Copying to the same disk
================================
Copy a file from hdc2 to hdc2. The result is _extremely_ interesting. So the
'barrier' thing is only relevant for large sets of files, it seems.
15:50 (none):/ # mount /dev/hdc2 /D -o barrier
15:50 (none):/ # cd /D/Video
15:50 (none):/D/Video # ls -l The*
-rw-r--r-- 1 jengelh users 6861772800 Mar 17 23:02 TLK3.iso
15:50 (none):/D/Video # time cp TLK3.iso ../tlk-0.iso
real 8m1.051s
user 0m0.670s
sys 0m41.120s
15:59 (none):/D/Video # cd /
15:59 (none):/ # umount /D
15:59 (none):/ # mount /dev/hdc2 /D -o nobarrier
15:59 (none):/ # Kcd /D/Video
15:59 (none):/D/Video # time cp TLK3.iso ../tlk-1.iso
real 7m53.275s
user 0m0.560s
sys 0m40.010s
16:07 (none):/D/Video # cd /
16:08 (none):/ # umount /D
16:08 (none):/ # exit
Script done on Mon May 22 16:08:09 2006
Jan Engelhardt
--
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-22 12:21 ` Jan Engelhardt
@ 2006-05-22 22:43 ` Nathan Scott
2006-05-23 13:23 ` Jan Engelhardt
2006-05-23 1:26 ` David Chinner
1 sibling, 1 reply; 12+ messages in thread
From: Nathan Scott @ 2006-05-22 22:43 UTC (permalink / raw)
To: Jan Engelhardt; +Cc: linux-xfs, Linux Kernel Mailing List
Hi Jan,
On Mon, May 22, 2006 at 02:21:48PM +0200, Jan Engelhardt wrote:
> I shuffled a bit in the source code and the word 'order' is quite often
> around 'barrier', so I thought the 'barrier' option makes the IO
> scheduler strictly ordered, that is, write log first, then data. (like
Not quite, theres an explanation here:
http://oss.sgi.com/projects/xfs/faq.html#wcache
> >Can you send the benchmark results themselves please? (as in,
> >the test(s) you've run that lead you to see 6-8x, and the data
> >those tests produced). Also, xfs_info output, and maybe list
> >the device driver(s) involved here too.
Thanks!
> xfs_info
> ========
> 14:20 shanghai:/root # xfs_info /
> meta-data=/dev/hda3 isize=256 agcount=19, agsize=462621 blks
> = sectsz=512 attr=1
> data = bsize=4096 blocks=8586742, imaxpct=25
> = sunit=0 swidth=0 blks, unwritten=1
> naming =version 2 bsize=4096
> log =internal bsize=4096 blocks=3614, version=1
> = sectsz=512 sunit=0 blks
> realtime =none extsz=65536 blocks=0, rtextents=0
> 14:20 shanghai:/root # xfs_info /D
> meta-data=/dev/hdc2 isize=256 agcount=16, agsize=3821963 blks
> = sectsz=512 attr=1
> data = bsize=4096 blocks=61151408, imaxpct=25
> = sunit=0 swidth=0 blks, unwritten=1
> naming =version 2 bsize=4096
> log =internal bsize=4096 blocks=29859, version=1
> = sectsz=512 sunit=0 blks
> realtime =none extsz=65536 blocks=0, rtextents=0
> ...
> Script started on Mon May 22 15:32:29 2006
> 15:32 (none):~ # mount /dev/hdc2 /D -o barrier
> 15:32 (none):~ # df -Th / /D
> Filesystem Type Size Used Avail Use% Mounted on
> /dev/hda3 xfs 33G 6.7G 27G 21% /
> /dev/hdc2 xfs 234G 142G 92G 61% /D
> A small nitpick BTW, the barrier flag is not listed in /proc/mounts.
Its the default, and so we don't list that. Quirk of the code I guess,
we could add code to distinguish default from requested but it doesn't
seem worth it really.
> hda is always in nobarrier mode.
> <6>hda: cache flushes not supported
> <6>hdc: cache flushes supported
> <5>Filesystem "hda3": Disabling barriers, not supported by the underlying
> device
> hdc can do barriers if wanted.
>
> CASE 1: Copying from one disk to another
> ========================================
> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090
> directories.
OK, we can call this a metadata intensive workload - lots of small
files, lots of creates. Barriers will hurt the most here, as we'd
already have been log I/O bound most likely, and I'd expect barriers
to only slow that further.
> 15:33 (none):/tmp # time rsync -PHSav kernel-source/ /D/kernel-0/ >/dev/null
> real 7m31.776s
> user 0m7.030s
> sys 0m14.450s
versus:
> 15:42 (none):/tmp # rsync -PHSav kernel-source/ /D/kernel-1/ >/dev/null
> real 2m14.401s
> user 0m7.290s
> sys 0m14.900s
Yep, note the user/sys shows no change, we're basically IO bound in
both tests, and barriers are hurting (as expected).
You may be able to combat this by switching to version 2 log format
(see the xfs_db version command in the _very_ latest xfs_db, as of a
few days ago, which allows this switch to be made on an unmounted fs).
Then you will be able to use larger incore log buffers, and this
should reduce the impact the barriers have (fewer, larger log IOs, so
fewer barriers).
> CASE 2: Removing
> ================
> Remove the copies we created in case 1.
>
> 15:45 (none):/tmp # mount /dev/hdc2 /D -o barrier
> 15:45 (none):/D # time rm -Rf kernel-0
> real 3m31.901s
> user 0m0.050s
> sys 0m3.140s
versus:
> 15:49 (none):/ # mount /dev/hdc2 /D -o nobarrier
> 15:49 (none):/D # time rm -Rf kernel-1
>
> real 0m53.471s
> user 0m0.070s
> sys 0m1.990s
> 15:50 (none):/D # cd /
> 15:50 (none):/ # umount /D
Also metadata intensive, of course. All the same issues as above,
and the same techniques should be used to address it.
Odd that the system time jumped here. Roughly the same decrease in
performance though (3-4x).
> CASE 3: Copying to the same disk
> ================================
> Copy a file from hdc2 to hdc2. The result is _extremely_ interesting. So the
> 'barrier' thing is only relevant for large sets of files, it seems.
Its consistent, and easily understood however...
> 15:50 (none):/ # mount /dev/hdc2 /D -o barrier
> 15:50 (none):/D/Video # ls -l The*
> -rw-r--r-- 1 jengelh users 6861772800 Mar 17 23:02 TLK3.iso
> 15:50 (none):/D/Video # time cp TLK3.iso ../tlk-0.iso
> real 8m1.051s
> user 0m0.670s
> sys 0m41.120s
versus:
> 15:59 (none):/ # mount /dev/hdc2 /D -o nobarrier
> 15:59 (none):/D/Video # time cp TLK3.iso ../tlk-1.iso
> real 7m53.275s
> user 0m0.560s
> sys 0m40.010s
This one is a file throughput workload, and so there are far fewer
metadata updates involved here. This means far fewer barrier
operations (far fewer log writes), so we should expect the results
to be as they are - slightly worse with barriers enabled, nothing
drastic though.
So, I agree, you're seeing the cost of write barriers here, but I
don't see anything unexpected (unfortunately for you, I guess).
The FAQ entry above will explain why they're enabled by default.
Try the v2 log change I suggested, hopefully that will mitigate
the problem somewhat. Alternatively, buy some decent hardware if
you're after performance. ;)
cheers.
--
Nathan
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-22 12:21 ` Jan Engelhardt
2006-05-22 22:43 ` Nathan Scott
@ 2006-05-23 1:26 ` David Chinner
1 sibling, 0 replies; 12+ messages in thread
From: David Chinner @ 2006-05-23 1:26 UTC (permalink / raw)
To: Jan Engelhardt; +Cc: Nathan Scott, linux-xfs, Linux Kernel Mailing List
On Mon, May 22, 2006 at 02:21:48PM +0200, Jan Engelhardt wrote:
> >You're a bit light on details here.
> >
> >Can you send the benchmark results themselves please? (as in,
> >the test(s) you've run that lead you to see 6-8x, and the data
> >those tests produced). Also, xfs_info output, and maybe list
> >the device driver(s) involved here too.
Jan, just out of curiousity, can you also run the tests with
barriers off and the drive write cache turned off? I would
expect the cache+barriers to be faster...
Cheers,
Dave.
--
Dave Chinner
R&D Software Enginner
SGI Australian Software Group
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-22 22:43 ` Nathan Scott
@ 2006-05-23 13:23 ` Jan Engelhardt
2006-05-23 22:22 ` Nathan Scott
2006-05-26 7:20 ` Helge Hafting
0 siblings, 2 replies; 12+ messages in thread
From: Jan Engelhardt @ 2006-05-23 13:23 UTC (permalink / raw)
To: Nathan Scott; +Cc: linux-xfs, Linux Kernel Mailing List
>> CASE 1: Copying from one disk to another
>> ========================================
>> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090
>> directories.
>
>OK, we can call this a metadata intensive workload - lots of small
>files, lots of creates. Barriers will hurt the most here, as we'd
>already have been log I/O bound most likely, and I'd expect barriers
>to only slow that further.
>
Yes and the most important thing is that someone made -o barrier the
default and did not notice. Someone else? :-D
>Yep, note the user/sys shows no change, we're basically IO bound in
>both tests, and barriers are hurting (as expected).
>> CASE 2: Removing
>> ================
>> Remove the copies we created in case 1.
>>
>> 15:45 (none):/tmp # mount /dev/hdc2 /D -o barrier
>> 15:45 (none):/D # time rm -Rf kernel-0
>> real 3m31.901s
>> user 0m0.050s
>> sys 0m3.140s
>
>versus:
>
>> 15:49 (none):/ # mount /dev/hdc2 /D -o nobarrier
>> 15:49 (none):/D # time rm -Rf kernel-1
>>
>> real 0m53.471s
>> user 0m0.070s
>> sys 0m1.990s
>> 15:50 (none):/D # cd /
>> 15:50 (none):/ # umount /D
>
>Also metadata intensive, of course. All the same issues as above,
>and the same techniques should be used to address it.
>
>Odd that the system time jumped here. Roughly the same decrease in
>performance though (3-4x).
You may, or may not, take that serious. I only ran each test once (since
3m31s vs 53s shows "enough" of the issue).
>So, I agree, you're seeing the cost of write barriers here, but I
>don't see anything unexpected (unfortunately for you, I guess).
>
>The FAQ entry above will explain why they're enabled by default.
>Try the v2 log change I suggested, hopefully that will mitigate
>the problem somewhat. Alternatively, buy some decent hardware if
>you're after performance. ;)
>
Well as mentioned, -o nobarrier solved it, and that's it. I do not actually
need barriers (or an UPS, to poke on another thread), because power
failures are rather rare in Germany.
Jan Engelhardt
--
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-23 13:23 ` Jan Engelhardt
@ 2006-05-23 22:22 ` Nathan Scott
2006-05-23 23:17 ` Nathan Scott
2006-05-26 7:20 ` Helge Hafting
1 sibling, 1 reply; 12+ messages in thread
From: Nathan Scott @ 2006-05-23 22:22 UTC (permalink / raw)
To: Jan Engelhardt; +Cc: xfs, linux-kernel
On Tue, May 23, 2006 at 03:23:31PM +0200, Jan Engelhardt wrote:
> >> CASE 1: Copying from one disk to another
> >> ========================================
> >> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090
> >> directories.
> >
> >OK, we can call this a metadata intensive workload - lots of small
> >files, lots of creates. Barriers will hurt the most here, as we'd
> >already have been log I/O bound most likely, and I'd expect barriers
> >to only slow that further.
> >
> Yes and the most important thing is that someone made -o barrier the
> default and did not notice. Someone else? :-D
Not sure what you're trying to say here. Yes, barriers are on
by default now if the hardware supports them, yes, they will
slow things down relative to write-cache-without-barriers, and
yes we all knew that ... its not the case that someone "did not
notice" or forgot about something. There is no doubt that this
is the right thing to be doing by default - there's no way that
I can tell from inside XFS in the kernel that you have a UPS. ;)
cheers.
--
Nathan
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-23 22:22 ` Nathan Scott
@ 2006-05-23 23:17 ` Nathan Scott
2006-05-24 7:16 ` Jeremy Higdon
0 siblings, 1 reply; 12+ messages in thread
From: Nathan Scott @ 2006-05-23 23:17 UTC (permalink / raw)
To: Jan Engelhardt; +Cc: xfs, linux-kernel
On Wed, May 24, 2006 at 08:22:19AM +1000, Nathan Scott wrote:
> On Tue, May 23, 2006 at 03:23:31PM +0200, Jan Engelhardt wrote:
> > >> CASE 1: Copying from one disk to another
> > >> ========================================
> > >> Copying a compiled 2.6.17-rc4 tree; 306907 KB in 28566 files in 2090
> > >> directories.
> > >
> > >OK, we can call this a metadata intensive workload - lots of small
> > >files, lots of creates. Barriers will hurt the most here, as we'd
> > >already have been log I/O bound most likely, and I'd expect barriers
> > >to only slow that further.
> > >
> > Yes and the most important thing is that someone made -o barrier the
> > default and did not notice. Someone else? :-D
>
> Not sure what you're trying to say here. Yes, barriers are on
> by default now if the hardware supports them, yes, they will
> slow things down relative to write-cache-without-barriers, and
> yes we all knew that ... its not the case that someone "did not
> notice" or forgot about something. There is no doubt that this
> is the right thing to be doing by default - there's no way that
> I can tell from inside XFS in the kernel that you have a UPS. ;)
Oh, I realised I've slightly misread your mail, you said...
| I do not actually need barriers (or an UPS, to poke on another thread),
| because power failures are rather rare in Germany.
Hmm, even harder for us to detect at runtime, in the kernel,
that you're in Germany. :)
Power failures aren't the only thing to cause crashes, however.
cheers.
--
Nathan
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-23 23:17 ` Nathan Scott
@ 2006-05-24 7:16 ` Jeremy Higdon
0 siblings, 0 replies; 12+ messages in thread
From: Jeremy Higdon @ 2006-05-24 7:16 UTC (permalink / raw)
To: Nathan Scott; +Cc: Jan Engelhardt, xfs, linux-kernel
On Wed, May 24, 2006 at 09:17:41AM +1000, Nathan Scott wrote:
> On Wed, May 24, 2006 at 08:22:19AM +1000, Nathan Scott wrote:
> >
> > Not sure what you're trying to say here. Yes, barriers are on
> > by default now if the hardware supports them, yes, they will
> > slow things down relative to write-cache-without-barriers, and
> > yes we all knew that ... its not the case that someone "did not
> > notice" or forgot about something. There is no doubt that this
> > is the right thing to be doing by default - there's no way that
> > I can tell from inside XFS in the kernel that you have a UPS. ;)
>
> Oh, I realised I've slightly misread your mail, you said...
>
> | I do not actually need barriers (or an UPS, to poke on another thread),
> | because power failures are rather rare in Germany.
>
> Hmm, even harder for us to detect at runtime, in the kernel,
> that you're in Germany. :)
>
> Power failures aren't the only thing to cause crashes, however.
There have been several examples of filesystem corruption without
power failures too, though that is a common cause. It can even happen
on a normal system shutdown if there is no synchronize operation to
the disk at shutdown time, depending on how much data is in the cache
and how long you have until the ATA chip is reset.
jeremy
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-23 13:23 ` Jan Engelhardt
2006-05-23 22:22 ` Nathan Scott
@ 2006-05-26 7:20 ` Helge Hafting
2006-05-26 8:51 ` Jan Engelhardt
1 sibling, 1 reply; 12+ messages in thread
From: Helge Hafting @ 2006-05-26 7:20 UTC (permalink / raw)
To: Jan Engelhardt; +Cc: Nathan Scott, linux-xfs, Linux Kernel Mailing List
Jan Engelhardt wrote:
> Well as mentioned, -o nobarrier solved it, and that's it. I do not actually
> need barriers (or an UPS, to poke on another thread), because power
> failures are rather rare in Germany.
Then there are mistakes like someone stepping on the power
cord, pulling out the wrong one, drilling holes in the wall,
there are kernel crashes, there is lightning,
there is the possibility that some
component inside the box just breaks.
Helge Hafting
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-26 7:20 ` Helge Hafting
@ 2006-05-26 8:51 ` Jan Engelhardt
2006-05-26 21:06 ` Bernd Eckenfels
0 siblings, 1 reply; 12+ messages in thread
From: Jan Engelhardt @ 2006-05-26 8:51 UTC (permalink / raw)
To: Helge Hafting; +Cc: Nathan Scott, linux-xfs, Linux Kernel Mailing List
>> Well as mentioned, -o nobarrier solved it, and that's it. I do not
>> actually need barriers (or an UPS, to poke on another thread), because
>> power failures are rather rare in Germany.
>
> Then there are mistakes like someone stepping on the power
> cord, pulling out the wrong one, drilling holes in the wall,
> there are kernel crashes, there is lightning,
> there is the possibility that some
> component inside the box just breaks.
>
A mathematician has fear of flying because there could possibly be a
terrorist bomb inside the plane. What does he do?
Taking an bomb with himself, because the probablity that there are two
bombs inside a plane is lower than one.
What I am wanting to say is that the factors you mentioned just do not happen
here, at this one - home - box.
Jan Engelhardt
--
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: XFS write speed drop
2006-05-26 8:51 ` Jan Engelhardt
@ 2006-05-26 21:06 ` Bernd Eckenfels
0 siblings, 0 replies; 12+ messages in thread
From: Bernd Eckenfels @ 2006-05-26 21:06 UTC (permalink / raw)
To: linux-kernel
Jan Engelhardt <jengelh@linux01.gwdg.de> wrote:
> What I am wanting to say is that the factors you mentioned just do not happen
> here, at this one - home - box.
They will as soon as you run the box in unsafe mode and you have important
data without backup :)
Gruss
Bernd
^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2006-05-26 21:06 UTC | newest]
Thread overview: 12+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2006-05-19 9:34 XFS write speed drop Jan Engelhardt
2006-05-22 0:53 ` Nathan Scott
2006-05-22 12:21 ` Jan Engelhardt
2006-05-22 22:43 ` Nathan Scott
2006-05-23 13:23 ` Jan Engelhardt
2006-05-23 22:22 ` Nathan Scott
2006-05-23 23:17 ` Nathan Scott
2006-05-24 7:16 ` Jeremy Higdon
2006-05-26 7:20 ` Helge Hafting
2006-05-26 8:51 ` Jan Engelhardt
2006-05-26 21:06 ` Bernd Eckenfels
2006-05-23 1:26 ` David Chinner
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox