* Write barriers and hardware RAID
@ 2009-07-17 10:35 J Pälve
2009-07-20 11:01 ` Michael Monnerie
0 siblings, 1 reply; 3+ messages in thread
From: J Pälve @ 2009-07-17 10:35 UTC (permalink / raw)
To: xfs
Hi,
I'm setting up a server to provide storage for a couple of VMware ESXi
servers. I'm using the latest stable Debian and I'm leaning towards using
NFS with XFS. However, I'm concerned about data integrity in the event of
power-out (we have UPS but they only last so long). Here are the specific
questions I have:
- The XFS FAQ states that with battery backup'd RAID hardware, both write
barriers and individual disk cache should be turned off. However, I'm
getting better benchmark results with both turned on. What I'm wondering
is, will write barriers work as intended when used with hardware RAID
controller (PERC 6/E)? Googling only turned up results relating to
software RAID.
- The XFS FAQ also states that virtualization products prevent write
barriers from working correctly. Is this still the case (specifically with
ESXi 3.5 and later) and is there anything that can be done about it? Does
VMFS somehow work around this, or is the problem then just "out of sight,
out of mind"?
Bregs,
JPa
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: Write barriers and hardware RAID
2009-07-17 10:35 Write barriers and hardware RAID J Pälve
@ 2009-07-20 11:01 ` Michael Monnerie
0 siblings, 0 replies; 3+ messages in thread
From: Michael Monnerie @ 2009-07-20 11:01 UTC (permalink / raw)
To: xfs
[-- Attachment #1.1: Type: text/plain, Size: 2704 bytes --]
I wrote that sections of the FAQ, so I should answer:
On Freitag 17 Juli 2009 J Pälve wrote:
> - The XFS FAQ states that with battery backup'd RAID hardware, both
> write barriers and individual disk cache should be turned off.
> However, I'm getting better benchmark results with both turned on.
I guess it's only the "hard disk cache" turned on leading to better
performance. But that is a very, very dangerous setup: If you use a RAID
with 16 hard disks, each having 32MB cache, on a power fail you can
loose up to 16*32 = 512MB of data, as on a power outage hard disks
simply drop their caches. And chances are *very* big that a significant
amount of filesystem metadata is in there, trashing your filesystem
badly. Never turn this on if you care about your data.
For write barriers, the performance should be a little bit lower if ON
instead OFF.
> What I'm wondering is, will write barriers work as intended when used
> with hardware RAID controller (PERC 6/E)? Googling only turned up
> results relating to software RAID.
No. RAID controllers simulate written data by telling the host that a
disk block has been written while it's only in the controller's cache.
The controller will write it later, when he has time. So basically
barriers only generate extra I/O there. This is valid if the controller
has writes "write back". If set to "write through", the RAID controller
simply does not cache writes, and directly writes them to disk, and only
afterwards tell the host that data has been written. This will drop your
write performance very significantly, on a server with much I/O you
don't want to use write through (aka write cache off).
> - The XFS FAQ also states that virtualization products prevent write
> barriers from working correctly. Is this still the case (specifically
> with ESXi 3.5 and later) and is there anything that can be done about
> it? Does VMFS somehow work around this, or is the problem then just
> "out of sight, out of mind"?
I found an entry for the ".vmx" config file:
scsi0:0.writeThrough = "TRUE"
That should do the desired "do not cache this disk", but I didn't test
it so far.
I wonder if someone knows of such a setting for XenServer?
If someone has a solution to "VM disk writes cached", I'd be happy to
hear how to do that.
mfg zmi
--
// Michael Monnerie, Ing.BSc ----- http://it-management.at
// Tel: 0660 / 415 65 31 .network.your.ideas.
// PGP Key: "curl -s http://zmi.at/zmi.asc | gpg --import"
// Fingerprint: AC19 F9D5 36ED CD8A EF38 500E CE14 91F7 1C12 09B4
// Keyserver: wwwkeys.eu.pgp.net Key-ID: 1C1209B4
[-- Attachment #1.2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 197 bytes --]
[-- Attachment #2: Type: text/plain, Size: 121 bytes --]
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
^ permalink raw reply [flat|nested] 3+ messages in thread
[parent not found: <mailman.323.1248105112.225101.xfs@oss.sgi.com>]
end of thread, other threads:[~2009-07-21 8:11 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-07-17 10:35 Write barriers and hardware RAID J Pälve
2009-07-20 11:01 ` Michael Monnerie
[not found] <mailman.323.1248105112.225101.xfs@oss.sgi.com>
[not found] ` <alpine.LRH.1.10.0907210918150.14559@pcuf.fi>
2009-07-21 8:12 ` Michael Monnerie
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox