public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
* Write barriers and hardware RAID
@ 2009-07-17 10:35 J Pälve
  2009-07-20 11:01 ` Michael Monnerie
  0 siblings, 1 reply; 3+ messages in thread
From: J Pälve @ 2009-07-17 10:35 UTC (permalink / raw)
  To: xfs

Hi,

I'm setting up a server to provide storage for a couple of VMware ESXi 
servers. I'm using the latest stable Debian and I'm leaning towards using 
NFS with XFS. However, I'm concerned about data integrity in the event of 
power-out (we have UPS but they only last so long). Here are the specific 
questions I have:

- The XFS FAQ states that with battery backup'd RAID hardware, both write 
barriers and individual disk cache should be turned off. However, I'm 
getting better benchmark results with both turned on. What I'm wondering 
is, will write barriers work as intended when used with hardware RAID 
controller (PERC 6/E)? Googling only turned up results relating to 
software RAID.

- The XFS FAQ also states that virtualization products prevent write 
barriers from working correctly. Is this still the case (specifically with 
ESXi 3.5 and later) and is there anything that can be done about it? Does 
VMFS somehow work around this, or is the problem then just "out of sight, 
out of mind"?

Bregs,
   JPa

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: Write barriers and hardware RAID
  2009-07-17 10:35 J Pälve
@ 2009-07-20 11:01 ` Michael Monnerie
  0 siblings, 0 replies; 3+ messages in thread
From: Michael Monnerie @ 2009-07-20 11:01 UTC (permalink / raw)
  To: xfs


[-- Attachment #1.1: Type: text/plain, Size: 2704 bytes --]

I wrote that sections of the FAQ, so I should answer:

On Freitag 17 Juli 2009 J Pälve wrote:
> - The XFS FAQ states that with battery backup'd RAID hardware, both
> write barriers and individual disk cache should be turned off.
> However, I'm getting better benchmark results with both turned on.

I guess it's only the "hard disk cache" turned on leading to better 
performance. But that is a very, very dangerous setup: If you use a RAID 
with 16 hard disks, each having 32MB cache, on a power fail you can 
loose up to 16*32 = 512MB of data, as on a power outage hard disks 
simply drop their caches. And chances are *very* big that a significant 
amount of filesystem metadata is in there, trashing your filesystem 
badly. Never turn this on if you care about your data.

For write barriers, the performance should be a little bit lower if ON 
instead OFF.

> What I'm wondering is, will write barriers work as intended when used
> with hardware RAID controller (PERC 6/E)? Googling only turned up
> results relating to software RAID.

No. RAID controllers simulate written data by telling the host that a 
disk block has been written while it's only in the controller's cache. 
The controller will write it later, when he has time. So basically 
barriers only generate extra I/O there. This is valid if the controller 
has writes "write back". If set to "write through", the RAID controller 
simply does not cache writes, and directly writes them to disk, and only 
afterwards tell the host that data has been written. This will drop your 
write performance very significantly, on a server with much I/O you 
don't want to use write through (aka write cache off).

> - The XFS FAQ also states that virtualization products prevent write
> barriers from working correctly. Is this still the case (specifically
> with ESXi 3.5 and later) and is there anything that can be done about
> it? Does VMFS somehow work around this, or is the problem then just
> "out of sight, out of mind"?

I found an entry for the ".vmx" config file:
scsi0:0.writeThrough = "TRUE"

That should do the desired "do not cache this disk", but I didn't test 
it so far.
I wonder if someone knows of such a setting for XenServer?

If someone has a solution to "VM disk writes cached", I'd be happy to 
hear how to do that.

mfg zmi
-- 
// Michael Monnerie, Ing.BSc    -----      http://it-management.at
// Tel: 0660 / 415 65 31                      .network.your.ideas.
// PGP Key:         "curl -s http://zmi.at/zmi.asc | gpg --import"
// Fingerprint: AC19 F9D5 36ED CD8A EF38  500E CE14 91F7 1C12 09B4
// Keyserver: wwwkeys.eu.pgp.net                  Key-ID: 1C1209B4


[-- Attachment #1.2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 197 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: Write barriers and hardware RAID
       [not found] ` <alpine.LRH.1.10.0907210918150.14559@pcuf.fi>
@ 2009-07-21  8:12   ` Michael Monnerie
  0 siblings, 0 replies; 3+ messages in thread
From: Michael Monnerie @ 2009-07-21  8:12 UTC (permalink / raw)
  To: J Pälve, xfs


[-- Attachment #1.1: Type: text/plain, Size: 2158 bytes --]

Please keep the discussion on the list also for others who search the 
same info later.

On Dienstag 21 Juli 2009 J Pälve wrote:
> Thank you for your clarification. I'm still wondering about this
> mention in the FAQ:
>
> "With a RAID controller with battery backed controller cache and
> cache in write back mode, you should turn off barriers - they are
> unnecessary in this case, and if the controller honors the cache
> flushes, it will be harmful to performance."
>
> If you issue a cache flush request to one of these controllers that
> honor them, will it only flush its own cache or also issue a cache
> flush request to the disks? In the latter case, wouldn't write
> barriers work correctly even with both controller and disk cache
> enabled?

If you use a RAID controller but then it allows cache flushes from the 
host to really flush it's cache, it's basically the same as setting the 
whole controller to "write through" mode (where he doesn't buffer 
anything). Your expensive RAID controller then acts like a dump switch 
and your only advantage is that you can connect several disks with 
parity. But your performance will be a mess. I can't believe anybody 
wanting performance will do that.
( We have some "up to 4 people" companies with HP servers where the 3-
disk RAID-5 is setup this way. As long as they only store word documents 
and such it's enough. We setup the whole controller write through then, 
battery backup is of course not needed. But disk write cache is always 
set to off, to prevent data loss on power fail.)

Back to your question: If there exists a hardware RAID controller 
honoring flushes, it should also flush the disks (I'd expect that). But 
I don't know any controller to really do that. If you do, please tell 
me.

mfg zmi
-- 
// Michael Monnerie, Ing.BSc    -----      http://it-management.at
// Tel: 0660 / 415 65 31                      .network.your.ideas.
// PGP Key:         "curl -s http://zmi.at/zmi.asc | gpg --import"
// Fingerprint: AC19 F9D5 36ED CD8A EF38  500E CE14 91F7 1C12 09B4
// Keyserver: wwwkeys.eu.pgp.net                  Key-ID: 1C1209B4


[-- Attachment #1.2: This is a digitally signed message part. --]
[-- Type: application/pgp-signature, Size: 197 bytes --]

[-- Attachment #2: Type: text/plain, Size: 121 bytes --]

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2009-07-21  8:11 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
     [not found] <mailman.323.1248105112.225101.xfs@oss.sgi.com>
     [not found] ` <alpine.LRH.1.10.0907210918150.14559@pcuf.fi>
2009-07-21  8:12   ` Write barriers and hardware RAID Michael Monnerie
2009-07-17 10:35 J Pälve
2009-07-20 11:01 ` Michael Monnerie

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox