From: "Hendrik ." <chasake@yahoo.com>
To: xfs@oss.sgi.com
Subject: Poor VMWare disk performance on XFS partition
Date: Sun, 23 Mar 2008 07:50:06 -0700 (PDT) [thread overview]
Message-ID: <876423.51989.qm@web52006.mail.re2.yahoo.com> (raw)
I've been converting some of my drives from EXT3 to
XFS a while ago. Now I notice poor disk performance
when using XFS as underlying filesystem for a VMware
virtual drive. I did some experiments and it really
seems to be the XFS filesystem 'trashing' the speed of
a VMware Windows XP guest.
The first thing that I noticed is that when I shut
down the virtual machine it takes a very long time
after the machine seems to be shut down until the
VMware window becomes responsive again. In the mean
time there is heavy disk I/O. I found out that VMware
seems to write some kind of memory map on the host
hard disk which get heavily fragmented. The removal of
this file is probably very I/O intensive which causes
the delay. This is very annoying but not really a
problem as it only happens when a virtual machine is
shut down.
But there seems to be another problem when running the
guest operating system itself. I made two exact copies
of a Windows XP virtual machine on two hard disks of
the same type, size and brand. The first hard disk had
a XFS partition to host the virtual machine, the
seconds harddisk was formatted as EXT3. The XFS
partition has no fragmentation at all thus all files
only consisted of 1 extent. The EXT3 files were a bit
fragmented but this was only marginal (some larger
disk image files consisted of 17 extents where 16 was
optimal, reported by 'filefrag').
Then I ran the virtual machines one by one and started
a defragmentation program to cause a lot of I/O on the
guest operating system. Defragmentation of the XP host
running from the EXT3 partition took only 2m36 but the
exact same guest on the XFS partition took 10m5 to
complete. On the EXT3 partition hardly any noise was
heard from the drive heads as if the host operating
system was caching and delaying the I/O operations. On
the XFS host however a lot of noise was heard as if
the harddisk was trashing heavily.
The machine I ran the tests on had the following
specifications:
- Dual core AMD Athlon64 5200+
- 2 GB memory
- 2x 80 GB Samsung IDE harddisk
- Ubuntu Gutsy Gibbon (7.10) 64-bit
- Standard stock Ubuntu kernel
- VMware workstation 6.0.2 build-59824
The guest was a bare Windows XP SP-2 installation,
running O&O defrag 10.
Can anyone give me some information what might be
causing this massive slowdown?
Regards,
Hendrik van den Boogaard
____________________________________________________________________________________
Looking for last minute shopping deals?
Find them fast with Yahoo! Search. http://tools.search.yahoo.com/newsearch/category.php?category=shopping
next reply other threads:[~2008-03-23 14:50 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2008-03-23 14:50 Hendrik . [this message]
2008-03-23 20:42 ` Poor VMWare disk performance on XFS partition Eric Sandeen
2008-03-24 4:02 ` Andi Kleen
2008-03-24 0:41 ` Jan Derfinak
2008-03-24 14:00 ` Russell Cattelan
2008-03-24 15:24 ` Peter Grandi
2008-03-24 18:56 ` Jan Derfinak
2008-03-24 19:12 ` Josef 'Jeff' Sipek
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=876423.51989.qm@web52006.mail.re2.yahoo.com \
--to=chasake@yahoo.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox