From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from eggs.gnu.org ([2001:4830:134:3::10]:43360) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1WBdAN-0000Pd-38 for qemu-devel@nongnu.org; Thu, 06 Feb 2014 23:35:58 -0500 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1WBdAF-0006Tv-Pg for qemu-devel@nongnu.org; Thu, 06 Feb 2014 23:35:51 -0500 Received: from mail-pd0-f171.google.com ([209.85.192.171]:45847) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1WBdAF-0006Te-Jm for qemu-devel@nongnu.org; Thu, 06 Feb 2014 23:35:43 -0500 Received: by mail-pd0-f171.google.com with SMTP id g10so2646067pdj.2 for ; Thu, 06 Feb 2014 20:35:41 -0800 (PST) Received: from aik.ozlabs.ibm.com (ibmaus65.lnk.telstra.net. [165.228.126.9]) by mx.google.com with ESMTPSA id yz5sm22913114pac.9.2014.02.06.20.35.38 for (version=TLSv1 cipher=ECDHE-RSA-RC4-SHA bits=128/128); Thu, 06 Feb 2014 20:35:39 -0800 (PST) Message-ID: <52F46298.8020602@ozlabs.ru> Date: Fri, 07 Feb 2014 15:35:36 +1100 From: Alexey Kardashevskiy MIME-Version: 1.0 Content-Type: text/plain; charset=KOI8-R Content-Transfer-Encoding: 7bit Subject: [Qemu-devel] migration question: disk images on nfs server List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: "qemu-devel@nongnu.org" Hi! I have yet another problem with migration. Or NFS. There is one NFS server and 2 test POWER8 machines. There is a shared NFS folder on the server, mounted to both test hosts. There is an qcow2 image (abc.qcow2) in that shared folder. We start a guest with this abc.qcow2 on the test machine #1. And start another guest on the test machine #2 with "-incoming ..." and same abc.qcow2. Now we start migration. In most cases it goes fine. But if we put some load on machine #1, the destination guest sometime crashes. I blame out-of-sync NFS on the test machines. I looked a bit further in QEMU and could not find a spot where it would fflush(abc.qcow2) or close it or do any other sync so it is up to the host NFS mountpoint to decide when to sync and it definitely does not get a clue when to do this. I do not really understand why the abc.qcow2 image is still open, should not it be closed after migration succeeded? What do I miss here? Should we switch from NFS to GlusterFS (is it always syncronized)? Or if we want NFS, should we just boot our guests with "root=/dev/nfs ip=dhcp nfsroot=..." and avoid using disk images in network disks? Thanks! -- Alexey