From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from [140.186.70.92] (port=44171 helo=eggs.gnu.org) by lists.gnu.org with esmtp (Exim 4.43) id 1OBq4D-0001I3-Q4 for qemu-devel@nongnu.org; Tue, 11 May 2010 10:04:19 -0400 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.69) (envelope-from ) id 1OBq49-0001gd-DT for qemu-devel@nongnu.org; Tue, 11 May 2010 10:04:13 -0400 Received: from mx1.redhat.com ([209.132.183.28]:10704) by eggs.gnu.org with esmtp (Exim 4.69) (envelope-from ) id 1OBq42-0001ej-HV for qemu-devel@nongnu.org; Tue, 11 May 2010 10:04:05 -0400 Message-ID: <4BE963C9.9090308@redhat.com> Date: Tue, 11 May 2010 17:03:53 +0300 From: Avi Kivity MIME-Version: 1.0 References: <1271872408-22842-1-git-send-email-cam@cs.ualberta.ca> <1271872408-22842-3-git-send-email-cam@cs.ualberta.ca> <1271872408-22842-4-git-send-email-cam@cs.ualberta.ca> <1271872408-22842-5-git-send-email-cam@cs.ualberta.ca> <4BE7F517.5010707@redhat.com> <4BE82623.4000905@redhat.com> <4BE82877.1040408@codemonkey.ws> <4BE83B69.4040904@redhat.com> <4BE84172.9080305@codemonkey.ws> <4BE847CB.7050503@codemonkey.ws> <4BE90E6D.7070007@redhat.com> <4BE9572B.3010104@codemonkey.ws> In-Reply-To: <4BE9572B.3010104@codemonkey.ws> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Subject: [Qemu-devel] Re: [PATCH v5 4/5] Inter-VM shared memory PCI device List-Id: qemu-devel.nongnu.org List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Anthony Liguori Cc: Cam Macdonell , qemu-devel@nongnu.org, kvm@vger.kernel.org On 05/11/2010 04:10 PM, Anthony Liguori wrote: > On 05/11/2010 02:59 AM, Avi Kivity wrote: >>> (Replying again to list) >>> >>> What data structure would you use? For a lockless ring queue, you >>> can only support a single producer and consumer. To achieve >>> bidirectional communication in virtio, we always use two queues. >> >> >> You don't have to use a lockless ring queue. You can use locks >> (spinlocks without interrupt support, full mutexes with interrupts) >> and any data structure you like. Say a hash table + LRU for a shared >> cache. > > Yeah, the mailslot enables this. > > I think the question boils down to whether we can support transparent > peer connections and disconnections. I think that's important in > order to support transparent live migration. > > If you have two peers that are disconnected and then connect to each > other, there's simply no way to choose who's content gets preserved. > It's necessary to designate one peer as a master in order to break the > tie. The master is the shared memory area. It's a completely separate entity that is represented by the backing file (or shared memory server handing out the fd to mmap). It can exists independently of any guest. > > So this could simply involve an additional option to the shared memory > driver: role=master|peer. If role=master, when a new shared memory > segment is mapped, the contents of the BAR ram is memcpy()'d to the > shared memory segment. In either case, the contents of the shared > memory segment should be memcpy()'d to the BAR ram whenever the shared > memory segment is disconnected. I don't understand why we need separate BAR ram and shared memory. Have just shared memory, exposed by the BAR when connected. When the PCI card is disconnected from shared memory, the BAR should discard writes and return all 1s for reads. Having a temporary RAM area while disconnected doesn't serve a purpose (since it exists only for a short while) and increases the RAM load. > I believe role=master should be default because I think a relationship > of master/slave is going to be much more common than peering. What if you have N guests? What if the master disconnects? > >>> >>> If you're adding additional queues to support other levels of >>> communication, you can always use different areas of shared memory. >> >> You'll need O(n^2) shared memory areas (n=peer count), and it is a >> lot less flexible that real shared memory. Consider using threading >> where the only communication among threads is a pipe (erlang?) > > I can't think of a use of multiple peers via shared memory today with > virtualization. I know lots of master/slave uses of shared memory > though. I agree that it's useful to support from an academic > perspective but I don't believe it's going to be the common use. Large shared cache. That use case even survives live migration if you use lockless algorithms. -- error compiling committee.c: too many arguments to function