From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from mailman by lists.gnu.org with tmda-scanned (Exim 4.43) id 1NmsEm-00021f-Me for qemu-devel@nongnu.org; Wed, 03 Mar 2010 12:19:56 -0500 Received: from [199.232.76.173] (port=36544 helo=monty-python.gnu.org) by lists.gnu.org with esmtp (Exim 4.43) id 1NmsEm-00020v-5h for qemu-devel@nongnu.org; Wed, 03 Mar 2010 12:19:56 -0500 Received: from Debian-exim by monty-python.gnu.org with spam-scanned (Exim 4.60) (envelope-from ) id 1NmsEk-00056N-Ok for qemu-devel@nongnu.org; Wed, 03 Mar 2010 12:19:56 -0500 Received: from mx1.redhat.com ([209.132.183.28]:25718) by monty-python.gnu.org with esmtp (Exim 4.60) (envelope-from ) id 1NmsEk-00056H-An for qemu-devel@nongnu.org; Wed, 03 Mar 2010 12:19:54 -0500 Date: Wed, 3 Mar 2010 19:16:32 +0200 From: "Michael S. Tsirkin" Message-ID: <977ef3684e3da8a68f8d2e5643c7cc6195b310d8.1267636215.git.mst@redhat.com> References: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Subject: [Qemu-devel] [PATCHv4 08/12] virtio-pci: fill in notifier support List-Id: qemu-devel.nongnu.org List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Anthony Liguori , qemu-devel@nongnu.org Cc: amit.shah@redhat.com, kraxel@redhat.com, quintela@redhat.com Support host/guest notifiers in virtio-pci. The last one only with kvm, that's okay because vhost relies on kvm anyway. Note on kvm usage: kvm ioeventfd API is implemented on non-kvm systems as well, this is the reason we don't need if (kvm_enabled()) around it. Signed-off-by: Michael S. Tsirkin --- hw/virtio-pci.c | 62 +++++++++++++++++++++++++++++++++++++++++++++++++++++++ 1 files changed, 62 insertions(+), 0 deletions(-) diff --git a/hw/virtio-pci.c b/hw/virtio-pci.c index fd9759a..596f056 100644 --- a/hw/virtio-pci.c +++ b/hw/virtio-pci.c @@ -24,6 +24,7 @@ #include "net.h" #include "block_int.h" #include "loader.h" +#include "kvm.h" /* from Linux's linux/virtio_pci.h */ @@ -392,6 +393,65 @@ static unsigned virtio_pci_get_features(void *opaque) return proxy->host_features; } +static void virtio_pci_guest_notifier_read(void *opaque) +{ + VirtQueue *vq = opaque; + EventNotifier *n = virtio_queue_guest_notifier(vq); + if (event_notifier_test_and_clear(n)) { + virtio_irq(vq); + } +} + +static int virtio_pci_guest_notifier(void *opaque, int n, bool assign) +{ + VirtIOPCIProxy *proxy = opaque; + VirtQueue *vq = virtio_queue(proxy->vdev, n); + EventNotifier *notifier = virtio_queue_guest_notifier(vq); + + if (assign) { + int r = event_notifier_init(notifier, 0); + if (r < 0) + return r; + qemu_set_fd_handler(event_notifier_get_fd(notifier), + virtio_pci_guest_notifier_read, NULL, vq); + } else { + qemu_set_fd_handler(event_notifier_get_fd(notifier), + NULL, NULL, NULL); + event_notifier_cleanup(notifier); + } + + return 0; +} + +static int virtio_pci_host_notifier(void *opaque, int n, bool assign) +{ + VirtIOPCIProxy *proxy = opaque; + VirtQueue *vq = virtio_queue(proxy->vdev, n); + EventNotifier *notifier = virtio_queue_host_notifier(vq); + int r; + if (assign) { + r = event_notifier_init(notifier, 1); + if (r < 0) { + return r; + } + r = kvm_set_ioeventfd_pio_word(event_notifier_get_fd(notifier), + proxy->addr + VIRTIO_PCI_QUEUE_NOTIFY, + n, assign); + if (r < 0) { + event_notifier_cleanup(notifier); + } + } else { + r = kvm_set_ioeventfd_pio_word(event_notifier_get_fd(notifier), + proxy->addr + VIRTIO_PCI_QUEUE_NOTIFY, + n, assign); + if (r < 0) { + return r; + } + event_notifier_cleanup(notifier); + } + return r; +} + static const VirtIOBindings virtio_pci_bindings = { .notify = virtio_pci_notify, .save_config = virtio_pci_save_config, @@ -399,6 +459,8 @@ static const VirtIOBindings virtio_pci_bindings = { .save_queue = virtio_pci_save_queue, .load_queue = virtio_pci_load_queue, .get_features = virtio_pci_get_features, + .host_notifier = virtio_pci_host_notifier, + .guest_notifier = virtio_pci_guest_notifier, }; static void virtio_init_pci(VirtIOPCIProxy *proxy, VirtIODevice *vdev, -- 1.7.0.18.g0d53a5