From: marcandre.lureau@redhat.com
To: qemu-devel@nongnu.org
Cc: mukawa@igel.co.jp, yuanhan.liu@linux.intel.com,
victork@redhat.com, jonshin@cisco.com, mst@redhat.com,
"Marc-André Lureau" <marcandre.lureau@redhat.com>
Subject: [Qemu-devel] [PATCH v5 12/31] vhost: use error_report() instead of fprintf(stderr, ...)
Date: Thu, 21 Jul 2016 12:57:31 +0400 [thread overview]
Message-ID: <20160721085750.30008-13-marcandre.lureau@redhat.com> (raw)
In-Reply-To: <20160721085750.30008-1-marcandre.lureau@redhat.com>
From: Marc-André Lureau <marcandre.lureau@redhat.com>
Let's use qemu proper error reporting API, this ensures the error is
reported at the right place (stderr or monitor), with a conventional
format.
Signed-off-by: Marc-André Lureau <marcandre.lureau@redhat.com>
---
hw/virtio/vhost.c | 19 +++++++++----------
1 file changed, 9 insertions(+), 10 deletions(-)
diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c
index a671ac8..09428bc 100644
--- a/hw/virtio/vhost.c
+++ b/hw/virtio/vhost.c
@@ -439,11 +439,11 @@ static int vhost_verify_ring_mappings(struct vhost_dev *dev,
l = vq->ring_size;
p = cpu_physical_memory_map(vq->ring_phys, &l, 1);
if (!p || l != vq->ring_size) {
- fprintf(stderr, "Unable to map ring buffer for ring %d\n", i);
+ error_report("Unable to map ring buffer for ring %d", i);
r = -ENOMEM;
}
if (p != vq->ring) {
- fprintf(stderr, "Ring buffer relocated for ring %d\n", i);
+ error_report("Ring buffer relocated for ring %d", i);
r = -EBUSY;
}
cpu_physical_memory_unmap(p, l, 0, 0);
@@ -1050,8 +1050,8 @@ int vhost_dev_init(struct vhost_dev *hdev, void *opaque,
}
if (used_memslots > hdev->vhost_ops->vhost_backend_memslots_limit(hdev)) {
- fprintf(stderr, "vhost backend memory slots limit is less"
- " than current number of present memory slots\n");
+ error_report("vhost backend memory slots limit is less"
+ " than current number of present memory slots");
r = -1;
goto fail;
}
@@ -1174,8 +1174,9 @@ int vhost_dev_enable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev)
VirtioBusState *vbus = VIRTIO_BUS(qbus);
VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus);
int i, r, e;
+
if (!k->ioeventfd_started) {
- fprintf(stderr, "binding does not support host notifiers\n");
+ error_report("binding does not support host notifiers");
r = -ENOSYS;
goto fail;
}
@@ -1184,7 +1185,7 @@ int vhost_dev_enable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev)
r = virtio_bus_set_host_notifier(VIRTIO_BUS(qbus), hdev->vq_index + i,
true);
if (r < 0) {
- fprintf(stderr, "vhost VQ %d notifier binding failed: %d\n", i, -r);
+ error_report("vhost VQ %d notifier binding failed: %d", i, -r);
goto fail_vq;
}
}
@@ -1195,8 +1196,7 @@ fail_vq:
e = virtio_bus_set_host_notifier(VIRTIO_BUS(qbus), hdev->vq_index + i,
false);
if (e < 0) {
- fprintf(stderr, "vhost VQ %d notifier cleanup error: %d\n", i, -r);
- fflush(stderr);
+ error_report("vhost VQ %d notifier cleanup error: %d", i, -r);
}
assert (e >= 0);
}
@@ -1218,8 +1218,7 @@ void vhost_dev_disable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev)
r = virtio_bus_set_host_notifier(VIRTIO_BUS(qbus), hdev->vq_index + i,
false);
if (r < 0) {
- fprintf(stderr, "vhost VQ %d notifier cleanup failed: %d\n", i, -r);
- fflush(stderr);
+ error_report("vhost VQ %d notifier cleanup failed: %d", i, -r);
}
assert (r >= 0);
}
--
2.9.0
next prev parent reply other threads:[~2016-07-21 8:58 UTC|newest]
Thread overview: 39+ messages / expand[flat|nested] mbox.gz Atom feed top
2016-07-21 8:57 [Qemu-devel] [PATCH v5 00/31] vhost-user reconnect fixes marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 01/31] misc: indentation marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 02/31] vhost-user: minor simplification marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 03/31] vhost: don't assume opaque is a fd, use backend cleanup marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 04/31] vhost: make vhost_log_put() idempotent marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 05/31] vhost: assert the log was cleaned up marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 06/31] vhost: fix cleanup on not fully initialized device marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 07/31] vhost: make vhost_dev_cleanup() idempotent marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 08/31] vhost-net: always call vhost_dev_cleanup() on failure marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 09/31] vhost: fix calling vhost_dev_cleanup() after vhost_dev_init() marcandre.lureau
2016-07-25 12:33 ` [Qemu-devel] [v5, " Ilya Maximets
2016-07-25 12:45 ` Marc-André Lureau
2016-07-25 12:52 ` Ilya Maximets
2016-07-25 13:05 ` Marc-André Lureau
2016-07-25 13:14 ` Ilya Maximets
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 10/31] vhost: do not assert() on vhost_ops failure marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 11/31] vhost: add missing VHOST_OPS_DEBUG marcandre.lureau
2016-07-21 8:57 ` marcandre.lureau [this message]
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 13/31] qemu-char: fix qemu_chr_fe_set_msgfds() crash when disconnected marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 14/31] vhost-user: call set_msgfds unconditionally marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 15/31] vhost-user: check qemu_chr_fe_set_msgfds() return value marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 16/31] vhost-user: check vhost_user_{read, write}() " marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 17/31] vhost-user: keep vhost_net after a disconnection marcandre.lureau
2016-07-25 12:48 ` [Qemu-devel] [v5, " Ilya Maximets
2016-07-25 13:09 ` Marc-André Lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 18/31] vhost-user: add get_vhost_net() assertions marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 19/31] Revert "vhost-net: do not crash if backend is not present" marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 20/31] vhost-net: vhost_migration_done is vhost-user specific marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 21/31] vhost: add assert() to check runtime behaviour marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 22/31] char: add chr_wait_connected callback marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 23/31] char: add and use tcp_chr_wait_connected marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 24/31] vhost-user: wait until backend init is completed marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 25/31] tests: plug some leaks in virtio-net-test marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 26/31] tests: fix vhost-user-test leak marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 27/31] tests: add /vhost-user/connect-fail test marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 28/31] tests: add a simple /vhost-user/multiqueue test marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 29/31] vhost-user: add error report in vhost_user_write() marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 30/31] vhost: add vhost_net_set_backend() marcandre.lureau
2016-07-21 8:57 ` [Qemu-devel] [PATCH v5 31/31] RFC: vhost: do not update last avail idx on get_vring_base() failure marcandre.lureau
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20160721085750.30008-13-marcandre.lureau@redhat.com \
--to=marcandre.lureau@redhat.com \
--cc=jonshin@cisco.com \
--cc=mst@redhat.com \
--cc=mukawa@igel.co.jp \
--cc=qemu-devel@nongnu.org \
--cc=victork@redhat.com \
--cc=yuanhan.liu@linux.intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).