* [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx()
@ 2022-10-14 13:20 Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 1/2] virtio-net: fix bottom-half packet TX on asynchronous completion Laurent Vivier
` (2 more replies)
0 siblings, 3 replies; 5+ messages in thread
From: Laurent Vivier @ 2022-10-14 13:20 UTC (permalink / raw)
To: qemu-devel
Cc: Stefano Brivio, Michael S. Tsirkin, David Gibson, Jason Wang,
Laurent Vivier
When virtio_net_flush_tx() reaches the tx_burst value all the queue is
not flushed and nothing restart the timer or the bottom half function.
For BH, this is only missing in the virtio_net_tx_complete() function.
For the timer, the same fix is needed in virtio_net_tx_complete() but
it must be also managed in the TX timer function.
v2:
- fix also tx timer
Laurent Vivier (2):
virtio-net: fix bottom-half packet TX on asynchronous completion
virtio-net: fix TX timer with tx_burst
hw/net/virtio-net.c | 68 +++++++++++++++++++++++++++++++++++++--------
1 file changed, 56 insertions(+), 12 deletions(-)
--
2.37.3
^ permalink raw reply [flat|nested] 5+ messages in thread
* [PATCH v2 1/2] virtio-net: fix bottom-half packet TX on asynchronous completion
2022-10-14 13:20 [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Laurent Vivier
@ 2022-10-14 13:20 ` Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 2/2] virtio-net: fix TX timer with tx_burst Laurent Vivier
2022-10-26 20:23 ` [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Michael S. Tsirkin
2 siblings, 0 replies; 5+ messages in thread
From: Laurent Vivier @ 2022-10-14 13:20 UTC (permalink / raw)
To: qemu-devel
Cc: Stefano Brivio, Michael S. Tsirkin, David Gibson, Jason Wang,
Laurent Vivier, alex.williamson
When virtio-net is used with the socket netdev backend, the backend
can be busy and not able to collect new packets.
In this case, net_socket_receive() returns 0 and registers a poll function
to detect when the socket is ready again.
In virtio_net_tx_bh(), virtio_net_flush_tx() forwards the 0, the virtio
notifications are disabled and the function is not re-scheduled, waiting
for the backend to be ready.
When the socket netdev backend is again able to send packets, the poll
function re-starts to flush remaining packets. This is done by
calling virtio_net_tx_complete(). It re-enables notifications and calls
again virtio_net_flush_tx().
But it seems if virtio_net_flush_tx() reaches the tx_burst value all
the queue is not flushed and no new notification is sent to re-schedule
virtio_net_tx_bh(). Nothing re-start to flush the queue and remaining
packets are stuck in the queue.
To fix that, detect in virtio_net_tx_complete() if virtio_net_flush_tx()
has been stopped by tx_burst and if yes re-schedule the bottom half
function virtio_net_tx_bh() to flush the remaining packets.
This is what is done in virtio_net_tx_bh() when the virtio_net_flush_tx()
is synchronous, and completly by-passed when the operation needs to be
asynchronous.
Fixes: a697a334b3c4 ("virtio-net: Introduce a new bottom half packet TX")
Cc: alex.williamson@redhat.com
Signed-off-by: Laurent Vivier <lvivier@redhat.com>
Acked-by: Michael S. Tsirkin <mst@redhat.com>
---
hw/net/virtio-net.c | 13 ++++++++++++-
1 file changed, 12 insertions(+), 1 deletion(-)
diff --git a/hw/net/virtio-net.c b/hw/net/virtio-net.c
index e9f696b4cfeb..1fbf2f3e19a7 100644
--- a/hw/net/virtio-net.c
+++ b/hw/net/virtio-net.c
@@ -2526,6 +2526,7 @@ static void virtio_net_tx_complete(NetClientState *nc, ssize_t len)
VirtIONet *n = qemu_get_nic_opaque(nc);
VirtIONetQueue *q = virtio_net_get_subqueue(nc);
VirtIODevice *vdev = VIRTIO_DEVICE(n);
+ int ret;
virtqueue_push(q->tx_vq, q->async_tx.elem, 0);
virtio_notify(vdev, q->tx_vq);
@@ -2534,7 +2535,17 @@ static void virtio_net_tx_complete(NetClientState *nc, ssize_t len)
q->async_tx.elem = NULL;
virtio_queue_set_notification(q->tx_vq, 1);
- virtio_net_flush_tx(q);
+ ret = virtio_net_flush_tx(q);
+ if (q->tx_bh && ret >= n->tx_burst) {
+ /*
+ * the flush has been stopped by tx_burst
+ * we will not receive notification for the
+ * remainining part, so re-schedule
+ */
+ virtio_queue_set_notification(q->tx_vq, 0);
+ qemu_bh_schedule(q->tx_bh);
+ q->tx_waiting = 1;
+ }
}
/* TX */
--
2.37.3
^ permalink raw reply related [flat|nested] 5+ messages in thread
* [PATCH v2 2/2] virtio-net: fix TX timer with tx_burst
2022-10-14 13:20 [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 1/2] virtio-net: fix bottom-half packet TX on asynchronous completion Laurent Vivier
@ 2022-10-14 13:20 ` Laurent Vivier
2022-10-20 4:20 ` Jason Wang
2022-10-26 20:23 ` [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Michael S. Tsirkin
2 siblings, 1 reply; 5+ messages in thread
From: Laurent Vivier @ 2022-10-14 13:20 UTC (permalink / raw)
To: qemu-devel
Cc: Stefano Brivio, Michael S. Tsirkin, David Gibson, Jason Wang,
Laurent Vivier, alex.williamson
When virtio_net_flush_tx() reaches the tx_burst value all
the queue is not flushed and nothing restart the timer.
Fix that by doing for TX timer as we do for bottom half TX:
rearming the timer if we find any packet to send during the
virtio_net_flush_tx() call.
Fixes: e3f30488e5f8 ("virtio-net: Limit number of packets sent per TX flush")
Cc: alex.williamson@redhat.com
Signed-off-by: Laurent Vivier <lvivier@redhat.com>
---
hw/net/virtio-net.c | 59 +++++++++++++++++++++++++++++++++++----------
1 file changed, 46 insertions(+), 13 deletions(-)
diff --git a/hw/net/virtio-net.c b/hw/net/virtio-net.c
index 1fbf2f3e19a7..b4964b821021 100644
--- a/hw/net/virtio-net.c
+++ b/hw/net/virtio-net.c
@@ -2536,14 +2536,19 @@ static void virtio_net_tx_complete(NetClientState *nc, ssize_t len)
virtio_queue_set_notification(q->tx_vq, 1);
ret = virtio_net_flush_tx(q);
- if (q->tx_bh && ret >= n->tx_burst) {
+ if (ret >= n->tx_burst) {
/*
* the flush has been stopped by tx_burst
* we will not receive notification for the
* remainining part, so re-schedule
*/
virtio_queue_set_notification(q->tx_vq, 0);
- qemu_bh_schedule(q->tx_bh);
+ if (q->tx_bh) {
+ qemu_bh_schedule(q->tx_bh);
+ } else {
+ timer_mod(q->tx_timer,
+ qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
+ }
q->tx_waiting = 1;
}
}
@@ -2644,6 +2649,8 @@ drop:
return num_packets;
}
+static void virtio_net_tx_timer(void *opaque);
+
static void virtio_net_handle_tx_timer(VirtIODevice *vdev, VirtQueue *vq)
{
VirtIONet *n = VIRTIO_NET(vdev);
@@ -2661,18 +2668,17 @@ static void virtio_net_handle_tx_timer(VirtIODevice *vdev, VirtQueue *vq)
}
if (q->tx_waiting) {
- virtio_queue_set_notification(vq, 1);
+ /* We already have queued packets, immediately flush */
timer_del(q->tx_timer);
- q->tx_waiting = 0;
- if (virtio_net_flush_tx(q) == -EINVAL) {
- return;
- }
- } else {
- timer_mod(q->tx_timer,
- qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
- q->tx_waiting = 1;
- virtio_queue_set_notification(vq, 0);
+ virtio_net_tx_timer(q);
+ return;
}
+
+ /* re-arm timer to flush it (and more) on next tick */
+ timer_mod(q->tx_timer,
+ qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
+ q->tx_waiting = 1;
+ virtio_queue_set_notification(vq, 0);
}
static void virtio_net_handle_tx_bh(VirtIODevice *vdev, VirtQueue *vq)
@@ -2702,6 +2708,8 @@ static void virtio_net_tx_timer(void *opaque)
VirtIONetQueue *q = opaque;
VirtIONet *n = q->n;
VirtIODevice *vdev = VIRTIO_DEVICE(n);
+ int ret;
+
/* This happens when device was stopped but BH wasn't. */
if (!vdev->vm_running) {
/* Make sure tx waiting is set, so we'll run when restarted. */
@@ -2716,8 +2724,33 @@ static void virtio_net_tx_timer(void *opaque)
return;
}
+ ret = virtio_net_flush_tx(q);
+ if (ret == -EBUSY || ret == -EINVAL) {
+ return;
+ }
+ /*
+ * If we flush a full burst of packets, assume there are
+ * more coming and immediately rearm
+ */
+ if (ret >= n->tx_burst) {
+ q->tx_waiting = 1;
+ timer_mod(q->tx_timer,
+ qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
+ return;
+ }
+ /*
+ * If less than a full burst, re-enable notification and flush
+ * anything that may have come in while we weren't looking. If
+ * we find something, assume the guest is still active and rearm
+ */
virtio_queue_set_notification(q->tx_vq, 1);
- virtio_net_flush_tx(q);
+ ret = virtio_net_flush_tx(q);
+ if (ret > 0) {
+ virtio_queue_set_notification(q->tx_vq, 0);
+ q->tx_waiting = 1;
+ timer_mod(q->tx_timer,
+ qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
+ }
}
static void virtio_net_tx_bh(void *opaque)
--
2.37.3
^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [PATCH v2 2/2] virtio-net: fix TX timer with tx_burst
2022-10-14 13:20 ` [PATCH v2 2/2] virtio-net: fix TX timer with tx_burst Laurent Vivier
@ 2022-10-20 4:20 ` Jason Wang
0 siblings, 0 replies; 5+ messages in thread
From: Jason Wang @ 2022-10-20 4:20 UTC (permalink / raw)
To: Laurent Vivier
Cc: qemu-devel, Stefano Brivio, Michael S. Tsirkin, David Gibson,
alex.williamson
On Fri, Oct 14, 2022 at 9:20 PM Laurent Vivier <lvivier@redhat.com> wrote:
>
> When virtio_net_flush_tx() reaches the tx_burst value all
> the queue is not flushed and nothing restart the timer.
>
> Fix that by doing for TX timer as we do for bottom half TX:
> rearming the timer if we find any packet to send during the
> virtio_net_flush_tx() call.
>
> Fixes: e3f30488e5f8 ("virtio-net: Limit number of packets sent per TX flush")
> Cc: alex.williamson@redhat.com
> Signed-off-by: Laurent Vivier <lvivier@redhat.com>
> ---
> hw/net/virtio-net.c | 59 +++++++++++++++++++++++++++++++++++----------
> 1 file changed, 46 insertions(+), 13 deletions(-)
>
> diff --git a/hw/net/virtio-net.c b/hw/net/virtio-net.c
> index 1fbf2f3e19a7..b4964b821021 100644
> --- a/hw/net/virtio-net.c
> +++ b/hw/net/virtio-net.c
> @@ -2536,14 +2536,19 @@ static void virtio_net_tx_complete(NetClientState *nc, ssize_t len)
>
> virtio_queue_set_notification(q->tx_vq, 1);
> ret = virtio_net_flush_tx(q);
> - if (q->tx_bh && ret >= n->tx_burst) {
> + if (ret >= n->tx_burst) {
> /*
> * the flush has been stopped by tx_burst
> * we will not receive notification for the
> * remainining part, so re-schedule
> */
> virtio_queue_set_notification(q->tx_vq, 0);
> - qemu_bh_schedule(q->tx_bh);
> + if (q->tx_bh) {
> + qemu_bh_schedule(q->tx_bh);
> + } else {
> + timer_mod(q->tx_timer,
> + qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
> + }
> q->tx_waiting = 1;
> }
> }
> @@ -2644,6 +2649,8 @@ drop:
> return num_packets;
> }
>
> +static void virtio_net_tx_timer(void *opaque);
> +
> static void virtio_net_handle_tx_timer(VirtIODevice *vdev, VirtQueue *vq)
> {
> VirtIONet *n = VIRTIO_NET(vdev);
> @@ -2661,18 +2668,17 @@ static void virtio_net_handle_tx_timer(VirtIODevice *vdev, VirtQueue *vq)
> }
>
> if (q->tx_waiting) {
> - virtio_queue_set_notification(vq, 1);
> + /* We already have queued packets, immediately flush */
> timer_del(q->tx_timer);
> - q->tx_waiting = 0;
> - if (virtio_net_flush_tx(q) == -EINVAL) {
> - return;
> - }
> - } else {
> - timer_mod(q->tx_timer,
> - qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
> - q->tx_waiting = 1;
> - virtio_queue_set_notification(vq, 0);
> + virtio_net_tx_timer(q);
> + return;
> }
> +
> + /* re-arm timer to flush it (and more) on next tick */
> + timer_mod(q->tx_timer,
> + qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
> + q->tx_waiting = 1;
> + virtio_queue_set_notification(vq, 0);
> }
Nit: if we stick the above in the else, we can avoid a lot of changes.
Others look good.
Thanks
>
> static void virtio_net_handle_tx_bh(VirtIODevice *vdev, VirtQueue *vq)
> @@ -2702,6 +2708,8 @@ static void virtio_net_tx_timer(void *opaque)
> VirtIONetQueue *q = opaque;
> VirtIONet *n = q->n;
> VirtIODevice *vdev = VIRTIO_DEVICE(n);
> + int ret;
> +
> /* This happens when device was stopped but BH wasn't. */
> if (!vdev->vm_running) {
> /* Make sure tx waiting is set, so we'll run when restarted. */
> @@ -2716,8 +2724,33 @@ static void virtio_net_tx_timer(void *opaque)
> return;
> }
>
> + ret = virtio_net_flush_tx(q);
> + if (ret == -EBUSY || ret == -EINVAL) {
> + return;
> + }
> + /*
> + * If we flush a full burst of packets, assume there are
> + * more coming and immediately rearm
> + */
> + if (ret >= n->tx_burst) {
> + q->tx_waiting = 1;
> + timer_mod(q->tx_timer,
> + qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
> + return;
> + }
> + /*
> + * If less than a full burst, re-enable notification and flush
> + * anything that may have come in while we weren't looking. If
> + * we find something, assume the guest is still active and rearm
> + */
> virtio_queue_set_notification(q->tx_vq, 1);
> - virtio_net_flush_tx(q);
> + ret = virtio_net_flush_tx(q);
> + if (ret > 0) {
> + virtio_queue_set_notification(q->tx_vq, 0);
> + q->tx_waiting = 1;
> + timer_mod(q->tx_timer,
> + qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + n->tx_timeout);
> + }
> }
>
> static void virtio_net_tx_bh(void *opaque)
> --
> 2.37.3
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx()
2022-10-14 13:20 [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 1/2] virtio-net: fix bottom-half packet TX on asynchronous completion Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 2/2] virtio-net: fix TX timer with tx_burst Laurent Vivier
@ 2022-10-26 20:23 ` Michael S. Tsirkin
2 siblings, 0 replies; 5+ messages in thread
From: Michael S. Tsirkin @ 2022-10-26 20:23 UTC (permalink / raw)
To: Laurent Vivier; +Cc: qemu-devel, Stefano Brivio, David Gibson, Jason Wang
On Fri, Oct 14, 2022 at 03:20:02PM +0200, Laurent Vivier wrote:
> When virtio_net_flush_tx() reaches the tx_burst value all the queue is
> not flushed and nothing restart the timer or the bottom half function.
>
> For BH, this is only missing in the virtio_net_tx_complete() function.
> For the timer, the same fix is needed in virtio_net_tx_complete() but
> it must be also managed in the TX timer function.
>
> v2:
> - fix also tx timer
Jason's area, and he wants a small nit fixed.
Looks good to me overall:
Reviewed-by: Michael S. Tsirkin <mst@redhat.com>
> Laurent Vivier (2):
> virtio-net: fix bottom-half packet TX on asynchronous completion
> virtio-net: fix TX timer with tx_burst
>
> hw/net/virtio-net.c | 68 +++++++++++++++++++++++++++++++++++++--------
> 1 file changed, 56 insertions(+), 12 deletions(-)
>
> --
> 2.37.3
>
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2022-10-26 20:28 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2022-10-14 13:20 [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 1/2] virtio-net: fix bottom-half packet TX on asynchronous completion Laurent Vivier
2022-10-14 13:20 ` [PATCH v2 2/2] virtio-net: fix TX timer with tx_burst Laurent Vivier
2022-10-20 4:20 ` Jason Wang
2022-10-26 20:23 ` [PATCH v2 0/2] virtio-net: re-arm/re-schedule when tx_burst stops virtio_net_flush_tx() Michael S. Tsirkin
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).