From: Jason Wang <jasowang@redhat.com>
To: "Michael S. Tsirkin" <mst@redhat.com>, linux-kernel@vger.kernel.org
Cc: netdev@vger.kernel.org,
virtualization@lists.linux-foundation.org,
maxime.coquelin@redhat.com, wexu@redhat.com,
"David S. Miller" <davem@davemloft.net>
Subject: Re: [PATCH RFC 1/2] virtio-net: bql support
Date: Thu, 6 Dec 2018 16:31:24 +0800 [thread overview]
Message-ID: <a7c57a15-825d-b7fa-8f6d-790cb285bdcc@redhat.com> (raw)
In-Reply-To: <21384cb5-99a6-7431-1039-b356521e1bc3@redhat.com>
[-- Attachment #1: Type: text/plain, Size: 5396 bytes --]
On 2018/12/6 下午4:17, Jason Wang wrote:
>
> On 2018/12/6 上午6:54, Michael S. Tsirkin wrote:
>> When use_napi is set, let's enable BQLs. Note: some of the issues are
>> similar to wifi. It's worth considering whether something similar to
>> commit 36148c2bbfbe ("mac80211: Adjust TSQ pacing shift") might be
>> benefitial.
>
>
> I've played a similar patch several days before. The tricky part is
> the mode switching between napi and no napi. We should make sure when
> the packet is sent and trakced by BQL, it should be consumed by BQL
> as well. I did it by tracking it through skb->cb. And deal with the
> freeze by reset the BQL status. Patch attached.
Add the patch.
Thanks
>
> But when testing with vhost-net, I don't very a stable performance, it
> was probably because we batch the used ring updating so tx interrupt
> may come randomly. We probably need to implement time bounded
> coalescing mechanism which could be configured from userspace.
>
> Btw, maybe it's time just enable napi TX by default. I get ~10% TCP_RR
> regression on machine without APICv, (haven't found time to test APICv
> machine). But consider it was for correctness, I think it's
> acceptable? Then we can do optimization on top?
>
>
> Thanks
>
>
>> Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
>> ---
>> drivers/net/virtio_net.c | 27 +++++++++++++++++++--------
>> 1 file changed, 19 insertions(+), 8 deletions(-)
>>
>> diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
>> index cecfd77c9f3c..b657bde6b94b 100644
>> --- a/drivers/net/virtio_net.c
>> +++ b/drivers/net/virtio_net.c
>> @@ -1325,7 +1325,8 @@ static int virtnet_receive(struct receive_queue
>> *rq, int budget,
>> return stats.packets;
>> }
>> -static void free_old_xmit_skbs(struct send_queue *sq)
>> +static void free_old_xmit_skbs(struct send_queue *sq, struct
>> netdev_queue *txq,
>> + bool use_napi)
>> {
>> struct sk_buff *skb;
>> unsigned int len;
>> @@ -1347,6 +1348,9 @@ static void free_old_xmit_skbs(struct
>> send_queue *sq)
>> if (!packets)
>> return;
>> + if (use_napi)
>> + netdev_tx_completed_queue(txq, packets, bytes);
>> +
>> u64_stats_update_begin(&sq->stats.syncp);
>> sq->stats.bytes += bytes;
>> sq->stats.packets += packets;
>> @@ -1364,7 +1368,7 @@ static void virtnet_poll_cleantx(struct
>> receive_queue *rq)
>> return;
>> if (__netif_tx_trylock(txq)) {
>> - free_old_xmit_skbs(sq);
>> + free_old_xmit_skbs(sq, txq, true);
>> __netif_tx_unlock(txq);
>> }
>> @@ -1440,7 +1444,7 @@ static int virtnet_poll_tx(struct napi_struct
>> *napi, int budget)
>> struct netdev_queue *txq = netdev_get_tx_queue(vi->dev,
>> vq2txq(sq->vq));
>> __netif_tx_lock(txq, raw_smp_processor_id());
>> - free_old_xmit_skbs(sq);
>> + free_old_xmit_skbs(sq, txq, true);
>> __netif_tx_unlock(txq);
>> virtqueue_napi_complete(napi, sq->vq, 0);
>> @@ -1505,13 +1509,15 @@ static netdev_tx_t start_xmit(struct sk_buff
>> *skb, struct net_device *dev)
>> struct send_queue *sq = &vi->sq[qnum];
>> int err;
>> struct netdev_queue *txq = netdev_get_tx_queue(dev, qnum);
>> - bool kick = !skb->xmit_more;
>> + bool more = skb->xmit_more;
>> bool use_napi = sq->napi.weight;
>> + unsigned int bytes = skb->len;
>> + bool kick;
>> /* Free up any pending old buffers before queueing new ones. */
>> - free_old_xmit_skbs(sq);
>> + free_old_xmit_skbs(sq, txq, use_napi);
>> - if (use_napi && kick)
>> + if (use_napi && !more)
>> virtqueue_enable_cb_delayed(sq->vq);
>> /* timestamp packet in software */
>> @@ -1552,7 +1558,7 @@ static netdev_tx_t start_xmit(struct sk_buff
>> *skb, struct net_device *dev)
>> if (!use_napi &&
>> unlikely(!virtqueue_enable_cb_delayed(sq->vq))) {
>> /* More just got used, free them then recheck. */
>> - free_old_xmit_skbs(sq);
>> + free_old_xmit_skbs(sq, txq, false);
>> if (sq->vq->num_free >= 2+MAX_SKB_FRAGS) {
>> netif_start_subqueue(dev, qnum);
>> virtqueue_disable_cb(sq->vq);
>> @@ -1560,7 +1566,12 @@ static netdev_tx_t start_xmit(struct sk_buff
>> *skb, struct net_device *dev)
>> }
>> }
>> - if (kick || netif_xmit_stopped(txq)) {
>> + if (use_napi)
>> + kick = __netdev_tx_sent_queue(txq, bytes, more);
>> + else
>> + kick = !more || netif_xmit_stopped(txq);
>> +
>> + if (kick) {
>> if (virtqueue_kick_prepare(sq->vq) &&
>> virtqueue_notify(sq->vq)) {
>> u64_stats_update_begin(&sq->stats.syncp);
>> sq->stats.kicks++;
> _______________________________________________
> Virtualization mailing list
> Virtualization@lists.linux-foundation.org
> https://lists.linuxfoundation.org/mailman/listinfo/virtualization
[-- Attachment #2: 0001-virtio-net-byte-queue-limit-support.patch --]
[-- Type: text/x-patch, Size: 4778 bytes --]
>From f1c27543dc412778e682b63addbb0a471afc5153 Mon Sep 17 00:00:00 2001
From: Jason Wang <jasowang@redhat.com>
Date: Tue, 20 Nov 2018 14:25:30 +0800
Subject: [PATCH] virtio-net: byte queue limit support
Signed-off-by: Jason Wang <jasowang@redhat.com>
---
drivers/net/virtio_net.c | 46 ++++++++++++++++++++++++++++++++++++++--------
1 file changed, 38 insertions(+), 8 deletions(-)
diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
index 47979fc..8712c11 100644
--- a/drivers/net/virtio_net.c
+++ b/drivers/net/virtio_net.c
@@ -279,6 +279,14 @@ static inline struct virtio_net_hdr_mrg_rxbuf *skb_vnet_hdr(struct sk_buff *skb)
return (struct virtio_net_hdr_mrg_rxbuf *)skb->cb;
}
+
+static inline int *skb_cb_bql(struct sk_buff *skb)
+{
+ BUILD_BUG_ON(sizeof(struct virtio_net_hdr_mrg_rxbuf) +
+ sizeof(int) > sizeof(skb->cb));
+ return (int *)(skb->cb + sizeof(struct virtio_net_hdr_mrg_rxbuf));
+}
+
/*
* private is used to chain pages for big packets, put the whole
* most recent used list in the beginning for reuse
@@ -1325,12 +1333,14 @@ static int virtnet_receive(struct receive_queue *rq, int budget,
return stats.packets;
}
-static void free_old_xmit_skbs(struct send_queue *sq)
+static void free_old_xmit_skbs(struct send_queue *sq,
+ struct netdev_queue *txq)
{
struct sk_buff *skb;
unsigned int len;
- unsigned int packets = 0;
- unsigned int bytes = 0;
+ unsigned int packets = 0, bql_packets = 0;
+ unsigned int bytes = 0, bql_bytes = 0;
+ int *bql;
while ((skb = virtqueue_get_buf(sq->vq, &len)) != NULL) {
pr_debug("Sent skb %p\n", skb);
@@ -1338,6 +1348,12 @@ static void free_old_xmit_skbs(struct send_queue *sq)
bytes += skb->len;
packets++;
+ bql = skb_cb_bql(skb);
+ if (*bql) {
+ bql_packets ++;
+ bql_bytes += skb->len;
+ }
+
dev_consume_skb_any(skb);
}
@@ -1351,6 +1367,8 @@ static void free_old_xmit_skbs(struct send_queue *sq)
sq->stats.bytes += bytes;
sq->stats.packets += packets;
u64_stats_update_end(&sq->stats.syncp);
+
+ netdev_tx_completed_queue(txq, bql_packets, bql_bytes);
}
static void virtnet_poll_cleantx(struct receive_queue *rq)
@@ -1364,7 +1382,7 @@ static void virtnet_poll_cleantx(struct receive_queue *rq)
return;
if (__netif_tx_trylock(txq)) {
- free_old_xmit_skbs(sq);
+ free_old_xmit_skbs(sq, txq);
__netif_tx_unlock(txq);
}
@@ -1440,7 +1458,7 @@ static int virtnet_poll_tx(struct napi_struct *napi, int budget)
struct netdev_queue *txq = netdev_get_tx_queue(vi->dev, vq2txq(sq->vq));
__netif_tx_lock(txq, raw_smp_processor_id());
- free_old_xmit_skbs(sq);
+ free_old_xmit_skbs(sq, txq);
__netif_tx_unlock(txq);
virtqueue_napi_complete(napi, sq->vq, 0);
@@ -1459,6 +1477,7 @@ static int xmit_skb(struct send_queue *sq, struct sk_buff *skb)
int num_sg;
unsigned hdr_len = vi->hdr_len;
bool can_push;
+ int *bql = skb_cb_bql(skb);
pr_debug("%s: xmit %p %pM\n", vi->dev->name, skb, dest);
@@ -1495,6 +1514,8 @@ static int xmit_skb(struct send_queue *sq, struct sk_buff *skb)
return num_sg;
num_sg++;
}
+
+ *bql = sq->napi.weight ? 1 : 0;
return virtqueue_add_outbuf(sq->vq, sq->sg, num_sg, skb, GFP_ATOMIC);
}
@@ -1509,7 +1530,7 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
bool use_napi = sq->napi.weight;
/* Free up any pending old buffers before queueing new ones. */
- free_old_xmit_skbs(sq);
+ free_old_xmit_skbs(sq, txq);
if (use_napi && kick)
virtqueue_enable_cb_delayed(sq->vq);
@@ -1537,6 +1558,9 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
nf_reset(skb);
}
+ if (use_napi)
+ netdev_tx_sent_queue(txq, skb->len);
+
/* If running out of space, stop queue to avoid getting packets that we
* are then unable to transmit.
* An alternative would be to force queuing layer to requeue the skb by
@@ -1552,7 +1576,7 @@ static netdev_tx_t start_xmit(struct sk_buff *skb, struct net_device *dev)
if (!use_napi &&
unlikely(!virtqueue_enable_cb_delayed(sq->vq))) {
/* More just got used, free them then recheck. */
- free_old_xmit_skbs(sq);
+ free_old_xmit_skbs(sq, txq);
if (sq->vq->num_free >= 2+MAX_SKB_FRAGS) {
netif_start_subqueue(dev, qnum);
virtqueue_disable_cb(sq->vq);
@@ -2275,8 +2299,14 @@ static void virtnet_freeze_down(struct virtio_device *vdev)
if (netif_running(vi->dev)) {
for (i = 0; i < vi->max_queue_pairs; i++) {
+ struct send_queue *sq = &vi->sq[i];
+ struct netdev_queue *txq =
+ netdev_get_tx_queue(vi->dev, i);
+
napi_disable(&vi->rq[i].napi);
- virtnet_napi_tx_disable(&vi->sq[i].napi);
+ virtnet_napi_tx_disable(&sq->napi);
+ if (sq->napi.weight)
+ netdev_tx_reset_queue(txq);
}
}
}
--
1.8.3.1
next prev parent reply other threads:[~2018-12-06 8:31 UTC|newest]
Thread overview: 25+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-12-05 22:54 [PATCH RFC 0/2] virtio-net: interrupt related improvements Michael S. Tsirkin
2018-12-05 22:54 ` [PATCH RFC 1/2] virtio-net: bql support Michael S. Tsirkin
2018-12-06 8:17 ` Jason Wang
2018-12-06 8:31 ` Jason Wang [this message]
2018-12-26 15:15 ` Michael S. Tsirkin
2018-12-27 9:56 ` Jason Wang
2018-12-26 15:19 ` Michael S. Tsirkin
2018-12-27 10:00 ` Jason Wang
2018-12-30 18:45 ` Michael S. Tsirkin
2019-01-02 3:28 ` Jason Wang
2019-01-02 13:59 ` Michael S. Tsirkin
2019-01-07 2:14 ` Jason Wang
2019-01-07 3:17 ` Michael S. Tsirkin
2019-01-07 3:51 ` Jason Wang
2019-01-07 4:01 ` Michael S. Tsirkin
2019-01-07 6:31 ` Jason Wang
2019-01-07 14:19 ` Michael S. Tsirkin
2019-01-08 10:06 ` Jason Wang
2018-12-26 15:22 ` Michael S. Tsirkin
2018-12-27 10:04 ` Jason Wang
2018-12-30 18:48 ` Michael S. Tsirkin
2019-01-02 3:30 ` Jason Wang
2019-01-02 13:54 ` Michael S. Tsirkin
2019-01-17 13:09 ` Jason Wang
2018-12-05 22:54 ` [PATCH RFC 2/2] virtio_net: bulk free tx skbs Michael S. Tsirkin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=a7c57a15-825d-b7fa-8f6d-790cb285bdcc@redhat.com \
--to=jasowang@redhat.com \
--cc=davem@davemloft.net \
--cc=linux-kernel@vger.kernel.org \
--cc=maxime.coquelin@redhat.com \
--cc=mst@redhat.com \
--cc=netdev@vger.kernel.org \
--cc=virtualization@lists.linux-foundation.org \
--cc=wexu@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).