From: Zoltan Kiss <zoltan.kiss@citrix.com>
To: Paul Durrant <Paul.Durrant@citrix.com>,
"xen-devel@lists.xenproject.org" <xen-devel@lists.xenproject.org>,
Ian Campbell <Ian.Campbell@citrix.com>,
"Wei Liu" <wei.liu2@citrix.com>,
"linux@eikelenboom.it" <linux@eikelenboom.it>
Cc: "netdev@vger.kernel.org" <netdev@vger.kernel.org>,
David Vrabel <david.vrabel@citrix.com>,
"davem@davemloft.net" <davem@davemloft.net>
Subject: Re: [PATCH net] xen-netback: Fix slot estimation
Date: Tue, 3 Jun 2014 15:04:33 +0100 [thread overview]
Message-ID: <538DD5F1.6060004@citrix.com> (raw)
In-Reply-To: <9AAE0902D5BC7E449B7C8E4E778ABCD038E22E@AMSPEX01CL01.citrite.net>
On 03/06/14 14:37, Paul Durrant wrote:
>> -----Original Message-----
>> From: Zoltan Kiss
>> Sent: 03 June 2014 14:32
>> To: xen-devel@lists.xenproject.org; Ian Campbell; Wei Liu; Paul Durrant;
>> linux@eikelenboom.it
>> Cc: netdev@vger.kernel.org; David Vrabel; davem@davemloft.net; Zoltan
>> Kiss
>> Subject: [PATCH net] xen-netback: Fix slot estimation
>>
>> A recent commit (a02eb4 "xen-netback: worse-case estimate in
>> xenvif_rx_action is
>> underestimating") capped the slot estimation to MAX_SKB_FRAGS, but that
>> triggers
>> the next BUG_ON a few lines down, as the packet consumes more slots than
>> estimated.
>> This patch remove that cap, and if the frontend doesn't provide enough slot,
>> put back the skb to the top of the queue and caps rx_last_skb_slots. When
>> the
>> next try also fails, it drops the packet.
>> Capping rx_last_skb_slots is needed because if the frontend never gives
>> enough
>> slots, the ring gets stalled.
>>
>> Signed-off-by: Zoltan Kiss <zoltan.kiss@citrix.com>
>> Cc: Paul Durrant <paul.durrant@citrix.com>
>> Cc: Wei Liu <wei.liu2@citrix.com>
>> Cc: Ian Campbell <ian.campbell@citrix.com>
>> Cc: David Vrabel <david.vrabel@citrix.com>
>>
>> diff --git a/drivers/net/xen-netback/netback.c b/drivers/net/xen-
>> netback/netback.c
>> index da85ffb..7164157 100644
>> --- a/drivers/net/xen-netback/netback.c
>> +++ b/drivers/net/xen-netback/netback.c
>> @@ -600,13 +600,6 @@ static void xenvif_rx_action(struct xenvif *vif)
>> PAGE_SIZE);
>> }
>>
>> - /* To avoid the estimate becoming too pessimal for some
>> - * frontends that limit posted rx requests, cap the estimate
>> - * at MAX_SKB_FRAGS.
>> - */
>> - if (max_slots_needed > MAX_SKB_FRAGS)
>> - max_slots_needed = MAX_SKB_FRAGS;
>> -
>> /* We may need one more slot for GSO metadata */
>> if (skb_is_gso(skb) &&
>> (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4 ||
>> @@ -615,9 +608,27 @@ static void xenvif_rx_action(struct xenvif *vif)
>>
>> /* If the skb may not fit then bail out now */
>> if (!xenvif_rx_ring_slots_available(vif, max_slots_needed)) {
>> + /* If the skb needs more than MAX_SKB_FRAGS
>> slots, it
>> + * can happen that the frontend never gives us
>> enough.
>> + * To avoid spining on that packet, first we put it back
>> + * to the top of the queue, but if the next try fail,
>> + * we drop it.
>> + */
>> + if (max_slots_needed > MAX_SKB_FRAGS &&
>> + vif->rx_last_skb_slots == MAX_SKB_FRAGS) {
>
> Isn't it sufficient to say:
>
> if (vif->rx_last_skb_slots != 0)
>
> here? We should not ordinarily wake before the requisite number of slots is available.
Yep, that would be enough
>
> Paul
>
>> + kfree_skb(skb);
>> + vif->rx_last_skb_slots = 0;
>> + continue;
>> + }
>> skb_queue_head(&vif->rx_queue, skb);
>> need_to_notify = true;
>> - vif->rx_last_skb_slots = max_slots_needed;
>> + /* Cap this otherwise if the guest never gives us
>> + * enough slot, rx_work_todo will spin
>> + */
>> + vif->rx_last_skb_slots =
>> + max_slots_needed > MAX_SKB_FRAGS ?
>> + MAX_SKB_FRAGS :
>> + max_slots_needed;
>> break;
>> } else
>> vif->rx_last_skb_slots = 0;
next prev parent reply other threads:[~2014-06-03 14:04 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-06-03 13:32 [PATCH net] xen-netback: Fix slot estimation Zoltan Kiss
2014-06-03 13:37 ` Paul Durrant
2014-06-03 14:04 ` Zoltan Kiss [this message]
2014-06-03 13:52 ` David Laight
2014-06-03 20:24 ` Zoltan Kiss
2014-06-05 22:02 ` David Miller
2014-06-06 10:20 ` Zoltan Kiss
2014-06-06 20:06 ` David Miller
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=538DD5F1.6060004@citrix.com \
--to=zoltan.kiss@citrix.com \
--cc=Ian.Campbell@citrix.com \
--cc=Paul.Durrant@citrix.com \
--cc=davem@davemloft.net \
--cc=david.vrabel@citrix.com \
--cc=linux@eikelenboom.it \
--cc=netdev@vger.kernel.org \
--cc=wei.liu2@citrix.com \
--cc=xen-devel@lists.xenproject.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).