From: Jesper Dangaard Brouer <brouer@redhat.com>
To: Eric Dumazet <eric.dumazet@gmail.com>
Cc: David Miller <davem@davemloft.net>,
hannes@stressinduktion.org, netdev@vger.kernel.org,
therbert@google.com, fw@strlen.de, dborkman@redhat.com,
jhs@mojatatu.com, alexander.duyck@gmail.com,
john.r.fastabend@intel.com, dave.taht@gmail.com, toke@toke.dk,
brouer@redhat.com
Subject: Re: Quota in __qdisc_run()
Date: Tue, 7 Oct 2014 21:34:23 +0200 [thread overview]
Message-ID: <20141007213423.37bfac8b@redhat.com> (raw)
In-Reply-To: <1412709015.11091.158.camel@edumazet-glaptop2.roam.corp.google.com>
On Tue, 07 Oct 2014 12:10:15 -0700
Eric Dumazet <eric.dumazet@gmail.com> wrote:
> On Tue, 2014-10-07 at 20:03 +0200, Jesper Dangaard Brouer wrote:
>
> > According to my measurements, at 10Gbit/s TCP_STREAM test the BQL limit
> > is 381528 bytes / 1514 = 252 packets, that will (potentially) be bulk
> > dequeued at once (with your version of the patch).
> >
>
> That's because you use a single queue maybe ?
No, I just double checked.
I was using a single netperf TCP_STREAM with GRO=off TSO=off.
> Here we have limits around 2 TSO packets.
Which should be approx 128k
Just tested with GRO=on TSO=on single TCP_STREAM, which is weird as I
should hit your 2xTSO limit right, and inflight shows 408780.
And I do have (which I guess is the 2xTSO):
/proc/sys/net/ipv4/tcp_limit_output_bytes:131072
$ grep -H . /sys/class/net/eth4/queues/tx-*/byte_queue_limits/{inflight,limit}
/sys/class/net/eth4/queues/tx-0/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-10/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-11/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-1/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-2/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-3/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-4/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-5/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-6/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-7/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-8/byte_queue_limits/inflight:408780
/sys/class/net/eth4/queues/tx-9/byte_queue_limits/inflight:0
/sys/class/net/eth4/queues/tx-0/byte_queue_limits/limit:340848
/sys/class/net/eth4/queues/tx-10/byte_queue_limits/limit:340650
/sys/class/net/eth4/queues/tx-11/byte_queue_limits/limit:367902
/sys/class/net/eth4/queues/tx-1/byte_queue_limits/limit:272520
/sys/class/net/eth4/queues/tx-2/byte_queue_limits/limit:204390
/sys/class/net/eth4/queues/tx-3/byte_queue_limits/limit:162856
/sys/class/net/eth4/queues/tx-4/byte_queue_limits/limit:158314
/sys/class/net/eth4/queues/tx-5/byte_queue_limits/limit:136260
/sys/class/net/eth4/queues/tx-6/byte_queue_limits/limit:140802
/sys/class/net/eth4/queues/tx-7/byte_queue_limits/limit:152258
/sys/class/net/eth4/queues/tx-8/byte_queue_limits/limit:340650
/sys/class/net/eth4/queues/tx-9/byte_queue_limits/limit:340650
Strange...
> Even with only 4 tx queues I have :
>
> # sar -n DEV 3 3 |grep eth1
> 12:05:19 PM eth1 147217.67 809066.67 9488.71 1196207.78 0.00 0.00 0.00
> 12:05:22 PM eth1 145958.00 807822.33 9407.48 1194366.73 0.00 0.00 0.00
> 12:05:25 PM eth1 147502.33 804739.33 9507.26 1189804.23 0.00 0.00 0.33
> Average: eth1 146892.67 807209.44 9467.82 1193459.58 0.00 0.00 0.11
>
>
> grep . /sys/class/net/eth1/queues/tx*/byte_queue_limits/{inflight,limit}
> /sys/class/net/eth1/queues/tx-0/byte_queue_limits/inflight:115064
> /sys/class/net/eth1/queues/tx-1/byte_queue_limits/inflight:0
> /sys/class/net/eth1/queues/tx-2/byte_queue_limits/inflight:0
> /sys/class/net/eth1/queues/tx-3/byte_queue_limits/inflight:0
> /sys/class/net/eth1/queues/tx-0/byte_queue_limits/limit:102952
> /sys/class/net/eth1/queues/tx-1/byte_queue_limits/limit:124148
> /sys/class/net/eth1/queues/tx-2/byte_queue_limits/limit:102952
> /sys/class/net/eth1/queues/tx-3/byte_queue_limits/limit:136260
Guess this is okay, 115064 / 1514 = 76 pkts which is closer to the 64 weight_p.
> > It seems to have the potential to exceed the weight_p(64) quite a lot.
> > And with e.g. TX ring size 512, we also also challenge the drivers at
> > this early adoption phase of tailptr writes. Just saying...
> >
>
> Yep, but remind we want to squeeze bugs out of the drivers, then add
> additional knobs later.
Okay, for squeezing out bugs, then I understand this more aggressive
bulking strategy. I'm all in then!
> Whatever limit we choose in core networking stack (being 64 packets for
> example), hardware might have different constraints that need to be
> taken care of in the driver.
--
Best regards,
Jesper Dangaard Brouer
MSc.CS, Sr. Network Kernel Developer at Red Hat
Author of http://www.iptv-analyzer.org
LinkedIn: http://www.linkedin.com/in/brouer
next prev parent reply other threads:[~2014-10-07 19:34 UTC|newest]
Thread overview: 42+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-10-01 20:35 [net-next PATCH V6 0/2] qdisc: bulk dequeue support Jesper Dangaard Brouer
2014-10-01 20:35 ` [net-next PATCH V6 1/2] qdisc: bulk dequeue support for qdiscs with TCQ_F_ONETXQUEUE Jesper Dangaard Brouer
2014-10-01 20:36 ` [net-next PATCH V6 2/2] qdisc: dequeue bulking also pickup GSO/TSO packets Jesper Dangaard Brouer
2014-10-02 14:35 ` Eric Dumazet
2014-10-02 14:38 ` Daniel Borkmann
2014-10-02 14:42 ` [net-next PATCH V6 0/2] qdisc: bulk dequeue support Tom Herbert
2014-10-02 15:04 ` Eric Dumazet
2014-10-02 15:24 ` [PATCH net-next] mlx4: add a new xmit_more counter Eric Dumazet
2014-10-05 0:04 ` David Miller
2014-10-02 15:27 ` [net-next PATCH V6 0/2] qdisc: bulk dequeue support Tom Herbert
2014-10-02 16:52 ` Florian Westphal
2014-10-02 17:32 ` Eric Dumazet
2014-10-02 17:35 ` Tom Herbert
2014-10-03 19:38 ` David Miller
2014-10-03 20:57 ` Eric Dumazet
2014-10-03 21:56 ` David Miller
2014-10-03 21:57 ` Eric Dumazet
2014-10-03 22:15 ` Eric Dumazet
2014-10-03 22:19 ` Tom Herbert
2014-10-03 22:56 ` Eric Dumazet
2014-10-03 22:30 ` David Miller
2014-10-03 22:31 ` [PATCH net-next] qdisc: validate skb without holding lock Eric Dumazet
2014-10-03 22:36 ` David Miller
2014-10-03 23:30 ` Eric Dumazet
2014-10-07 7:34 ` Quota in __qdisc_run() (was: qdisc: validate skb without holding lock) Jesper Dangaard Brouer
2014-10-07 12:47 ` Eric Dumazet
2014-10-07 13:30 ` Jesper Dangaard Brouer
2014-10-07 14:43 ` Hannes Frederic Sowa
2014-10-07 15:01 ` Eric Dumazet
2014-10-07 15:06 ` Eric Dumazet
2014-10-07 17:19 ` Quota in __qdisc_run() David Miller
2014-10-07 17:32 ` Eric Dumazet
2014-10-07 18:37 ` Jesper Dangaard Brouer
2014-10-07 20:07 ` Jesper Dangaard Brouer
2014-10-07 18:03 ` Jesper Dangaard Brouer
2014-10-07 19:10 ` Eric Dumazet
2014-10-07 19:34 ` Jesper Dangaard Brouer [this message]
2014-10-07 15:26 ` Quota in __qdisc_run() (was: qdisc: validate skb without holding lock) Jesper Dangaard Brouer
2014-10-08 17:38 ` Quota in __qdisc_run() John Fastabend
2014-10-06 14:12 ` [PATCH net-next] qdisc: validate skb without holding lock Jesper Dangaard Brouer
2014-10-04 3:59 ` [PATCH net-next] net: skb_segment() provides list head and tail Eric Dumazet
2014-10-06 4:38 ` David Miller
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20141007213423.37bfac8b@redhat.com \
--to=brouer@redhat.com \
--cc=alexander.duyck@gmail.com \
--cc=dave.taht@gmail.com \
--cc=davem@davemloft.net \
--cc=dborkman@redhat.com \
--cc=eric.dumazet@gmail.com \
--cc=fw@strlen.de \
--cc=hannes@stressinduktion.org \
--cc=jhs@mojatatu.com \
--cc=john.r.fastabend@intel.com \
--cc=netdev@vger.kernel.org \
--cc=therbert@google.com \
--cc=toke@toke.dk \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).