* [ANNOUNCE] ndiv: line-rate network traffic processing @ 2016-09-21 11:28 Willy Tarreau 2016-09-21 16:26 ` Jesper Dangaard Brouer 2016-09-21 17:16 ` Tom Herbert 0 siblings, 2 replies; 5+ messages in thread From: Willy Tarreau @ 2016-09-21 11:28 UTC (permalink / raw) To: netdev Hi, Over the last 3 years I've been working a bit on high traffic processing for various reasons. It started with the wish to capture line-rate GigE traffic on very small fanless ARM machines and the framework has evolved to be used at my company as a basis for our anti-DDoS engine capable of dealing with multiple 10G links saturated with floods. I know it comes a bit late now that there is XDP, but it's my first vacation since then and I needed to have a bit of calm time to collect the patches from the various branches and put them together. Anyway I'm sending this here in case it can be of interest to anyone, for use or just to study it. I presented it in 2014 at kernel recipes : http://kernel-recipes.org/en/2014/ndiv-a-low-overhead-network-traffic-diverter/ It now supports drivers mvneta, ixgbe, e1000e, e1000 and igb. It is very light, and retrieves the packets in the NIC's driver before they are converted to an skb, then submits them to a registered RX handler running in softirq context so we have the best of all worlds by benefitting from CPU scalability, delayed processing, and not paying the cost of switching to userland. Also an rx_done() function allows handlers to batch their processing. The RX handler returns an action among accepting the packet as-is, accepting it modified (eg: vlan or tunnel decapsulation), dropping it, postponing the processing (equivalent to EAGAIN), or building a new packet to send back. This last function is the one requiring the most changes in existing drivers, but offers the widest range of possibilities. We use it to send SYN cookies, but I have also implemented a stateless HTTP server supporting keep-alive using it, achieving line-rate traffic processing on a single CPU core when the NIC supports it. It's very convenient to test various stateful TCP components as it's easy to sustain millions of connections per second on it. It does not support forwarding between NICs. It was my first goal because I wanted to implement a TAP with it, bridging the traffic between two ports, but figured it was adding some complexity to the system back then. However since then we've implemented traffic capture in our product, exploiting this framework to capture without losses at 14 Mpps. I may find some time to try to extract it later. It uses the /sys API so that you can simply plug tcpdump -r on a file there, though there's also an mmap version which uses less CPU (that's important at 10G). In its current form since the initial code's intent was to limit core changes, it happens not to modify anything in the kernel by default and to reuse the net_device's ax25_ptr to attach devices (idea borrowed from netmap), so it can be used on an existing kernel just by loading the patched network drivers (yes, I know it's not a valid solution for the long term). The current code is available here : http://git.kernel.org/cgit/linux/kernel/git/wtarreau/ndiv.git/ Please let me know if there could be some interest in rebasing it on more recent versions (currently 3.10, 3.14 and 4.4 are supported). I don't have much time to assign to it since it works fine as-is, but will be glad to do so if that can be useful. Also the stateless HTTP server provided in it definitely is a nice use case for testing such a framework. Regards, Willy ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [ANNOUNCE] ndiv: line-rate network traffic processing 2016-09-21 11:28 [ANNOUNCE] ndiv: line-rate network traffic processing Willy Tarreau @ 2016-09-21 16:26 ` Jesper Dangaard Brouer 2016-09-21 18:00 ` Willy Tarreau 2016-09-21 17:16 ` Tom Herbert 1 sibling, 1 reply; 5+ messages in thread From: Jesper Dangaard Brouer @ 2016-09-21 16:26 UTC (permalink / raw) To: Willy Tarreau Cc: brouer, netdev, Tom Herbert, Alexei Starovoitov, Brenden Blanco, Tariq Toukan On Wed, 21 Sep 2016 13:28:52 +0200 Willy Tarreau <w@1wt.eu> wrote: > Over the last 3 years I've been working a bit on high traffic processing > for various reasons. It started with the wish to capture line-rate GigE > traffic on very small fanless ARM machines and the framework has evolved > to be used at my company as a basis for our anti-DDoS engine capable of > dealing with multiple 10G links saturated with floods. > > I know it comes a bit late now that there is XDP, but it's my first > vacation since then and I needed to have a bit of calm time to collect > the patches from the various branches and put them together. Anyway I'm > sending this here in case it can be of interest to anyone, for use or > just to study it. I definitely want to study it! You mention XDP. If you didn't notice, I've created some documentation on XDP (it is very "live" documentation at this point and it will hopefully "materialized" later in the process). But it should be a good starting point for understanding XDP: https://prototype-kernel.readthedocs.io/en/latest/networking/XDP/index.html > I presented it in 2014 at kernel recipes : > http://kernel-recipes.org/en/2014/ndiv-a-low-overhead-network-traffic-diverter/ Cool, and it even have a video! > It now supports drivers mvneta, ixgbe, e1000e, e1000 and igb. It is > very light, and retrieves the packets in the NIC's driver before they > are converted to an skb, then submits them to a registered RX handler > running in softirq context so we have the best of all worlds by > benefitting from CPU scalability, delayed processing, and not paying > the cost of switching to userland. Also an rx_done() function allows > handlers to batch their processing. Wow - it does sound a lot like XDP! I would say that is sort of validate the current direction of XDP, and that there are real use-cases for this stuff. > The RX handler returns an action > among accepting the packet as-is, accepting it modified (eg: vlan or > tunnel decapsulation), dropping it, postponing the processing > (equivalent to EAGAIN), or building a new packet to send back. I'll be very interested in studying in-details how you implemented and choose what actions to implement. What was the need for postponing the processing (EAGAIN)? > This last function is the one requiring the most changes in existing > drivers, but offers the widest range of possibilities. We use it to > send SYN cookies, but I have also implemented a stateless HTTP server > supporting keep-alive using it, achieving line-rate traffic processing > on a single CPU core when the NIC supports it. It's very convenient to > test various stateful TCP components as it's easy to sustain millions > of connections per second on it. Interesting, and controversial use-case. One controversial use-case for XDP, that I imagine was implementing a DNS accelerator, what answers simple and frequent requests. You took it a step further with a HTTP server! > It does not support forwarding between NICs. It was my first goal > because I wanted to implement a TAP with it, bridging the traffic > between two ports, but figured it was adding some complexity to the > system back then. With all the XDP features at the moment, we have avoided going through the page allocator, by relying on different page recycling tricks. When doing forwarding between NICs is it harder to do these page recycling tricks. I've measured that page allocators fast-path ("recycling" same page) cost approx 270 cycles, and the 14Mpps cycle count on this 4GHz CPU is 268 cycles. Thus, it is a non-starter... Did you have to modify the page allocator? Or implement some kind of recycling? > However since then we've implemented traffic > capture in our product, exploiting this framework to capture without > losses at 14 Mpps. I may find some time to try to extract it later. > It uses the /sys API so that you can simply plug tcpdump -r on a > file there, though there's also an mmap version which uses less CPU > (that's important at 10G). Interesting. I do see a XDP use-case for RAW packet capture, but I've postponed that work until later. I would interested in how you solved it? E.g. Do you support zero-copy? > In its current form since the initial code's intent was to limit > core changes, it happens not to modify anything in the kernel by > default and to reuse the net_device's ax25_ptr to attach devices > (idea borrowed from netmap), so it can be used on an existing > kernel just by loading the patched network drivers (yes, I know > it's not a valid solution for the long term). > > The current code is available here : > > http://git.kernel.org/cgit/linux/kernel/git/wtarreau/ndiv.git/ I was just about to complain that the link was broken... but it fixed itself while writing this email ;-) Can you instead explain what branch to look at? > > Please let me know if there could be some interest in rebasing it > on more recent versions (currently 3.10, 3.14 and 4.4 are supported). What, no support for 2.4 ;-) > I don't have much time to assign to it since it works fine as-is, > but will be glad to do so if that can be useful. > > Also the stateless HTTP server provided in it definitely is a nice > use case for testing such a framework. -- Best regards, Jesper Dangaard Brouer MSc.CS, Principal Kernel Engineer at Red Hat Author of http://www.iptv-analyzer.org LinkedIn: http://www.linkedin.com/in/brouer ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [ANNOUNCE] ndiv: line-rate network traffic processing 2016-09-21 16:26 ` Jesper Dangaard Brouer @ 2016-09-21 18:00 ` Willy Tarreau 0 siblings, 0 replies; 5+ messages in thread From: Willy Tarreau @ 2016-09-21 18:00 UTC (permalink / raw) To: Jesper Dangaard Brouer Cc: netdev, Tom Herbert, Alexei Starovoitov, Brenden Blanco, Tariq Toukan Hi Jesper! On Wed, Sep 21, 2016 at 06:26:39PM +0200, Jesper Dangaard Brouer wrote: > I definitely want to study it! Great, at least I've not put this online for nothing :-) > You mention XDP. If you didn't notice, I've created some documentation > on XDP (it is very "live" documentation at this point and it will > hopefully "materialized" later in the process). But it should be a good > starting point for understanding XDP: > > https://prototype-kernel.readthedocs.io/en/latest/networking/XDP/index.html Thanks, I'll read it. We'll need to educate ourselves to see how to port our anti-ddos to XDP in the future I guess, so better ensure the design is fit from the beginning! > > I presented it in 2014 at kernel recipes : > > http://kernel-recipes.org/en/2014/ndiv-a-low-overhead-network-traffic-diverter/ > > Cool, and it even have a video! Yep, with a horrible english accent :-) > > It now supports drivers mvneta, ixgbe, e1000e, e1000 and igb. It is > > very light, and retrieves the packets in the NIC's driver before they > > are converted to an skb, then submits them to a registered RX handler > > running in softirq context so we have the best of all worlds by > > benefitting from CPU scalability, delayed processing, and not paying > > the cost of switching to userland. Also an rx_done() function allows > > handlers to batch their processing. > > Wow - it does sound a lot like XDP! I would say that is sort of > validate the current direction of XDP, and that there are real > use-cases for this stuff. Absolutely! In fact what drove use to this architecture is that we first wrote our anti-ddos in userland using netmap. While userland might be OK for switches and routers, in our case we have haproxy listening on TCP sockets and waiting for these packets. So the packets were bouncing from kernel to user, then to kernel again, losing checksums, GRO, GSO, etc... We modified it to support all of these but the performance was still poor, capping at about 8 Gbps of forwarded traffic instead of ~40. Thus we thought that the processing would definitely need to be placed in the kernel to avoid this bouncing, and to avoid turning rings into newer rings all the time. That's when I realized that it could possibly also cover my needs for a sniffer and we redesigned the initial code to support both use cases. Now we don't even see it in regular traffic, which is pretty nice. > > The RX handler returns an action > > among accepting the packet as-is, accepting it modified (eg: vlan or > > tunnel decapsulation), dropping it, postponing the processing > > (equivalent to EAGAIN), or building a new packet to send back. > > I'll be very interested in studying in-details how you implemented and > choose what actions to implement. OK. The HTTP server is a good use case to study because it lets packets pass through, being dropped, or being responded to, and the code is very small, so easy to analyse. > What was the need for postponing the processing (EAGAIN)? Our SYN cookie generator. If the NIC's Tx queue is full and we cannot build a SYN-ACK, we prefer to break out of the Rx loop because there's still room in the Rx ring (statistically speaking). > > This last function is the one requiring the most changes in existing > > drivers, but offers the widest range of possibilities. We use it to > > send SYN cookies, but I have also implemented a stateless HTTP server > > supporting keep-alive using it, achieving line-rate traffic processing > > on a single CPU core when the NIC supports it. It's very convenient to > > test various stateful TCP components as it's easy to sustain millions > > of connections per second on it. > > Interesting, and controversial use-case. One controversial use-case > for XDP, that I imagine was implementing a DNS accelerator, what > answers simple and frequent requests. We thought about such a use case as well, just like of a ping responder (rate limited to avoid serving as DDoS responders). > You took it a step further with a HTTP server! It's a fake HTTP server. You ask it to return 1kB of data and it sends you 1kB. It can even do multiple segments but then you're facing the risk of losses that you'd preferably avoid. But in our case it's very useful to test various things including netfilter, LVS and haproxy because it consumes so little power to reach performance levels that they cannot even reach that you can set it up on a small machine (eg: a cheap USB-powered ARM board saturates the GigE link with 340 kcps, 663 krps). However I found that it *could* be fun to improve it to deliver favicons or small error pages. > > It does not support forwarding between NICs. It was my first goal > > because I wanted to implement a TAP with it, bridging the traffic > > between two ports, but figured it was adding some complexity to the > > system back then. > > With all the XDP features at the moment, we have avoided going through > the page allocator, by relying on different page recycling tricks. > > When doing forwarding between NICs is it harder to do these page > recycling tricks. I've measured that page allocators fast-path > ("recycling" same page) cost approx 270 cycles, and the 14Mpps cycle > count on this 4GHz CPU is 268 cycles. Thus, it is a non-starter... Wow indeed. We're doing complete stateful inspection and policy-based filtering on less than 67 ns, so indeed here it would be far too long. > Did you have to modify the page allocator? > Or implement some kind of recycling? We "simply" implemented our own Tx ring, depending on what drivers and hardware support. This is the most complicated part of the code because it is very hardware-dependant and because you want to deal with conflicts between the packets being generated on the Rx path and other packets being delivered by other cores on the regular Tx path. In some drivers we cheat on the skb pointer in the descriptors, we set bit 0 to 1 to mark it as being ours so that we recycle it into our ring after it's sent instead of releasing an skb. That's why it would be hard to implement forwarding. I thought that I could at least implement it between two NICs making use of the same driver, but it could lead to starvation of certain tx rings and other ones filling up. However I don't have a solution for now because I decided to stop thinking about it at the moment. Over the long term I would love to see my mirabox being used as an inline tap logging to USB3 :-) Another important design choice that comes to my mind is that we purposely decide to support optimal devices. We decided this after seeing how netmap uses only the least common denominator between all supportable NICs resulting in any NIC to become dumb. In our case, the driver has to feed the checksum status, L3/L4 protocol types etc. If the NIC is too dumb to support this, it just has to be implemented in software for this NIC only. And in practice all NICs that matter support L3/L4 protocol identification as well as checksum verification/computation, so it's not a problem and the hardware continues to work for us for free. > > However since then we've implemented traffic > > capture in our product, exploiting this framework to capture without > > losses at 14 Mpps. I may find some time to try to extract it later. > > It uses the /sys API so that you can simply plug tcpdump -r on a > > file there, though there's also an mmap version which uses less CPU > > (that's important at 10G). > > Interesting. I do see a XDP use-case for RAW packet capture, but I've > postponed that work until later. I would interested in how you solved > it? E.g. Do you support zero-copy? No, we intentionally copy. In fact on xeon processors, the memory bandwidth is so huge that you don't even notice the copy. And by using small buffers, you can even ensure that the memory blocks stays in L3 cache. We had most difficulties with the /sys API because it only supports page-size transfers and uses lots of CPU just for this, hence we had to implement mmap support to present the packets to user-space (without copy here). But even the regular /sys API with a double copy supports line-rate with high CPU usage. I'll ask my coworker Emeric who did the sniffer if he can take it out as a standalone component. It will take a bit of time because we're moving to a new office and that significantly mangles our priorities as you can expect, but that's definitely something we'd like to do. > > In its current form since the initial code's intent was to limit > > core changes, it happens not to modify anything in the kernel by > > default and to reuse the net_device's ax25_ptr to attach devices > > (idea borrowed from netmap), so it can be used on an existing > > kernel just by loading the patched network drivers (yes, I know > > it's not a valid solution for the long term). > > > > The current code is available here : > > > > http://git.kernel.org/cgit/linux/kernel/git/wtarreau/ndiv.git/ > > I was just about to complain that the link was broken... but it fixed > itself while writing this email ;-) I noticed the same thing a few times already, and whatever I do, the description is not updated. I suspect there's some load balancing with one server not being updated as fast as the other ones. > Can you instead explain what branch to look at? Sure! For the most up-to-date code, better use ndiv_v5-4.4. It contains the core (a single .h file), and support for ixgbe, e1000e, e1000, igb, and the dummy HTTP server (slhttpd). For a more readable version, better use ndiv_v5-3.14 which also contains the mvneta driver, it's much simpler than the other ones and makes the code more readable. I'll have to port it to 4.4 soon but didn't have time yet. We don't support mlx4 yet, and it's a chicken-and-egg problem : by lack of time we don't work on porting it and since we don't support it we don't use it in our products. That's too bad because from what I've been told we should be able to reach high packet rates there as well. > > Please let me know if there could be some interest in rebasing it > > on more recent versions (currently 3.10, 3.14 and 4.4 are supported). > > What, no support for 2.4 ;-) Not yet :-) Jokes aside, given that the API is very simple, it could be done if anyone needed, as it really doesn't rely on any existing infrastructure. The API is reasonably OS-agnostic as it only wants pointers and lengths. For sniffing and/or filtering on Rx/Tx paths only, the code basically only is (synthetic code, just to illustrate) : ndiv = netdev_get_ndiv(dev); if (ndiv) { ret = ndiv->handle_rx(ndiv, l3ptr, l3len, l2len, l2ptr, NULL); if (ret & NDIV_RX_R_F_DROP) continue; } Best regards, Willy ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [ANNOUNCE] ndiv: line-rate network traffic processing 2016-09-21 11:28 [ANNOUNCE] ndiv: line-rate network traffic processing Willy Tarreau 2016-09-21 16:26 ` Jesper Dangaard Brouer @ 2016-09-21 17:16 ` Tom Herbert 2016-09-21 18:06 ` Willy Tarreau 1 sibling, 1 reply; 5+ messages in thread From: Tom Herbert @ 2016-09-21 17:16 UTC (permalink / raw) To: Willy Tarreau; +Cc: Linux Kernel Network Developers On Wed, Sep 21, 2016 at 4:28 AM, Willy Tarreau <w@1wt.eu> wrote: > Hi, > > Over the last 3 years I've been working a bit on high traffic processing > for various reasons. It started with the wish to capture line-rate GigE > traffic on very small fanless ARM machines and the framework has evolved > to be used at my company as a basis for our anti-DDoS engine capable of > dealing with multiple 10G links saturated with floods. > > I know it comes a bit late now that there is XDP, but it's my first > vacation since then and I needed to have a bit of calm time to collect > the patches from the various branches and put them together. Anyway I'm > sending this here in case it can be of interest to anyone, for use or > just to study it. > > I presented it in 2014 at kernel recipes : > http://kernel-recipes.org/en/2014/ndiv-a-low-overhead-network-traffic-diverter/ > > It now supports drivers mvneta, ixgbe, e1000e, e1000 and igb. It is > very light, and retrieves the packets in the NIC's driver before they > are converted to an skb, then submits them to a registered RX handler > running in softirq context so we have the best of all worlds by > benefitting from CPU scalability, delayed processing, and not paying > the cost of switching to userland. Also an rx_done() function allows > handlers to batch their processing. The RX handler returns an action > among accepting the packet as-is, accepting it modified (eg: vlan or > tunnel decapsulation), dropping it, postponing the processing > (equivalent to EAGAIN), or building a new packet to send back. > > This last function is the one requiring the most changes in existing > drivers, but offers the widest range of possibilities. We use it to > send SYN cookies, but I have also implemented a stateless HTTP server > supporting keep-alive using it, achieving line-rate traffic processing > on a single CPU core when the NIC supports it. It's very convenient to > test various stateful TCP components as it's easy to sustain millions > of connections per second on it. > > It does not support forwarding between NICs. It was my first goal > because I wanted to implement a TAP with it, bridging the traffic > between two ports, but figured it was adding some complexity to the > system back then. However since then we've implemented traffic > capture in our product, exploiting this framework to capture without > losses at 14 Mpps. I may find some time to try to extract it later. > It uses the /sys API so that you can simply plug tcpdump -r on a > file there, though there's also an mmap version which uses less CPU > (that's important at 10G). > > In its current form since the initial code's intent was to limit > core changes, it happens not to modify anything in the kernel by > default and to reuse the net_device's ax25_ptr to attach devices > (idea borrowed from netmap), so it can be used on an existing > kernel just by loading the patched network drivers (yes, I know > it's not a valid solution for the long term). > > The current code is available here : > > http://git.kernel.org/cgit/linux/kernel/git/wtarreau/ndiv.git/ > > Please let me know if there could be some interest in rebasing it > on more recent versions (currently 3.10, 3.14 and 4.4 are supported). > I don't have much time to assign to it since it works fine as-is, > but will be glad to do so if that can be useful. > Hi Willy, This does seem interesting and indeed the driver datapath looks very much like XDP. It would be quite interesting if you could rebase and then maybe look at how this can work with XDP that would be helpful. The return actions are identical, but processing descriptor meta data (like checksum, vlan) is not yet implemented in XDP-- maybe this is something we can leverage from ndiv? Tom > Also the stateless HTTP server provided in it definitely is a nice > use case for testing such a framework. > > Regards, > Willy ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [ANNOUNCE] ndiv: line-rate network traffic processing 2016-09-21 17:16 ` Tom Herbert @ 2016-09-21 18:06 ` Willy Tarreau 0 siblings, 0 replies; 5+ messages in thread From: Willy Tarreau @ 2016-09-21 18:06 UTC (permalink / raw) To: Tom Herbert; +Cc: Linux Kernel Network Developers Hi Tom, On Wed, Sep 21, 2016 at 10:16:45AM -0700, Tom Herbert wrote: > This does seem interesting and indeed the driver datapath looks very > much like XDP. It would be quite interesting if you could rebase and > then maybe look at how this can work with XDP that would be helpful. OK I'll assign some time to rebase it then. > The return actions are identical, I'm not surprized that the same needs lead to the same designs when these designs are constrained by CPU cycle count :-) > but processing descriptor meta data > (like checksum, vlan) is not yet implemented in XDP-- maybe this is > something we can leverage from ndiv? Yes possibly. It's not a big work but it's absolutely mandatory if you don't want to waste some smart devices' valuable performance improvements. We changed our API when porting it to ixgbe to support what this NIC (and many other ones) supports so that the application code doesn't have to deal with checksums etc. By the way, VLAN is not yet implemented in the mvneta driver. But this choice ensures that no application has to deal nor to create bugs. Cheers, Willy ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2016-09-21 18:07 UTC | newest] Thread overview: 5+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2016-09-21 11:28 [ANNOUNCE] ndiv: line-rate network traffic processing Willy Tarreau 2016-09-21 16:26 ` Jesper Dangaard Brouer 2016-09-21 18:00 ` Willy Tarreau 2016-09-21 17:16 ` Tom Herbert 2016-09-21 18:06 ` Willy Tarreau
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).