From: Jonathan Lemon <jonathan.lemon@gmail.com>
To: intel-wired-lan@osuosl.org
Subject: [Intel-wired-lan] [PATCH 00/11] XDP unaligned chunk placement support
Date: Tue, 02 Jul 2019 09:33:52 -0700 [thread overview]
Message-ID: <3510BE85-6B1B-4BB4-9640-ECEE2572DB4E@gmail.com> (raw)
In-Reply-To: <59AF69C657FD0841A61C55336867B5B07ED8B210@IRSMSX103.ger.corp.intel.com>
On 2 Jul 2019, at 2:27, Richardson, Bruce wrote:
>> -----Original Message-----
>> From: Jakub Kicinski [mailto:jakub.kicinski at netronome.com]
>> Sent: Monday, July 1, 2019 10:20 PM
>> To: Laatz, Kevin <kevin.laatz@intel.com>
>> Cc: Jonathan Lemon <jonathan.lemon@gmail.com>;
>> netdev at vger.kernel.org;
>> ast at kernel.org; daniel at iogearbox.net; Topel, Bjorn
>> <bjorn.topel@intel.com>; Karlsson, Magnus
>> <magnus.karlsson@intel.com>;
>> bpf at vger.kernel.org; intel-wired-lan at lists.osuosl.org; Richardson,
>> Bruce
>> <bruce.richardson@intel.com>; Loftus, Ciara <ciara.loftus@intel.com>
>> Subject: Re: [PATCH 00/11] XDP unaligned chunk placement support
>>
>> On Mon, 1 Jul 2019 15:44:29 +0100, Laatz, Kevin wrote:
>>> On 28/06/2019 21:29, Jonathan Lemon wrote:
>>>> On 28 Jun 2019, at 9:19, Laatz, Kevin wrote:
>>>>> On 27/06/2019 22:25, Jakub Kicinski wrote:
>>>>>> I think that's very limiting.? What is the challenge in
>>>>>> providing
>>>>>> aligned addresses, exactly?
>>>>> The challenges are two-fold:
>>>>> 1) it prevents using arbitrary buffer sizes, which will be an
>>>>> issue
>>>>> supporting e.g. jumbo frames in future.
>>>>> 2) higher level user-space frameworks which may want to use
>>>>> AF_XDP,
>>>>> such as DPDK, do not currently support having buffers with 'fixed'
>>>>> alignment.
>>>>> ??? The reason that DPDK uses arbitrary placement is that:
>>>>> ??? ??? - it would stop things working on certain NICs which
>>>>> need
>>>>> the actual writable space specified in units of 1k - therefore we
>>>>> need 2k
>>>>> + metadata space.
>>>>> ??? ??? - we place padding between buffers to avoid
>>>>> constantly
>>>>> hitting the same memory channels when accessing memory.
>>>>> ??? ??? - it allows the application to choose the actual
>>>>> buffer
>>>>> size it wants to use.
>>>>> ??? We make use of the above to allow us to speed up processing
>>>>> significantly and also reduce the packet buffer memory size.
>>>>>
>>>>> ??? Not having arbitrary buffer alignment also means an AF_XDP
>>>>> driver for DPDK cannot be a drop-in replacement for existing
>>>>> drivers in those frameworks. Even with a new capability to allow
>>>>> an
>>>>> arbitrary buffer alignment, existing apps will need to be modified
>>>>> to use that new capability.
>>>>
>>>> Since all buffers in the umem are the same chunk size, the original
>>>> buffer address can be recalculated with some multiply/shift math.
>>>> However, this is more expensive than just a mask operation.
>>>
>>> Yes, we can do this.
>>
>> That'd be best, can DPDK reasonably guarantee the slicing is uniform?
>> E.g. it's not desperate buffer pools with different bases?
>
> It's generally uniform, but handling the crossing of (huge)page
> boundaries
> complicates things a bit. Therefore I think the final option below
> is best as it avoids any such problems.
>
>>
>>> Another option we have is to add a socket option for querying the
>>> metadata length from the driver (assuming it doesn't vary per
>>> packet).
>>> We can use that information to get back to the original address
>>> using
>>> subtraction.
>>
>> Unfortunately the metadata depends on the packet and how much info
>> the
>> device was able to extract. So it's variable length.
>>
>>> Alternatively, we can change the Rx descriptor format to include the
>>> metadata length. We could do this in a couple of ways, for example,
>>> rather than returning the address as the start of the packet,
>>> instead
>>> return the buffer address that was passed in, and adding another
>>> 16-bit field to specify the start of packet offset with that buffer.
>>> If using another 16-bits of the descriptor space is not desirable,
>>> an
>>> alternative could be to limit umem sizes to e.g. 2^48 bits (256
>>> terabytes should be enough, right :-) ) and use the remaining 16
>>> bits
>>> of the address as a packet offset. Other variations on these
>>> approach
>>> are obviously possible too.
>>
>> Seems reasonable to me..
>
> I think this is probably the best solution, and also has the advantage
> that
> a buffer retains its base address the full way through the cycle of Rx
> and Tx.
I like this as well - it also has the advantage that drivers can keep
performing adjustments on the handle, which ends up just modifying the
offset.
--
Jonathan
next prev parent reply other threads:[~2019-07-02 16:33 UTC|newest]
Thread overview: 28+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-06-20 8:39 [Intel-wired-lan] [PATCH 00/11] XDP unaligned chunk placement support Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 01/11] i40e: simplify Rx buffer recycle Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 02/11] ixgbe: " Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 03/11] xdp: add offset param to zero_copy_allocator Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 04/11] i40e: add offset to zca_free Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 05/11] ixgbe: " Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 06/11] xsk: add support to allow unaligned chunk placement Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 07/11] libbpf: add flags to umem config Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 08/11] samples/bpf: add unaligned chunks mode support to xdpsock Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 09/11] samples/bpf: add buffer recycling for unaligned chunks " Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 10/11] samples/bpf: use hugepages in xdpsock app Kevin Laatz
2019-06-20 8:39 ` [Intel-wired-lan] [PATCH 11/11] doc/af_xdp: include unaligned chunk case Kevin Laatz
2019-06-24 15:38 ` [Intel-wired-lan] [PATCH 00/11] XDP unaligned chunk placement support =?unknown-8bit?q?Bj=C3=B6rn_T=C3=B6pel?=
2019-06-25 13:12 ` Laatz, Kevin
2019-06-25 18:44 ` Jonathan Lemon
2019-06-27 11:14 ` Laatz, Kevin
2019-06-27 21:25 ` Jakub Kicinski
2019-06-28 16:19 ` Laatz, Kevin
2019-06-28 16:51 ` =?unknown-8bit?q?Bj=C3=B6rn_T=C3=B6pel?=
2019-06-28 20:08 ` Jakub Kicinski
2019-06-28 20:25 ` Jakub Kicinski
2019-06-28 20:29 ` Jonathan Lemon
2019-07-01 14:44 ` Laatz, Kevin
2019-07-01 21:20 ` Jakub Kicinski
2019-07-02 9:27 ` Richardson, Bruce
2019-07-02 16:33 ` Jonathan Lemon [this message]
2019-07-01 14:58 ` Laatz, Kevin
-- strict thread matches above, loose matches on Subject: below --
2019-06-20 9:09 Kevin Laatz
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=3510BE85-6B1B-4BB4-9640-ECEE2572DB4E@gmail.com \
--to=jonathan.lemon@gmail.com \
--cc=intel-wired-lan@osuosl.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox