From: Tariq Toukan <ttoukan.linux@gmail.com>
To: Jesper Dangaard Brouer <jbrouer@redhat.com>,
Alexei Starovoitov <ast@kernel.org>,
John Fastabend <john.fastabend@gmail.com>,
Jakub Kicinski <kuba@kernel.org>
Cc: brouer@redhat.com, Daniel Borkmann <daniel@iogearbox.net>,
Jesper Dangaard Brouer <hawk@kernel.org>,
bpf@vger.kernel.org, "David S. Miller" <davem@davemloft.net>,
Eric Dumazet <edumazet@google.com>,
netdev@vger.kernel.org, Gal Pressman <gal@nvidia.com>,
Nimrod Oren <noren@nvidia.com>, Tariq Toukan <tariqt@nvidia.com>,
Lorenzo Bianconi <lorenzo.bianconi@redhat.com>,
drosen@google.com, Joanne Koong <joannelkoong@gmail.com>,
henning.fehrmann@aei.mpg.de, oliver.behnke@aei.mpg.de
Subject: Re: [PATCH bpf-next 1/2] samples/bpf: fixup xdp_redirect tool to be able to support xdp multibuffer
Date: Tue, 30 May 2023 16:40:41 +0300 [thread overview]
Message-ID: <dc19366d-8516-9f2a-b6ed-d9323e9250c9@gmail.com> (raw)
In-Reply-To: <3168b14c-c9c1-b11b-2500-2ff2451eb81c@redhat.com>
On 30/05/2023 15:40, Jesper Dangaard Brouer wrote:
>
>
> On 30/05/2023 14.17, Tariq Toukan wrote:
>>
>> On 30/05/2023 14:33, Jesper Dangaard Brouer wrote:
>>>
>>>
>>> On 29/05/2023 13.06, Tariq Toukan wrote:
>>>> Expand the xdp multi-buffer support to xdp_redirect tool.
>>>> Similar to what's done in commit
>>>> 772251742262 ("samples/bpf: fixup some tools to be able to support
>>>> xdp multibuffer")
>>>> and its fix commit
>>>> 7a698edf954c ("samples/bpf: Fix MAC address swapping in xdp2_kern").
>>>>
>>>
>>> Have you tested if this cause a performance degradation?
>>>
>>> (Also found possible bug below)
>>>
>>
>> Hi Jesper,
>>
>> This introduces the same known perf degradation we already have in
>> xdp1 and xdp2.
>
> Did a quick test with xdp1, the performance degradation is around 18%.
>
> Before: 22,917,961 pps
> After: 18,798,336 pps
>
> (1-(18798336/22917961))*100 = 17.97%
>
>
>> Unfortunately, this is the API we have today to safely support
>> multi-buffer.
>> Note that both perf and functional (noted below) degradation should be
>> eliminated once replacing the load/store operations with dynptr logic
>> that returns a pointer to the scatter entry instead of copying it.
>>
>
> Well, should we use dynptr logic in this patch then?
>
AFAIU it's not there ready to be used...
Not sure what parts are missing, I'll need to review it a bit deeper.
> Does it make sense to add sample code that does thing in a way that is
> sub-optimal and we want to replace?
> ... (I fear people will copy paste the sample code).
>
I get your point.
As xdp1 and xdp2 are already there, I thought that we'd want to expose
multi-buffer samples in XDP_REDIRECT as well. We use these samples for
internal testing.
>> I initiated a discussion on this topic a few months ago. dynptr was
>> accepted since then, but I'm not aware of any in-progress followup
>> work that addresses this.
>>
>
> Are you saying some more work is needed on dynptr?
>
AFAIU yes.
But I might be wrong... I need to revisit this.
Do you think/know that dynptr can be used immediately?
>>>> Signed-off-by: Tariq Toukan <tariqt@nvidia.com>
>>>> Reviewed-by: Nimrod Oren <noren@nvidia.com>
>>>> ---
>>>> samples/bpf/xdp_redirect.bpf.c | 16 ++++++++++++----
>>>> 1 file changed, 12 insertions(+), 4 deletions(-)
>>>>
>>>> diff --git a/samples/bpf/xdp_redirect.bpf.c
>>>> b/samples/bpf/xdp_redirect.bpf.c
>>>> index 7c02bacfe96b..620163eb7e19 100644
>>>> --- a/samples/bpf/xdp_redirect.bpf.c
>>>> +++ b/samples/bpf/xdp_redirect.bpf.c
>>>> @@ -16,16 +16,21 @@
>>>> const volatile int ifindex_out;
>>>> -SEC("xdp")
>>>> +#define XDPBUFSIZE 64
>>>
>>> Pktgen sample scripts will default send with 60 pkt length, because the
>>> 4 bytes FCS (end-frame checksum) is added by hardware.
>>>
>>> Will this result in an error when bpf_xdp_load_bytes() tries to copy 64
>>> bytes from a 60 bytes packet?
>>>
>>
>> Yes.
>>
>> This can be resolved by reducing XDPBUFSIZE to 60.
>> Need to check if it's OK to disregard these last 4 bytes without
>> hurting the XDP program logic.
>>
>> If so, do you suggest changing xdp1 and xdp2 as well?
>>
>
> I can take care of reducing XDPBUFSIZE to 60 on xpd1 and xdp2, as I
> already had to make these changes for the above quick bench work ;-)
> I'll send out patches shortly.
>
>
Thanks.
Are we fine with the above?
Should we just change the array size to 60 and re-spin?
>>>> +SEC("xdp.frags")
>>>> int xdp_redirect_prog(struct xdp_md *ctx)
>>>> {
>>>> - void *data_end = (void *)(long)ctx->data_end;
>>>> - void *data = (void *)(long)ctx->data;
>>>> + __u8 pkt[XDPBUFSIZE] = {};
>>>> + void *data_end = &pkt[XDPBUFSIZE-1];
>>>> + void *data = pkt;
>>>> u32 key = bpf_get_smp_processor_id();
>>>> struct ethhdr *eth = data;
>>>> struct datarec *rec;
>>>> u64 nh_off;
>>>> + if (bpf_xdp_load_bytes(ctx, 0, pkt, sizeof(pkt)))
>>>> + return XDP_DROP;
>>>
>>> E.g. sizeof(pkt) = 64 bytes here.
>>>
>>>> +
>>>> nh_off = sizeof(*eth);
>>>> if (data + nh_off > data_end)
>>>> return XDP_DROP;
>>>> @@ -36,11 +41,14 @@ int xdp_redirect_prog(struct xdp_md *ctx)
>>>> NO_TEAR_INC(rec->processed);
>>>> swap_src_dst_mac(data);
>>>> + if (bpf_xdp_store_bytes(ctx, 0, pkt, sizeof(pkt)))
>>>> + return XDP_DROP;
>>>> +
>>>> return bpf_redirect(ifindex_out, 0);
>>>> }
>>>> /* Redirect require an XDP bpf_prog loaded on the TX device */
>>>> -SEC("xdp")
>>>> +SEC("xdp.frags")
>>>> int xdp_redirect_dummy_prog(struct xdp_md *ctx)
>>>> {
>>>> return XDP_PASS;
>>>
>>
>
next prev parent reply other threads:[~2023-05-30 13:40 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-05-29 11:06 [PATCH bpf-next 0/2] multi-buffer support for XDP_REDIRECT samples Tariq Toukan
2023-05-29 11:06 ` [PATCH bpf-next 1/2] samples/bpf: fixup xdp_redirect tool to be able to support xdp multibuffer Tariq Toukan
2023-05-30 11:33 ` Jesper Dangaard Brouer
2023-05-30 12:17 ` Tariq Toukan
2023-05-30 12:40 ` Jesper Dangaard Brouer
2023-05-30 13:40 ` Tariq Toukan [this message]
2023-05-30 15:05 ` Toke Høiland-Jørgensen
2023-05-30 17:22 ` Martin KaFai Lau
2023-05-31 12:02 ` Tariq Toukan
2023-05-29 11:06 ` [PATCH bpf-next 2/2] samples/bpf: fixup xdp_redirect_map " Tariq Toukan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=dc19366d-8516-9f2a-b6ed-d9323e9250c9@gmail.com \
--to=ttoukan.linux@gmail.com \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=brouer@redhat.com \
--cc=daniel@iogearbox.net \
--cc=davem@davemloft.net \
--cc=drosen@google.com \
--cc=edumazet@google.com \
--cc=gal@nvidia.com \
--cc=hawk@kernel.org \
--cc=henning.fehrmann@aei.mpg.de \
--cc=jbrouer@redhat.com \
--cc=joannelkoong@gmail.com \
--cc=john.fastabend@gmail.com \
--cc=kuba@kernel.org \
--cc=lorenzo.bianconi@redhat.com \
--cc=netdev@vger.kernel.org \
--cc=noren@nvidia.com \
--cc=oliver.behnke@aei.mpg.de \
--cc=tariqt@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).