From: Zhu Yanjun <yanjun.zhu@linux.dev>
To: Jakub Sitnicki <jakub@cloudflare.com>, lsf-pc@lists.linux-foundation.org
Cc: bpf@vger.kernel.org, kernel-team@cloudflare.com
Subject: Re: [LSF/MM/BPF TOPIC] BPF local storage for every packet
Date: Tue, 3 Mar 2026 07:06:46 -0800 [thread overview]
Message-ID: <aa2fada2-3f61-4f3d-9f55-36538f566333@linux.dev> (raw)
In-Reply-To: <87ecmffopy.fsf@cloudflare.com>
在 2026/2/20 6:56, Jakub Sitnicki 写道:
> In the upcoming days we are going to post an RFC which proposes to
> extend the concept of BPF local storage to socket buffers (sk_buff, skb)
> as means to attach arbitrary metadata to packets from BPF programs [1]
> (slides 41-55).
>
> Design wise, BPF local storage is a great fit for a packet metadata
> container, as it that avoids some of the shortcoming of the the XDP
> metadata interface:
>
> 1. Users interact with storage through BPF maps and can take advantage
> of existing built-in BPF map types, while still being able to
> implement a custom data format,
>
> 2. Maps within local storage can have different properties controlled by
> map flags. For example, maps with BPF_F_CLONE set can survive packet
> cloning. Other flags could allow map contents to survive sk_buff
> scrubbing during encapsulation/decapsulation or pass across network
> namespace boundaries.
>
> 3. Local storage supports multiple users out of the box - each user
> creates their own map, eliminating the need to coordinate data
> layout,
>
> 4. Local storage has its own backing memory, so persisting it across
> network stack layers requires no changes to the network stack.
>
> However, this flexibility comes at a cost. While XDP metadata requires
> no allocations [2], an initial write to BPF local storage requires two:
> one for bpf_local_storage_elem, and one for bpf_local_storage itself.
>
> We would like to align this work with the needs of other BPF local
> storage users (socks, cgroups, tasks, inodes), where allocation overhead
> has been a concern as well [2].
>
> Optimization ideas we would like to put up for discussion:
> - slimming down bpf_local_storage so it can be embedded as an skb
> extension chunk,
Interested in this topic. I hope to join this meeting.
Zhu Yanjun
> - making the bpf_local_storage cache size configurable,
> - allowing bpf_local_storage to be pre-allocated,
> - co-allocating bpf_local_storage and bpf_local_storage_elem for the
> single-map case.
>
> Thanks,
> -jkbs
>
> [1] https://fosdem.org/2026/schedule/event/DSC9L3-rich-packet-metadata/
> [2] Assuming sufficient free headroom in the skb linear buffer.
> [3] http://msgid.link/ad835a9b-e544-48d3-b6e2-ffe172fcfa6d@linux.dev
next prev parent reply other threads:[~2026-03-03 15:06 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-02-20 14:56 [LSF/MM/BPF TOPIC] BPF local storage for every packet Jakub Sitnicki
2026-02-20 18:34 ` Martin KaFai Lau
2026-02-21 13:42 ` Jakub Sitnicki
2026-02-23 19:26 ` Martin KaFai Lau
2026-02-24 11:58 ` Jakub Sitnicki
2026-03-03 15:06 ` Zhu Yanjun [this message]
2026-03-03 21:07 ` Jakub Sitnicki
2026-03-16 3:02 ` Zhu Yanjun
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aa2fada2-3f61-4f3d-9f55-36538f566333@linux.dev \
--to=yanjun.zhu@linux.dev \
--cc=bpf@vger.kernel.org \
--cc=jakub@cloudflare.com \
--cc=kernel-team@cloudflare.com \
--cc=lsf-pc@lists.linux-foundation.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox