bpf.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Yonghong Song <yonghong.song@linux.dev>
To: Anton Protopopov <a.s.protopopov@gmail.com>
Cc: bpf@vger.kernel.org, Alexei Starovoitov <ast@kernel.org>,
	Andrii Nakryiko <andrii@kernel.org>,
	Anton Protopopov <aspsk@isovalent.com>,
	Daniel Borkmann <daniel@iogearbox.net>,
	Eduard Zingerman <eddyz87@gmail.com>,
	Quentin Monnet <qmo@kernel.org>
Subject: Re: [PATCH v1 bpf-next 00/11] BPF indirect jumps
Date: Fri, 5 Sep 2025 08:39:02 -0700	[thread overview]
Message-ID: <c3f6482e-fce0-4d9d-a1dd-53bafa901fe0@linux.dev> (raw)
In-Reply-To: <aLqdTM9CL+eOdWBs@mail.gmail.com>



On 9/5/25 1:20 AM, Anton Protopopov wrote:
> On 25/09/04 01:27PM, Yonghong Song wrote:
>>
>> On 8/16/25 11:06 AM, Anton Protopopov wrote:
>>> This patchset implements a new type of map, instruction set, and uses
>>> it to build support for indirect branches in BPF (on x86). (The same
>>> map will be later used to provide support for indirect calls and static
>>> keys.) See [1], [2] for more context.
>>>
>>> This patch set is a follow-up on the initial RFC [3], now converted to
>>> normal version to trigger CI. Note that GCC and non-x86 archs are not
>>> supposed to work.
>>>
>>> Short table of contents:
>>>
>>>     * Patches 1-6 implement the new map of type
>>>       BPF_MAP_TYPE_INSN_SET and corresponding selftests. This map can
>>>       be used to track the "original -> xlated -> jitted mapping" for
>>>       a given program. Patches 5,6 add support for "blinded" variant.
>>>
>>>     * Patches 7,8,9 implement the support for indirect jumps
>>>
>>>     * Patches 10,11 add support for LLVM-compiled programs containing
>>>       indirect jumps.
>>>
>>> A special LLVM should be used for that, see [4] for the details and
>>> some related discussions. Due to this fact, selftests for indirect
>>> jumps which directly use `goto *rX` are commented out (such that
>>> CI can run).
>>>
>>> There is a list of TBDs (mostly, more selftests + some limitations
>>> like maximal map size), however, all the selftests which compile
>>> to contain an indirect jump work with this patchset.
>>>
>>> See individual patches for more details on implementation details.
>>>
>>> Changes since RFC:
>>>
>>>     * I've tried to address all the comments provided by Alexei and
>>>       Eduard in RFC. Will try to list the most important of them below.
>>>
>>>     * One big change: move from older LLVM version [5] to newer [4].
>>>       Now LLVM generates jump tables as symbols in the new special
>>>       section ".jumptables". Another part of this change is that
>>>       libbpf now doesn't try to link map load and goto *rX, as
>>>       1) this is absolutely not reliable 2) for some use cases this
>>>       is impossible (namely, when more than one jump table can be used
>>>       in the same gotox instruction).
>>>
>>>     * Added insn_successors() support (Alexei, Eduard). This includes
>>>       getting rid of the ugly bpf_insn_set_iter_xlated_offset()
>>>       interface (Eduard).
>>>
>>>     * Removed hack for the unreachable instruction, as new LLVM thank to
>>>       Eduard doesn't generate it.
>>>
>>>     * Set mem_size for direct map access properly instead of hacking.
>>>       Remove off>0 check. (Alexei)
>>>
>>>     * Do not allocate new memory for min_index/max_index (Alexei, Eduard)
>>>
>>>     * Information required during check_cfg is now cached to be reused
>>>       later (Alexei + general logic for supporting multiple JT per jump)
>>>
>>>     * Properly compare registers in regsafe (Alexei, Eduard)
>>>
>>>     * Remove support for JMP32 (Eduard)
>>>
>>>     * Better checks in adjust_ptr_min_max_vals (Eduard)
>>>
>>>     * More selftests were added (but still there's room for more) which
>>>       directly use gotox (Alexei)
>>>
>>>     * More checks and verbose messages added
>>>
>>>     * "unique pointers" are no more in the map
>>>
>>> Links:
>>>     1. https://lpc.events/event/18/contributions/1941/
>>>     2. https://lwn.net/Articles/1017439/
>>>     3. https://lore.kernel.org/bpf/20250615085943.3871208-1-a.s.protopopov@gmail.com/
>>>     4. https://github.com/llvm/llvm-project/pull/149715
>>>
>>> Anton Protopopov (11):
>>>     bpf: fix the return value of push_stack
>>>     bpf: save the start of functions in bpf_prog_aux
>>>     bpf, x86: add new map type: instructions array
>>>     selftests/bpf: add selftests for new insn_array map
>>>     bpf: support instructions arrays with constants blinding
>>>     selftests/bpf: test instructions arrays with blinding
>>>     bpf, x86: allow indirect jumps to r8...r15
>>>     bpf, x86: add support for indirect jumps
>>>     bpf: disasm: add support for BPF_JMP|BPF_JA|BPF_X
>>>     libbpf: support llvm-generated indirect jumps
>>>     selftests/bpf: add selftests for indirect jumps
>>>
>>>    arch/x86/net/bpf_jit_comp.c                   |  39 +-
>>>    include/linux/bpf.h                           |  30 +
>>>    include/linux/bpf_types.h                     |   1 +
>>>    include/linux/bpf_verifier.h                  |  20 +-
>>>    include/uapi/linux/bpf.h                      |  11 +
>>>    kernel/bpf/Makefile                           |   2 +-
>>>    kernel/bpf/bpf_insn_array.c                   | 350 ++++++++++
>>>    kernel/bpf/core.c                             |  20 +
>>>    kernel/bpf/disasm.c                           |   9 +
>>>    kernel/bpf/syscall.c                          |  22 +
>>>    kernel/bpf/verifier.c                         | 603 ++++++++++++++++--
>>>    .../bpf/bpftool/Documentation/bpftool-map.rst |   2 +-
>>>    tools/bpf/bpftool/map.c                       |   2 +-
>>>    tools/include/uapi/linux/bpf.h                |  11 +
>>>    tools/lib/bpf/libbpf.c                        | 159 ++++-
>>>    tools/lib/bpf/libbpf_probes.c                 |   4 +
>>>    tools/lib/bpf/linker.c                        |  12 +-
>>>    tools/testing/selftests/bpf/Makefile          |   4 +-
>>>    .../selftests/bpf/prog_tests/bpf_goto_x.c     | 132 ++++
>>>    .../selftests/bpf/prog_tests/bpf_insn_array.c | 498 +++++++++++++++
>>>    .../testing/selftests/bpf/progs/bpf_goto_x.c  | 384 +++++++++++
>>>    21 files changed, 2230 insertions(+), 85 deletions(-)
>>>    create mode 100644 kernel/bpf/bpf_insn_array.c
>>>    create mode 100644 tools/testing/selftests/bpf/prog_tests/bpf_goto_x.c
>>>    create mode 100644 tools/testing/selftests/bpf/prog_tests/bpf_insn_array.c
>>>    create mode 100644 tools/testing/selftests/bpf/progs/bpf_goto_x.c
>>>
>> After indirect jumps, the next natural steps will be supporting callx
>> and static key in bpf programs.
>>
>> For static keys, currently, llvm supports gotol_or_nop/nop_or_gotol insns
>> (https://github.com/llvm/llvm-project/compare/main...aspsk:llvm-project:static-keys)
>> and these insns can only be used in inline asm.
>>
>> For callx, there are two patterns, one is calling a particular func with
>> flow sensitive analysis, another is calling through call stable.
>>
>> The following two examples are to call a partuclar func with current
>> variable to tracing register.
>>
>> Example 1:
>>
>> typedef int (*op_t)(int, int); static int add(int a, int b) { return a + b;
>> } static int mul(int a, int b) { return a * b; } static int apply(op_t f,
>> int a, int b) { // indirect call via function pointer return f(a, b); } int
>> result(int i, int j) { op_t f; if (i + j) f = add; else f = mul; return
>> apply(f, i, j); } The asm code:
>> result:                                 # @result
>> # %bb.0:
>>          w4 = w2
>>          w4 = -w4
>>          r3 = mul ll
>>          if w1 == w4 goto LBB0_2
>> # %bb.1:
>>          r3 = add ll
>> LBB0_2:
>>          callx r3
>>          exit
>>
>> Example 2:
>>
>> typedef int (*op_t)(int, int);
>>
>> __attribute__((section("_add"))) static int add(int a, int b) { return a + b; }
>> __attribute__((section("_mul"))) static int mul(int a, int b) { return a * b; }
>>
>> struct ctx {
>>    op_t f;
>> };
>>
>> __attribute__((noinline)) static int apply(struct ctx *ctx, int a, int b) {
>>      // indirect call via function pointer
>>      return ctx->f(a, b);
>> }
>>
>> int result(int i, int j) {
>>      int x = 2, y = 3;
>>      struct ctx ctx;
>>
>>      if (i&2) ctx.f = add;
>>      else ctx.f = mul;
>>      int r1 = apply(&ctx, x, y);
>>      int r2 = apply(&ctx, x, y);
>>
>>      return r1 + r2;
>> }
>>
>> asm code:
>>
>> result:                                 # @result
>> # %bb.0:
>>          w1 &= 2
>>          r2 = mul ll
>>          if w1 == 0 goto LBB0_2
>> # %bb.1:
>>          r2 = add ll
>> LBB0_2:
>>          *(u64 *)(r10 - 8) = r2
>>          r6 = r10
>>          r6 += -8
>>          r1 = r6
>>          call apply
>>          w7 = w0
>>          r1 = r6
>>          call apply
>>          w0 += w7
>>          exit
>> ...
>> apply:                                  # @apply
>> # %bb.0:
>>          r3 = *(u64 *)(r1 + 0)
>>          w1 = 2
>>          w2 = 3
>>          callx r3
>>          exit
>>
>> In the above two cases, current verifier can be enhanced to
>> track functions and eventuall 'callx r3' can find proper
>> targets.
>>
>> Another pattern is to have a calltable (similar to jump table)
>> and callx will call one of functions based on calltable base
>> and an index. The example is below:
>>
>> typedef int (*op_t)(int, int);
>>
>> __attribute__((section("_add"))) static int add(int a, int b) { return a + b; }
>> __attribute__((section("_mul"))) static int mul(int a, int b) { return a * b; }
>>
>> __attribute__((noinline)) static int apply(op_t *ops, int index, int a, int b) {
>>      // indirect call via function pointer
>>      return ops[index](a, b);
>> }
>>
>> int result(int i, int j) {
>>      op_t ops[] = { add, mul, add, add, mul, mul };
>>      int x = 2, y = 3;
>>
>>      int r1 = apply(ops, 0, x, y);
>>      int r2 = apply(ops, 4, x, y);
>>
>>      return r1 + r2;
>> }
>>
>> int result2(int i, int j) {
>>      op_t ops[] = { add, add, add, mul, mul };
>>      int x = 3, y = 2;
>>
>>      int r1 = apply(ops, 1, x, y);
>>      int r2 = apply(ops, 2, x, y);
>>
>>      return r1 + r2;
>> }
>>
>>
>> The related llvm IR:
>>
>> @__const.result.ops = private unnamed_addr constant [6 x ptr] [ptr @add, ptr @mul, ptr @add, ptr @add, ptr @mul, ptr @mul], align 8
>> @__const.result2.ops = private unnamed_addr constant [5 x ptr] [ptr @add, ptr @add, ptr @add, ptr @mul, ptr @mul], align 8
>>
>> ; Function Attrs: nounwind
>> define dso_local i32 @result(i32 noundef %0, i32 noundef %1) local_unnamed_addr #0 {
>>    %3 = tail call fastcc i32 @apply(ptr noundef @__const.result.ops, i32 noundef 0, i32 noundef 2, i32 noundef 3)
>>    %4 = tail call fastcc i32 @apply(ptr noundef @__const.result.ops, i32 noundef 4, i32 noundef 2, i32 noundef 3)
>>    %5 = add nsw i32 %4, %3
>>    ret i32 %5
>> }
>> ...
>> ; Function Attrs: noinline nounwind
>> define internal fastcc i32 @apply(ptr noundef nonnull readonly captures(none) %0, i32 noundef range(i32 0, 5) %1, i32 noundef range(i32 2, 4) %2, i32 noundef range(i32 2, 4) %3) unnamed_addr #2 {
>>    %5 = zext nneg i32 %1 to i64
>>    %6 = getelementptr inbounds nuw ptr, ptr %0, i64 %5
>>    %7 = load ptr, ptr %6, align 8, !tbaa !3
>>    %8 = tail call i32 %7(i32 noundef %2, i32 noundef %3) #3
>>    ret i32 %8
>> }
>>
>> ; Function Attrs: nounwind
>> define dso_local i32 @result2(i32 noundef %0, i32 noundef %1) local_unnamed_addr #0 {
>>    %3 = tail call fastcc i32 @apply(ptr noundef @__const.result2.ops, i32 noundef 1, i32 noundef 3, i32 noundef 2)
>>    %4 = tail call fastcc i32 @apply(ptr noundef @__const.result2.ops, i32 noundef 2, i32 noundef 3, i32 noundef 2)
>>    %5 = add nsw i32 %4, %3
>>    ret i32 %5
>> }
>>
>> To make
>>     @__const.result.ops = private unnamed_addr constant [6 x ptr] [ptr @add, ptr @mul, ptr @add, ptr @add, ptr @mul, ptr @mul], align 8
>> explicit for call table, the llvm changed the call table name (with a llvm hack) to be like
>>     BPF.__const.result.ops
>>
>> The llvm hack on top of https://github.com/llvm/llvm-project/pull/149715:
> Thanks for the details Yonghong! I am planning to work on this after
> indirect jumps. At the moment I have a simple change to the verifier
> which allows to do `callx rx` for any rx having type PTR_TO_FUNCTION
> (see https://github.com/aspsk/bpf-next/tree/wip/indirect-calls).

Sounds good. My above hack is only for compiler generated globals (private linkage).
I will handle static and external globals as well. Will post it to llvm-project
once indirect jump pull request lands.

>
> Also, I've recently rebased static keys branch on top of the current
> indirect jumps implementation, the one main thing left is to address
> Andrii's comments and implement "global", i.e., per object, static
> keys, such that any program in the object can use it. (This would be
> a map of maps containing pointers to instruction arrays, which must
> be unique per program.)
>
Sound good. Thanks!


      reply	other threads:[~2025-09-05 15:39 UTC|newest]

Thread overview: 41+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-08-16 18:06 [PATCH v1 bpf-next 00/11] BPF indirect jumps Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 01/11] bpf: fix the return value of push_stack Anton Protopopov
2025-08-25 18:12   ` Eduard Zingerman
2025-08-26 15:00     ` Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 02/11] bpf: save the start of functions in bpf_prog_aux Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 03/11] bpf, x86: add new map type: instructions array Anton Protopopov
2025-08-25 21:05   ` Eduard Zingerman
2025-08-26 15:52     ` Anton Protopopov
2025-08-26 16:04       ` Eduard Zingerman
2025-08-16 18:06 ` [PATCH v1 bpf-next 04/11] selftests/bpf: add selftests for new insn_array map Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 05/11] bpf: support instructions arrays with constants blinding Anton Protopopov
2025-08-17  5:50   ` kernel test robot
2025-08-18  8:24     ` Anton Protopopov
2025-08-25 23:29   ` Eduard Zingerman
2025-08-27  9:20     ` Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 06/11] selftests/bpf: test instructions arrays with blinding Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 07/11] bpf, x86: allow indirect jumps to r8...r15 Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 08/11] bpf, x86: add support for indirect jumps Anton Protopopov
2025-08-18  7:57   ` Dan Carpenter
2025-08-18  8:22     ` Anton Protopopov
2025-08-25 23:15   ` Eduard Zingerman
2025-08-27 15:34     ` Anton Protopopov
2025-08-27 18:58       ` Eduard Zingerman
2025-08-28  9:58     ` Anton Protopopov
2025-08-28 14:15       ` Anton Protopopov
2025-08-28 16:10         ` Eduard Zingerman
2025-08-28 16:30       ` Eduard Zingerman
2025-08-16 18:06 ` [PATCH v1 bpf-next 09/11] bpf: disasm: add support for BPF_JMP|BPF_JA|BPF_X Anton Protopopov
2025-08-16 18:06 ` [PATCH v1 bpf-next 10/11] libbpf: support llvm-generated indirect jumps Anton Protopopov
2025-08-21  0:20   ` Andrii Nakryiko
2025-08-21 13:05     ` Anton Protopopov
2025-08-21 18:14       ` Andrii Nakryiko
2025-08-21 19:12         ` Anton Protopopov
2025-08-26  0:06   ` Eduard Zingerman
2025-08-26 16:15     ` Anton Protopopov
2025-08-26 16:51       ` Anton Protopopov
2025-08-26 16:47         ` Eduard Zingerman
2025-08-16 18:06 ` [PATCH v1 bpf-next 11/11] selftests/bpf: add selftests for " Anton Protopopov
2025-09-04 20:27 ` [PATCH v1 bpf-next 00/11] BPF " Yonghong Song
2025-09-05  8:20   ` Anton Protopopov
2025-09-05 15:39     ` Yonghong Song [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=c3f6482e-fce0-4d9d-a1dd-53bafa901fe0@linux.dev \
    --to=yonghong.song@linux.dev \
    --cc=a.s.protopopov@gmail.com \
    --cc=andrii@kernel.org \
    --cc=aspsk@isovalent.com \
    --cc=ast@kernel.org \
    --cc=bpf@vger.kernel.org \
    --cc=daniel@iogearbox.net \
    --cc=eddyz87@gmail.com \
    --cc=qmo@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).