From: Eduard Zingerman <eddyz87@gmail.com>
To: Dimitar Kanaliev <dimitar.kanaliev@siteground.com>, bpf@vger.kernel.org
Cc: Alexei Starovoitov <ast@kernel.org>,
Daniel Borkmann <daniel@iogearbox.net>,
John Fastabend <john.fastabend@gmail.com>,
Andrii Nakryiko <andrii@kernel.org>,
Martin KaFai Lau <martin.lau@linux.dev>,
Song Liu <song@kernel.org>,
Yonghong Song <yonghong.song@linux.dev>,
KP Singh <kpsingh@kernel.org>,
Stanislav Fomichev <sdf@fomichev.me>,
Hao Luo <haoluo@google.com>, Jiri Olsa <jolsa@kernel.org>,
Mykola Lysenko <mykolal@fb.com>,
Shung-Hsi Yu <shung-hsi.yu@suse.com>
Subject: Re: [PATCH v1 2/3] bpf: verifier: Simplify register sign extension with tnum_scast
Date: Mon, 01 Dec 2025 15:49:54 -0800 [thread overview]
Message-ID: <cad6577291b778e6caad2f06fae304b2ec07f752.camel@gmail.com> (raw)
In-Reply-To: <20251125125634.2671-3-dimitar.kanaliev@siteground.com>
On Tue, 2025-11-25 at 14:56 +0200, Dimitar Kanaliev wrote:
[...]
> diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c
> index 766695491bc5..c9a6bf85b4ad 100644
> --- a/kernel/bpf/verifier.c
> +++ b/kernel/bpf/verifier.c
> @@ -6876,147 +6876,57 @@ static void coerce_reg_to_size(struct bpf_reg_state *reg, int size)
> reg_bounds_sync(reg);
> }
>
> -static void set_sext64_default_val(struct bpf_reg_state *reg, int size)
> -{
> - if (size == 1) {
> - reg->smin_value = reg->s32_min_value = S8_MIN;
> - reg->smax_value = reg->s32_max_value = S8_MAX;
> - } else if (size == 2) {
> - reg->smin_value = reg->s32_min_value = S16_MIN;
> - reg->smax_value = reg->s32_max_value = S16_MAX;
> - } else {
> - /* size == 4 */
> - reg->smin_value = reg->s32_min_value = S32_MIN;
> - reg->smax_value = reg->s32_max_value = S32_MAX;
> - }
> - reg->umin_value = reg->u32_min_value = 0;
> - reg->umax_value = U64_MAX;
> - reg->u32_max_value = U32_MAX;
> - reg->var_off = tnum_unknown;
> -}
> -
> static void coerce_reg_to_size_sx(struct bpf_reg_state *reg, int size)
> {
> - s64 init_s64_max, init_s64_min, s64_max, s64_min, u64_cval;
> - u64 top_smax_value, top_smin_value;
> - u64 num_bits = size * 8;
> + s64 smin_value, smax_value;
>
> - if (tnum_is_const(reg->var_off)) {
> - u64_cval = reg->var_off.value;
> - if (size == 1)
> - reg->var_off = tnum_const((s8)u64_cval);
> - else if (size == 2)
> - reg->var_off = tnum_const((s16)u64_cval);
> - else
> - /* size == 4 */
> - reg->var_off = tnum_const((s32)u64_cval);
> -
> - u64_cval = reg->var_off.value;
> - reg->smax_value = reg->smin_value = u64_cval;
> - reg->umax_value = reg->umin_value = u64_cval;
> - reg->s32_max_value = reg->s32_min_value = u64_cval;
> - reg->u32_max_value = reg->u32_min_value = u64_cval;
> + if (size >= 8)
> return;
> - }
>
> - top_smax_value = ((u64)reg->smax_value >> num_bits) << num_bits;
> - top_smin_value = ((u64)reg->smin_value >> num_bits) << num_bits;
> + reg->var_off = tnum_scast(reg->var_off, size);
>
> - if (top_smax_value != top_smin_value)
> - goto out;
> + smin_value = -(1LL << (size * 8 - 1));
> + smax_value = (1LL << (size * 8 - 1)) - 1;
>
> - /* find the s64_min and s64_min after sign extension */
> - if (size == 1) {
> - init_s64_max = (s8)reg->smax_value;
> - init_s64_min = (s8)reg->smin_value;
> - } else if (size == 2) {
> - init_s64_max = (s16)reg->smax_value;
> - init_s64_min = (s16)reg->smin_value;
> - } else {
> - init_s64_max = (s32)reg->smax_value;
> - init_s64_min = (s32)reg->smin_value;
> - }
> -
> - s64_max = max(init_s64_max, init_s64_min);
> - s64_min = min(init_s64_max, init_s64_min);
> + reg->smin_value = smin_value;
> + reg->smax_value = smax_value;
>
> - /* both of s64_max/s64_min positive or negative */
> - if ((s64_max >= 0) == (s64_min >= 0)) {
> - reg->s32_min_value = reg->smin_value = s64_min;
> - reg->s32_max_value = reg->smax_value = s64_max;
> - reg->u32_min_value = reg->umin_value = s64_min;
> - reg->u32_max_value = reg->umax_value = s64_max;
> - reg->var_off = tnum_range(s64_min, s64_max);
> - return;
> - }
> + reg->s32_min_value = (s32)smin_value;
> + reg->s32_max_value = (s32)smax_value;
>
> -out:
> - set_sext64_default_val(reg, size);
> -}
Assume that size == 1, s64_min = 0b000, s64_max == 0b100.
This corresponds to tnum with value == 0b000 and mask == 0b111.
Old algorithm computes more precise range in this situation.
Old:
0: (85) call bpf_get_prandom_u32#7 ; R0=scalar()
1: (25) if r0 > 0x4 goto pc+2 ; R0=scalar(smin=smin32=0,smax=umax=smax32=umax32=4,var_off=(0x0; 0x7))
2: (7b) *(u64 *)(r10 -8) = r0 ; R0=scalar(id=1,smin=smin32=0,smax=umax=smax32=umax32=4,var_off=(0x0; 0x7)) ...
3: (91) r0 = *(s8 *)(r10 -8) ; R0=scalar(id=1,smin=smin32=0,smax=umax=smax32=umax32=4,var_off=(0x0; 0x7)) ...
4: (b7) r0 = 0 ; R0=0
5: (95) exit
New:
0: (85) call bpf_get_prandom_u32#7 ; R0=scalar()
1: (25) if r0 > 0x4 goto pc+2 ; R0=scalar(smin=smin32=0,smax=umax=smax32=umax32=4,var_off=(0x0; 0x7))
2: (7b) *(u64 *)(r10 -8) = r0 ; R0=scalar(id=1,smin=smin32=0,smax=umax=smax32=umax32=4,var_off=(0x0; 0x7)) ...
3: (91) r0 = *(s8 *)(r10 -8) ; R0=scalar(id=1,smin=smin32=0,smax=umax=smax32=umax32=7,var_off=(0x0; 0x7)) ...
4: (b7) r0 = 0 ; R0=0
5: (95) exit
Note that range for R0 at (3) is 0..4 for old algorithm and 0..7 for
new algorithm.
Can we keep both algorithms by e.g. replacing set_sext64_default_val()
implementation with tnum_scast() adding tnum_scast() in
coerce_reg_to_size_sx()?
In general, for such kinds of patch-sets it is interesting to see how
much precision is gained/lost with the change. It shouldn't be hard to
collect such data for e.g. complete s8 range by writing a small
user-space program that enumerates the s8 x s8 range and applies both
old an new range computations.
[...]
next prev parent reply other threads:[~2025-12-01 23:49 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-11-25 12:56 [PATCH v1 0/3] Add tnum_scast helper Dimitar Kanaliev
2025-11-25 12:56 ` [PATCH v1 1/3] bpf: Introduce tnum_scast as a tnum native sign extension helper Dimitar Kanaliev
2025-11-25 13:22 ` bot+bpf-ci
2025-11-26 8:56 ` Shung-Hsi Yu
2025-12-01 7:43 ` Dimitar Kanaliev
2025-12-15 2:40 ` Shung-Hsi Yu
2025-11-25 12:56 ` [PATCH v1 2/3] bpf: verifier: Simplify register sign extension with tnum_scast Dimitar Kanaliev
2025-11-25 13:22 ` bot+bpf-ci
2025-11-26 10:32 ` Dimitar Kanaliev
2025-12-01 23:49 ` Eduard Zingerman [this message]
2025-12-02 10:53 ` Dimitar Kanaliev
2025-12-02 18:03 ` Eduard Zingerman
2025-12-04 6:50 ` Shung-Hsi Yu
2025-11-25 12:56 ` [PATCH v1 3/3] selftests/bpf: Add verifier bounds checks for sign extension Dimitar Kanaliev
2025-11-26 9:04 ` [PATCH v1 0/3] Add tnum_scast helper Shung-Hsi Yu
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=cad6577291b778e6caad2f06fae304b2ec07f752.camel@gmail.com \
--to=eddyz87@gmail.com \
--cc=andrii@kernel.org \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=dimitar.kanaliev@siteground.com \
--cc=haoluo@google.com \
--cc=john.fastabend@gmail.com \
--cc=jolsa@kernel.org \
--cc=kpsingh@kernel.org \
--cc=martin.lau@linux.dev \
--cc=mykolal@fb.com \
--cc=sdf@fomichev.me \
--cc=shung-hsi.yu@suse.com \
--cc=song@kernel.org \
--cc=yonghong.song@linux.dev \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox