* [PATCH bpf-next 1/3] riscv, bpf: Fix support for BPF_SDIV and BPF_SMOD in RV32 JIT
2026-04-29 12:35 [PATCH bpf-next 0/3] riscv, bpf: Fix signed operations and add 32 bit atomics Kuan-Wei Chiu
@ 2026-04-29 12:35 ` Kuan-Wei Chiu
2026-04-29 13:13 ` bot+bpf-ci
2026-04-29 12:35 ` [PATCH bpf-next 2/3] riscv, bpf: Fix support for BPF_MOVSX " Kuan-Wei Chiu
2026-04-29 12:35 ` [PATCH bpf-next 3/3] riscv, bpf: Add 32 bit atomic operations to " Kuan-Wei Chiu
2 siblings, 1 reply; 7+ messages in thread
From: Kuan-Wei Chiu @ 2026-04-29 12:35 UTC (permalink / raw)
To: ast, daniel, andrii, eddyz87, memxor, luke.r.nels, xi.wang, pjw,
palmer, aou
Cc: martin.lau, song, yonghong.song, jolsa, alex, jserv, eleanor15x,
marscheng, bpf, linux-riscv, linux-kernel, Kuan-Wei Chiu
The current rv32 bpf jit compiler incorrectly treats BPF_SDIV and
BPF_SMOD as unsigned operations. The BPF instruction set allows
signed division and modulo by reusing the BPF_DIV and BPF_MOD opcodes
with the instruction offset set to 1.
Update the emit_alu_r32() function to accept an 'is_sdiv' variable and
emit the correct div and rem instructions when the offset is 1.
Before this patch:
[ 44.161771] test_bpf: #165 ALU_SDIV_X: -6 / 2 = -3 jited:1 ret 2147483645 != -3 (0x7ffffffd != 0xfffffffd)FAIL (1 times)
[ 44.167385] test_bpf: #166 ALU_SDIV_K: -6 / 2 = -3 jited:1 ret 2147483645 != -3 (0x7ffffffd != 0xfffffffd)FAIL (1 times)
[ 44.171053] test_bpf: #169 ALU_SMOD_X: -7 % 2 = -1 jited:1 ret 1 != -1 (0x1 != 0xffffffff)FAIL (1 times)
[ 44.172081] test_bpf: #170 ALU_SMOD_K: -7 % 2 = -1 jited:1 ret 1 != -1 (0x1 != 0xffffffff)FAIL (1 times)
After this patch:
[ 16.002192] test_bpf: #165 ALU_SDIV_X: -6 / 2 = -3 jited:1 95 PASS
[ 16.002983] test_bpf: #166 ALU_SDIV_K: -6 / 2 = -3 jited:1 1059 PASS
[ 16.017167] test_bpf: #169 ALU_SMOD_X: -7 % 2 = -1 jited:1 136 PASS
[ 16.023002] test_bpf: #170 ALU_SMOD_K: -7 % 2 = -1 jited:1 109 PASS
Signed-off-by: Kuan-Wei Chiu <visitorckw@gmail.com>
---
arch/riscv/net/bpf_jit_comp32.c | 19 ++++++++++++++-----
1 file changed, 14 insertions(+), 5 deletions(-)
diff --git a/arch/riscv/net/bpf_jit_comp32.c b/arch/riscv/net/bpf_jit_comp32.c
index 592dd86fbf81..7396899ea276 100644
--- a/arch/riscv/net/bpf_jit_comp32.c
+++ b/arch/riscv/net/bpf_jit_comp32.c
@@ -509,7 +509,7 @@ static void emit_alu_r64(const s8 *dst, const s8 *src,
}
static void emit_alu_r32(const s8 *dst, const s8 *src,
- struct rv_jit_context *ctx, const u8 op)
+ struct rv_jit_context *ctx, const u8 op, bool is_sdiv)
{
const s8 *tmp1 = bpf2rv32[TMP_REG_1];
const s8 *tmp2 = bpf2rv32[TMP_REG_2];
@@ -539,10 +539,16 @@ static void emit_alu_r32(const s8 *dst, const s8 *src,
emit(rv_mul(lo(rd), lo(rd), lo(rs)), ctx);
break;
case BPF_DIV:
- emit(rv_divu(lo(rd), lo(rd), lo(rs)), ctx);
+ if (is_sdiv)
+ emit(rv_div(lo(rd), lo(rd), lo(rs)), ctx);
+ else
+ emit(rv_divu(lo(rd), lo(rd), lo(rs)), ctx);
break;
case BPF_MOD:
- emit(rv_remu(lo(rd), lo(rd), lo(rs)), ctx);
+ if (is_sdiv)
+ emit(rv_rem(lo(rd), lo(rd), lo(rs)), ctx);
+ else
+ emit(rv_remu(lo(rd), lo(rd), lo(rs)), ctx);
break;
case BPF_LSH:
emit(rv_sll(lo(rd), lo(rd), lo(rs)), ctx);
@@ -959,6 +965,7 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
u8 code = insn->code;
s16 off = insn->off;
s32 imm = insn->imm;
+ bool is_sdiv = false;
const s8 *dst = bpf2rv32[insn->dst_reg];
const s8 *src = bpf2rv32[insn->src_reg];
@@ -1041,7 +1048,9 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
emit_imm32(tmp2, imm, ctx);
src = tmp2;
}
- emit_alu_r32(dst, src, ctx, BPF_OP(code));
+ if ((BPF_OP(code) == BPF_DIV || BPF_OP(code) == BPF_MOD) && insn->off == 1)
+ is_sdiv = true;
+ emit_alu_r32(dst, src, ctx, BPF_OP(code), is_sdiv);
break;
case BPF_ALU | BPF_MOV | BPF_K:
@@ -1065,7 +1074,7 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
* src is ignored---choose tmp2 as a dummy register since it
* is not on the stack.
*/
- emit_alu_r32(dst, tmp2, ctx, BPF_OP(code));
+ emit_alu_r32(dst, tmp2, ctx, BPF_OP(code), false);
break;
case BPF_ALU | BPF_END | BPF_FROM_LE:
--
2.54.0.545.g6539524ca2-goog
^ permalink raw reply related [flat|nested] 7+ messages in thread* Re: [PATCH bpf-next 1/3] riscv, bpf: Fix support for BPF_SDIV and BPF_SMOD in RV32 JIT
2026-04-29 12:35 ` [PATCH bpf-next 1/3] riscv, bpf: Fix support for BPF_SDIV and BPF_SMOD in RV32 JIT Kuan-Wei Chiu
@ 2026-04-29 13:13 ` bot+bpf-ci
0 siblings, 0 replies; 7+ messages in thread
From: bot+bpf-ci @ 2026-04-29 13:13 UTC (permalink / raw)
To: visitorckw, ast, daniel, andrii, eddyz87, memxor, luke.r.nels,
xi.wang, pjw, palmer, aou
Cc: martin.lau, song, yonghong.song, jolsa, alex, jserv, eleanor15x,
marscheng, bpf, linux-riscv, linux-kernel, visitorckw, ast,
andrii, daniel, martin.lau, eddyz87, yonghong.song, clm,
ihor.solodrai
[-- Attachment #1: Type: text/plain, Size: 2485 bytes --]
> diff --git a/arch/riscv/net/bpf_jit_comp32.c b/arch/riscv/net/bpf_jit_comp32.c
> index 592dd86fbf81..7396899ea276 100644
> --- a/arch/riscv/net/bpf_jit_comp32.c
> +++ b/arch/riscv/net/bpf_jit_comp32.c
> @@ -509,7 +509,7 @@ static void emit_alu_r64(const s8 *dst, const s8 *src,
> }
>
> static void emit_alu_r32(const s8 *dst, const s8 *src,
> - struct rv_jit_context *ctx, const u8 op)
> + struct rv_jit_context *ctx, const u8 op, bool is_sdiv)
> {
> const s8 *tmp1 = bpf2rv32[TMP_REG_1];
> const s8 *tmp2 = bpf2rv32[TMP_REG_2];
> @@ -539,10 +539,16 @@ static void emit_alu_r32(const s8 *dst, const s8 *src,
> emit(rv_mul(lo(rd), lo(rd), lo(rs)), ctx);
> break;
> case BPF_DIV:
> - emit(rv_divu(lo(rd), lo(rd), lo(rs)), ctx);
> + if (is_sdiv)
> + emit(rv_div(lo(rd), lo(rd), lo(rs)), ctx);
> + else
> + emit(rv_divu(lo(rd), lo(rd), lo(rs)), ctx);
> break;
> case BPF_MOD:
> - emit(rv_remu(lo(rd), lo(rd), lo(rs)), ctx);
> + if (is_sdiv)
> + emit(rv_rem(lo(rd), lo(rd), lo(rs)), ctx);
> + else
> + emit(rv_remu(lo(rd), lo(rd), lo(rs)), ctx);
> break;
> case BPF_LSH:
> emit(rv_sll(lo(rd), lo(rd), lo(rs)), ctx);
> @@ -959,6 +965,7 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
> u8 code = insn->code;
> s16 off = insn->off;
> s32 imm = insn->imm;
> + bool is_sdiv = false;
>
> const s8 *dst = bpf2rv32[insn->dst_reg];
> const s8 *src = bpf2rv32[insn->src_reg];
> @@ -1041,7 +1048,9 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
> emit_imm32(tmp2, imm, ctx);
> src = tmp2;
> }
> - emit_alu_r32(dst, src, ctx, BPF_OP(code));
> + if ((BPF_OP(code) == BPF_DIV || BPF_OP(code) == BPF_MOD) && insn->off == 1)
> + is_sdiv = true;
> + emit_alu_r32(dst, src, ctx, BPF_OP(code), is_sdiv);
> break;
>
> case BPF_ALU | BPF_MOV | BPF_K:
> @@ -1065,7 +1074,7 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
> * src is ignored---choose tmp2 as a dummy register since it
> * is not on the stack.
> */
> - emit_alu_r32(dst, tmp2, ctx, BPF_OP(code));
> + emit_alu_r32(dst, tmp2, ctx, BPF_OP(code), false);
> break;
>
> case BPF_ALU | BPF_END | BPF_FROM_LE:
---
AI reviewed your patch. Please fix the bug or email reply why it's not a bug.
See: https://github.com/kernel-patches/vmtest/blob/master/ci/claude/README.md
CI run summary: https://github.com/kernel-patches/bpf/actions/runs/25109627626
^ permalink raw reply [flat|nested] 7+ messages in thread
* [PATCH bpf-next 2/3] riscv, bpf: Fix support for BPF_MOVSX in RV32 JIT
2026-04-29 12:35 [PATCH bpf-next 0/3] riscv, bpf: Fix signed operations and add 32 bit atomics Kuan-Wei Chiu
2026-04-29 12:35 ` [PATCH bpf-next 1/3] riscv, bpf: Fix support for BPF_SDIV and BPF_SMOD in RV32 JIT Kuan-Wei Chiu
@ 2026-04-29 12:35 ` Kuan-Wei Chiu
2026-04-29 13:13 ` bot+bpf-ci
2026-04-29 12:35 ` [PATCH bpf-next 3/3] riscv, bpf: Add 32 bit atomic operations to " Kuan-Wei Chiu
2 siblings, 1 reply; 7+ messages in thread
From: Kuan-Wei Chiu @ 2026-04-29 12:35 UTC (permalink / raw)
To: ast, daniel, andrii, eddyz87, memxor, luke.r.nels, xi.wang, pjw,
palmer, aou
Cc: martin.lau, song, yonghong.song, jolsa, alex, jserv, eleanor15x,
marscheng, bpf, linux-riscv, linux-kernel, Kuan-Wei Chiu
The current rv32 bpf jit compiler incorrectly treats BPF_MOVSX as a
standard zero-extended move operation. The bpf instruction set allows
sign-extension moves by reusing the BPF_MOV opcode with the instruction
offset set to 8, 16, or 32.
Update the bpf_jit_emit_insn() function to check the offset field for
both ALU and ALU64 MOV operations. If the offset is non-zero, emit the
correct slli and srai instructions to perform the sign extension.
Before this patch:
[ 19.549705] test_bpf: #82 ALU_MOVSX | BPF_B jited:1 ret 2 != 1 (0x2 != 0x1)FAIL (1 times)
[ 19.551354] test_bpf: #83 ALU_MOVSX | BPF_H jited:1 ret 2 != 1 (0x2 != 0x1)FAIL (1 times)
[ 19.552576] test_bpf: #84 ALU64_MOVSX | BPF_B jited:1 ret 2 != 1 (0x2 != 0x1)FAIL (1 times)
[ 19.553542] test_bpf: #85 ALU64_MOVSX | BPF_H jited:1 ret 2 != 1 (0x2 != 0x1)FAIL (1 times)
[ 19.554807] test_bpf: #86 ALU64_MOVSX | BPF_W jited:1 ret 2 != 1 (0x2 != 0x1)FAIL (1 times)
After this patch:
[ 17.931172] test_bpf: #82 ALU_MOVSX | BPF_B jited:1 125 PASS
[ 17.932198] test_bpf: #83 ALU_MOVSX | BPF_H jited:1 124 PASS
[ 17.933039] test_bpf: #84 ALU64_MOVSX | BPF_B jited:1 124 PASS
[ 17.933918] test_bpf: #85 ALU64_MOVSX | BPF_H jited:1 124 PASS
[ 17.934751] test_bpf: #86 ALU64_MOVSX | BPF_W jited:1 122 PASS
Signed-off-by: Kuan-Wei Chiu <visitorckw@gmail.com>
---
arch/riscv/net/bpf_jit_comp32.c | 32 ++++++++++++++++++++++++++++++++
1 file changed, 32 insertions(+)
diff --git a/arch/riscv/net/bpf_jit_comp32.c b/arch/riscv/net/bpf_jit_comp32.c
index 7396899ea276..f8509950fed4 100644
--- a/arch/riscv/net/bpf_jit_comp32.c
+++ b/arch/riscv/net/bpf_jit_comp32.c
@@ -974,6 +974,24 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
switch (code) {
case BPF_ALU64 | BPF_MOV | BPF_X:
+ if (insn->off != 0) {
+ const s8 *rd = bpf_get_reg64(dst, tmp1, ctx);
+ const s8 *rs = bpf_get_reg64(src, tmp2, ctx);
+
+ if (insn->off == 8) {
+ emit(rv_slli(lo(rd), lo(rs), 24), ctx);
+ emit(rv_srai(lo(rd), lo(rd), 24), ctx);
+ } else if (insn->off == 16) {
+ emit(rv_slli(lo(rd), lo(rs), 16), ctx);
+ emit(rv_srai(lo(rd), lo(rd), 16), ctx);
+ } else {
+ emit(rv_addi(lo(rd), lo(rs), 0), ctx);
+ }
+ emit(rv_srai(hi(rd), lo(rd), 31), ctx);
+ bpf_put_reg64(dst, rd, ctx);
+ break;
+ }
+ fallthrough;
case BPF_ALU64 | BPF_ADD | BPF_X:
case BPF_ALU64 | BPF_ADD | BPF_K:
@@ -1024,6 +1042,20 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
emit_zext64(dst, ctx);
break;
}
+ if (insn->off != 0) {
+ const s8 *rd = bpf_get_reg32(dst, tmp1, ctx);
+ const s8 *rs = bpf_get_reg32(src, tmp2, ctx);
+
+ if (insn->off == 8) {
+ emit(rv_slli(lo(rd), lo(rs), 24), ctx);
+ emit(rv_srai(lo(rd), lo(rd), 24), ctx);
+ } else if (insn->off == 16) {
+ emit(rv_slli(lo(rd), lo(rs), 16), ctx);
+ emit(rv_srai(lo(rd), lo(rd), 16), ctx);
+ }
+ bpf_put_reg32(dst, rd, ctx);
+ break;
+ }
fallthrough;
case BPF_ALU | BPF_ADD | BPF_X:
--
2.54.0.545.g6539524ca2-goog
^ permalink raw reply related [flat|nested] 7+ messages in thread* Re: [PATCH bpf-next 2/3] riscv, bpf: Fix support for BPF_MOVSX in RV32 JIT
2026-04-29 12:35 ` [PATCH bpf-next 2/3] riscv, bpf: Fix support for BPF_MOVSX " Kuan-Wei Chiu
@ 2026-04-29 13:13 ` bot+bpf-ci
0 siblings, 0 replies; 7+ messages in thread
From: bot+bpf-ci @ 2026-04-29 13:13 UTC (permalink / raw)
To: visitorckw, ast, daniel, andrii, eddyz87, memxor, luke.r.nels,
xi.wang, pjw, palmer, aou
Cc: martin.lau, song, yonghong.song, jolsa, alex, jserv, eleanor15x,
marscheng, bpf, linux-riscv, linux-kernel, visitorckw, ast,
andrii, daniel, martin.lau, eddyz87, yonghong.song, clm,
ihor.solodrai
[-- Attachment #1: Type: text/plain, Size: 608 bytes --]
> riscv, bpf: Fix support for BPF_MOVSX in RV32 JIT
The subject indicates this is a fix, but there's no Fixes: tag in the
commit message. Should this include:
Fixes: 8100928c8814 ("bpf: Support new sign-extension mov insns")
The RV32 BPF JIT never implemented support for BPF_MOVSX instructions,
which were added to the BPF instruction set in commit 8100928c8814.
---
AI reviewed your patch. Please fix the bug or email reply why it's not a bug.
See: https://github.com/kernel-patches/vmtest/blob/master/ci/claude/README.md
CI run summary: https://github.com/kernel-patches/bpf/actions/runs/25109627626
^ permalink raw reply [flat|nested] 7+ messages in thread
* [PATCH bpf-next 3/3] riscv, bpf: Add 32 bit atomic operations to RV32 JIT
2026-04-29 12:35 [PATCH bpf-next 0/3] riscv, bpf: Fix signed operations and add 32 bit atomics Kuan-Wei Chiu
2026-04-29 12:35 ` [PATCH bpf-next 1/3] riscv, bpf: Fix support for BPF_SDIV and BPF_SMOD in RV32 JIT Kuan-Wei Chiu
2026-04-29 12:35 ` [PATCH bpf-next 2/3] riscv, bpf: Fix support for BPF_MOVSX " Kuan-Wei Chiu
@ 2026-04-29 12:35 ` Kuan-Wei Chiu
2026-04-29 13:13 ` bot+bpf-ci
2 siblings, 1 reply; 7+ messages in thread
From: Kuan-Wei Chiu @ 2026-04-29 12:35 UTC (permalink / raw)
To: ast, daniel, andrii, eddyz87, memxor, luke.r.nels, xi.wang, pjw,
palmer, aou
Cc: martin.lau, song, yonghong.song, jolsa, alex, jserv, eleanor15x,
marscheng, bpf, linux-riscv, linux-kernel, Kuan-Wei Chiu
The RV32 BPF JIT compiler currently only supports the BPF_ADD atomic
operation. Other 32 bit atomic operations (and, or, xor, xchg) and
their BPF_FETCH variants are not supported and gracefully fall back to
the interpreter.
Since the RISC-V A extension is required for Linux on RV32, we can
natively support these 32-bit BPF atomic operations by mapping them
directly to the corresponding RISC-V amo*.w instructions.
Implement BPF_ADD, BPF_AND, BPF_OR, BPF_XOR, and BPF_XCHG with and
without BPF_FETCH. BPF_CMPXCHG requires a more complex lr.w/sc.w
loop and is left to fall back to the interpreter.
Before this patch:
[ 138.862161] test_bpf: Summary: 1054 PASSED, 0 FAILED, [843/1042 JIT'ed]
After this patch:
[ 157.024124] test_bpf: Summary: 1054 PASSED, 0 FAILED, [902/1042 JIT'ed]
Signed-off-by: Kuan-Wei Chiu <visitorckw@gmail.com>
---
arch/riscv/net/bpf_jit_comp32.c | 47 +++++++++++++++++++++++++--------
1 file changed, 36 insertions(+), 11 deletions(-)
diff --git a/arch/riscv/net/bpf_jit_comp32.c b/arch/riscv/net/bpf_jit_comp32.c
index f8509950fed4..710c56b20f98 100644
--- a/arch/riscv/net/bpf_jit_comp32.c
+++ b/arch/riscv/net/bpf_jit_comp32.c
@@ -877,7 +877,7 @@ static int emit_load_r64(const s8 *dst, const s8 *src, s16 off,
static int emit_store_r64(const s8 *dst, const s8 *src, s16 off,
struct rv_jit_context *ctx, const u8 size,
- const u8 mode)
+ const u8 mode, s32 imm)
{
const s8 *tmp1 = bpf2rv32[TMP_REG_1];
const s8 *tmp2 = bpf2rv32[TMP_REG_2];
@@ -902,11 +902,40 @@ static int emit_store_r64(const s8 *dst, const s8 *src, s16 off,
case BPF_MEM:
emit(rv_sw(RV_REG_T0, 0, lo(rs)), ctx);
break;
- case BPF_ATOMIC: /* Only BPF_ADD supported */
- emit(rv_amoadd_w(RV_REG_ZERO, lo(rs), RV_REG_T0, 0, 0),
- ctx);
+ case BPF_ATOMIC:
+ {
+ s8 fetch_reg = (imm & BPF_FETCH) ? lo(rs) : RV_REG_ZERO;
+
+ switch (imm) {
+ case BPF_ADD:
+ case BPF_ADD | BPF_FETCH:
+ emit(rv_amoadd_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
+ break;
+ case BPF_AND:
+ case BPF_AND | BPF_FETCH:
+ emit(rv_amoand_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
+ break;
+ case BPF_OR:
+ case BPF_OR | BPF_FETCH:
+ emit(rv_amoor_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
+ break;
+ case BPF_XOR:
+ case BPF_XOR | BPF_FETCH:
+ emit(rv_amoxor_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
+ break;
+ case BPF_XCHG:
+ emit(rv_amoswap_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
+ break;
+ default:
+ return -1;
+ }
+ if (imm & BPF_FETCH) {
+ emit(rv_addi(hi(rs), RV_REG_ZERO, 0), ctx);
+ bpf_put_reg64(src, rs, ctx);
+ }
break;
}
+ }
break;
case BPF_DW:
emit(rv_sw(RV_REG_T0, 0, lo(rs)), ctx);
@@ -1308,20 +1337,16 @@ int bpf_jit_emit_insn(const struct bpf_insn *insn, struct rv_jit_context *ctx,
}
if (emit_store_r64(dst, src, off, ctx, BPF_SIZE(code),
- BPF_MODE(code)))
+ BPF_MODE(code), 0))
return -1;
break;
case BPF_STX | BPF_ATOMIC | BPF_W:
- if (insn->imm != BPF_ADD) {
- pr_info_once(
- "bpf-jit: not supported: atomic operation %02x ***\n",
- insn->imm);
+ if (insn->imm == BPF_CMPXCHG)
return -EFAULT;
- }
if (emit_store_r64(dst, src, off, ctx, BPF_SIZE(code),
- BPF_MODE(code)))
+ BPF_MODE(code), insn->imm))
return -1;
break;
--
2.54.0.545.g6539524ca2-goog
^ permalink raw reply related [flat|nested] 7+ messages in thread* Re: [PATCH bpf-next 3/3] riscv, bpf: Add 32 bit atomic operations to RV32 JIT
2026-04-29 12:35 ` [PATCH bpf-next 3/3] riscv, bpf: Add 32 bit atomic operations to " Kuan-Wei Chiu
@ 2026-04-29 13:13 ` bot+bpf-ci
0 siblings, 0 replies; 7+ messages in thread
From: bot+bpf-ci @ 2026-04-29 13:13 UTC (permalink / raw)
To: visitorckw, ast, daniel, andrii, eddyz87, memxor, luke.r.nels,
xi.wang, pjw, palmer, aou
Cc: martin.lau, song, yonghong.song, jolsa, alex, jserv, eleanor15x,
marscheng, bpf, linux-riscv, linux-kernel, visitorckw, ast,
andrii, daniel, martin.lau, eddyz87, yonghong.song, clm,
ihor.solodrai
[-- Attachment #1: Type: text/plain, Size: 2730 bytes --]
> diff --git a/arch/riscv/net/bpf_jit_comp32.c b/arch/riscv/net/bpf_jit_comp32.c
> index f8509950fed4..710c56b20f98 100644
> --- a/arch/riscv/net/bpf_jit_comp32.c
> +++ b/arch/riscv/net/bpf_jit_comp32.c
[ ... ]
> @@ -902,11 +902,40 @@ static int emit_store_r64(const s8 *dst, const s8 *src, s16 off,
> case BPF_MEM:
> emit(rv_sw(RV_REG_T0, 0, lo(rs)), ctx);
> break;
> - case BPF_ATOMIC: /* Only BPF_ADD supported */
> - emit(rv_amoadd_w(RV_REG_ZERO, lo(rs), RV_REG_T0, 0, 0),
> - ctx);
> + case BPF_ATOMIC:
> + {
> + s8 fetch_reg = (imm & BPF_FETCH) ? lo(rs) : RV_REG_ZERO;
> +
> + switch (imm) {
> + case BPF_ADD:
> + case BPF_ADD | BPF_FETCH:
> + emit(rv_amoadd_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
> + break;
> + case BPF_AND:
> + case BPF_AND | BPF_FETCH:
> + emit(rv_amoand_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
> + break;
> + case BPF_OR:
> + case BPF_OR | BPF_FETCH:
> + emit(rv_amoor_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
> + break;
> + case BPF_XOR:
> + case BPF_XOR | BPF_FETCH:
> + emit(rv_amoxor_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
> + break;
> + case BPF_XCHG:
> + emit(rv_amoswap_w(fetch_reg, lo(rs), RV_REG_T0, 0, 0), ctx);
> + break;
Do the aq=0, rl=0 arguments provide the correct memory ordering for the
FETCH variants and XCHG?
Looking at the BPF specification in include/linux/filter.h, these BPF
atomic operations map to Linux kernel atomic_fetch_* and atomic_xchg
primitives, which require full memory ordering.
The RV64 BPF JIT in arch/riscv/net/bpf_jit_comp64.c emits these same
operations with aq=1, rl=1 for FETCH variants and XCHG:
case BPF_ADD | BPF_FETCH:
emit(is64 ? rv_amoadd_d(rs, rs, rd, 1, 1) :
rv_amoadd_w(rs, rs, rd, 1, 1), ctx);
case BPF_XCHG:
emit(is64 ? rv_amoswap_d(rs, rs, rd, 1, 1) :
rv_amoswap_w(rs, rs, rd, 1, 1), ctx);
The kernel's arch_atomic_fetch_add in arch/riscv/include/asm/atomic.h also
uses amoadd.w.aqrl (which corresponds to aq=1, rl=1).
With aq=0, rl=0, the atomic operations use relaxed memory ordering, which
could allow loads and stores to move across the atomic. Could this cause
BPF programs using atomic_fetch_*/atomic_xchg on RV32 to observe
memory-ordering violations that wouldn't occur on other architectures?
Note: For the non-FETCH variants of BPF_ADD/AND/OR/XOR (the return-void
operations), aq=0, rl=0 appears correct since those map to relaxed
atomic primitives.
---
AI reviewed your patch. Please fix the bug or email reply why it's not a bug.
See: https://github.com/kernel-patches/vmtest/blob/master/ci/claude/README.md
CI run summary: https://github.com/kernel-patches/bpf/actions/runs/25109627626
^ permalink raw reply [flat|nested] 7+ messages in thread