// SPDX-License-Identifier: GPL-2.0 /* Copyright (C) 2023 SUSE LLC */ #include #include #include "../../../include/linux/filter.h" #include "bpf_misc.h" struct { __uint(type, BPF_MAP_TYPE_ARRAY); __uint(max_entries, 1); __type(key, __u32); __type(value, __u64); } precision_map SEC(".maps"); SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 3: (bf) r1 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 2: (55) if r2 != 0xfffffff8 goto pc+2") __msg("mark_precise: frame0: regs=r2 stack= before 1: (87) r2 = -r2") __msg("mark_precise: frame0: regs=r2 stack= before 0: (b7) r2 = 8") __naked int bpf_neg(void) { asm volatile ( "r2 = 8;" "r2 = -r2;" "if r2 != -8 goto 1f;" "r1 = r10;" "r1 += r2;" "1:" "r0 = 0;" "exit;" ::: __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 3: (bf) r1 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 2: (55) if r2 != 0x0 goto pc+2") __msg("mark_precise: frame0: regs=r2 stack= before 1: (d4) r2 = le16 r2") __msg("mark_precise: frame0: regs=r2 stack= before 0: (b7) r2 = 0") __naked int bpf_end_to_le(void) { asm volatile ( "r2 = 0;" "r2 = le16 r2;" "if r2 != 0 goto 1f;" "r1 = r10;" "r1 += r2;" "1:" "r0 = 0;" "exit;" ::: __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 3: (bf) r1 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 2: (55) if r2 != 0x0 goto pc+2") __msg("mark_precise: frame0: regs=r2 stack= before 1: (dc) r2 = be16 r2") __msg("mark_precise: frame0: regs=r2 stack= before 0: (b7) r2 = 0") __naked int bpf_end_to_be(void) { asm volatile ( "r2 = 0;" "r2 = be16 r2;" "if r2 != 0 goto 1f;" "r1 = r10;" "r1 += r2;" "1:" "r0 = 0;" "exit;" ::: __clobber_all); } #if (defined(__TARGET_ARCH_arm64) || defined(__TARGET_ARCH_x86) || \ (defined(__TARGET_ARCH_riscv) && __riscv_xlen == 64) || \ defined(__TARGET_ARCH_arm) || defined(__TARGET_ARCH_s390)) && \ __clang_major__ >= 18 SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 3: (bf) r1 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 2: (55) if r2 != 0x0 goto pc+2") __msg("mark_precise: frame0: regs=r2 stack= before 1: (d7) r2 = bswap16 r2") __msg("mark_precise: frame0: regs=r2 stack= before 0: (b7) r2 = 0") __naked int bpf_end_bswap(void) { asm volatile ( "r2 = 0;" "r2 = bswap16 r2;" "if r2 != 0 goto 1f;" "r1 = r10;" "r1 += r2;" "1:" "r0 = 0;" "exit;" ::: __clobber_all); } #ifdef CAN_USE_LOAD_ACQ_STORE_REL SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 3: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 2: (db) r2 = load_acquire((u64 *)(r10 -8))") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_load_acquire(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" ".8byte %[load_acquire_insn];" /* r2 = load_acquire((u64 *)(r10 - 8)); */ "r3 = r10;" "r3 += r2;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(load_acquire_insn, BPF_ATOMIC_OP(BPF_DW, BPF_LOAD_ACQ, BPF_REG_2, BPF_REG_10, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r1 stack= before 3: (bf) r2 = r10") __msg("mark_precise: frame0: regs=r1 stack= before 2: (79) r1 = *(u64 *)(r10 -8)") __msg("mark_precise: frame0: regs= stack=-8 before 1: (db) store_release((u64 *)(r10 -8), r1)") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_store_release(void) { asm volatile ( "r1 = 8;" ".8byte %[store_release_insn];" /* store_release((u64 *)(r10 - 8), r1); */ "r1 = *(u64 *)(r10 - 8);" "r2 = r10;" "r2 += r1;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(store_release_insn, BPF_ATOMIC_OP(BPF_DW, BPF_STORE_REL, BPF_REG_10, BPF_REG_1, -8)) : __clobber_all); } #endif /* CAN_USE_LOAD_ACQ_STORE_REL */ #endif /* v4 instruction */ SEC("?raw_tp") __success __log_level(2) /* * Without the bug fix there will be no history between "last_idx 3 first_idx 3" * and "parent state regs=" lines. "R0=6" parts are here to help anchor * expected log messages to the one specific mark_chain_precision operation. * * This is quite fragile: if verifier checkpointing heuristic changes, this * might need adjusting. */ __msg("2: (07) r0 += 1 ; R0=6") __msg("3: (35) if r0 >= 0xa goto pc+1") __msg("mark_precise: frame0: last_idx 3 first_idx 3 subseq_idx -1") __msg("mark_precise: frame0: regs=r0 stack= before 2: (07) r0 += 1") __msg("mark_precise: frame0: regs=r0 stack= before 1: (07) r0 += 1") __msg("mark_precise: frame0: regs=r0 stack= before 4: (05) goto pc-4") __msg("mark_precise: frame0: regs=r0 stack= before 3: (35) if r0 >= 0xa goto pc+1") __msg("mark_precise: frame0: parent state regs= stack=: R0=P4") __msg("3: R0=6") __naked int state_loop_first_last_equal(void) { asm volatile ( "r0 = 0;" "l0_%=:" "r0 += 1;" "r0 += 1;" /* every few iterations we'll have a checkpoint here with * first_idx == last_idx, potentially confusing precision * backtracking logic */ "if r0 >= 10 goto l1_%=;" /* checkpoint + mark_precise */ "goto l0_%=;" "l1_%=:" "exit;" ::: __clobber_common ); } __used __naked static void __bpf_cond_op_r10(void) { asm volatile ( "r2 = 2314885393468386424 ll;" "goto +0;" "if r2 <= r10 goto +3;" "if r1 >= -1835016 goto +0;" "if r2 <= 8 goto +0;" "if r3 <= 0 goto +0;" "exit;" ::: __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("8: (bd) if r2 <= r10 goto pc+3") __msg("9: (35) if r1 >= 0xffe3fff8 goto pc+0") __msg("10: (b5) if r2 <= 0x8 goto pc+0") __msg("mark_precise: frame1: last_idx 10 first_idx 0 subseq_idx -1") __msg("mark_precise: frame1: regs=r2 stack= before 9: (35) if r1 >= 0xffe3fff8 goto pc+0") __msg("mark_precise: frame1: regs=r2 stack= before 8: (bd) if r2 <= r10 goto pc+3") __msg("mark_precise: frame1: regs=r2 stack= before 7: (05) goto pc+0") __naked void bpf_cond_op_r10(void) { asm volatile ( "r3 = 0 ll;" "call __bpf_cond_op_r10;" "r0 = 0;" "exit;" ::: __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("3: (bf) r3 = r10") __msg("4: (bd) if r3 <= r2 goto pc+1") __msg("5: (b5) if r2 <= 0x8 goto pc+2") __msg("mark_precise: frame0: last_idx 5 first_idx 0 subseq_idx -1") __msg("mark_precise: frame0: regs=r2 stack= before 4: (bd) if r3 <= r2 goto pc+1") __msg("mark_precise: frame0: regs=r2 stack= before 3: (bf) r3 = r10") __naked void bpf_cond_op_not_r10(void) { asm volatile ( "r0 = 0;" "r2 = 2314885393468386424 ll;" "r3 = r10;" "if r3 <= r2 goto +1;" "if r2 <= 8 goto +2;" "r0 = 2 ll;" "exit;" ::: __clobber_all); } SEC("lsm.s/socket_connect") __success __log_level(2) __msg("0: (b7) r0 = 1 ; R0=1") __msg("1: (84) w0 = -w0 ; R0=0xffffffff") __msg("mark_precise: frame0: last_idx 2 first_idx 0 subseq_idx -1") __msg("mark_precise: frame0: regs=r0 stack= before 1: (84) w0 = -w0") __msg("mark_precise: frame0: regs=r0 stack= before 0: (b7) r0 = 1") __naked int bpf_neg_2(void) { /* * lsm.s/socket_connect requires a return value within [-4095, 0]. * Returning -1 is allowed */ asm volatile ( "r0 = 1;" "w0 = -w0;" "exit;" ::: __clobber_all); } SEC("lsm.s/socket_connect") __failure __msg("At program exit the register R0 has") __naked int bpf_neg_3(void) { /* * lsm.s/socket_connect requires a return value within [-4095, 0]. * Returning -10000 is not allowed. */ asm volatile ( "r0 = 10000;" "w0 = -w0;" "exit;" ::: __clobber_all); } SEC("lsm.s/socket_connect") __success __log_level(2) __msg("0: (b7) r0 = 1 ; R0=1") __msg("1: (87) r0 = -r0 ; R0=-1") __msg("mark_precise: frame0: last_idx 2 first_idx 0 subseq_idx -1") __msg("mark_precise: frame0: regs=r0 stack= before 1: (87) r0 = -r0") __msg("mark_precise: frame0: regs=r0 stack= before 0: (b7) r0 = 1") __naked int bpf_neg_4(void) { /* * lsm.s/socket_connect requires a return value within [-4095, 0]. * Returning -1 is allowed */ asm volatile ( "r0 = 1;" "r0 = -r0;" "exit;" ::: __clobber_all); } SEC("lsm.s/socket_connect") __failure __msg("At program exit the register R0 has") __naked int bpf_neg_5(void) { /* * lsm.s/socket_connect requires a return value within [-4095, 0]. * Returning -10000 is not allowed. */ asm volatile ( "r0 = 10000;" "r0 = -r0;" "exit;" ::: __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 4: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 3: (db) r2 = atomic64_fetch_add((u64 *)(r10 -8), r2)") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_fetch_add_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r2 = 0;" ".8byte %[fetch_add_insn];" /* r2 = atomic_fetch_add(*(u64 *)(r10 - 8), r2) */ "r3 = r10;" "r3 += r2;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(fetch_add_insn, BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 4: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 3: (db) r2 = atomic64_xchg((u64 *)(r10 -8), r2)") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_xchg_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r2 = 0;" ".8byte %[xchg_insn];" /* r2 = atomic_xchg(*(u64 *)(r10 - 8), r2) */ "r3 = r10;" "r3 += r2;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(xchg_insn, BPF_ATOMIC_OP(BPF_DW, BPF_XCHG, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 4: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 3: (db) r2 = atomic64_fetch_or((u64 *)(r10 -8), r2)") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_fetch_or_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r2 = 0;" ".8byte %[fetch_or_insn];" /* r2 = atomic_fetch_or(*(u64 *)(r10 - 8), r2) */ "r3 = r10;" "r3 += r2;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(fetch_or_insn, BPF_ATOMIC_OP(BPF_DW, BPF_OR | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 4: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 3: (db) r2 = atomic64_fetch_and((u64 *)(r10 -8), r2)") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_fetch_and_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r2 = 0;" ".8byte %[fetch_and_insn];" /* r2 = atomic_fetch_and(*(u64 *)(r10 - 8), r2) */ "r3 = r10;" "r3 += r2;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(fetch_and_insn, BPF_ATOMIC_OP(BPF_DW, BPF_AND | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2 stack= before 4: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r2 stack= before 3: (db) r2 = atomic64_fetch_xor((u64 *)(r10 -8), r2)") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_fetch_xor_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r2 = 0;" ".8byte %[fetch_xor_insn];" /* r2 = atomic_fetch_xor(*(u64 *)(r10 - 8), r2) */ "r3 = r10;" "r3 += r2;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(fetch_xor_insn, BPF_ATOMIC_OP(BPF_DW, BPF_XOR | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r0 stack= before 5: (bf) r3 = r10") __msg("mark_precise: frame0: regs=r0 stack= before 4: (db) r0 = atomic64_cmpxchg((u64 *)(r10 -8), r0, r2)") __msg("mark_precise: frame0: regs= stack=-8 before 3: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r0 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_cmpxchg_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r0 = 0;" "r2 = 0;" ".8byte %[cmpxchg_insn];" /* r0 = atomic_cmpxchg(*(u64 *)(r10 - 8), r0, r2) */ "r3 = r10;" "r3 += r0;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(cmpxchg_insn, BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } /* Regression test for dual precision: Both the fetched value (r2) and * a reread of the same stack slot (r3) are tracked for precision. After * the atomic operation, the stack slot is STACK_MISC. Thus, the ldx at * insn 4 does NOT set INSN_F_STACK_ACCESS. Precision for the stack slot * propagates solely through the atomic fetch's load side (insn 3). */ SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r2,r3 stack= before 4: (79) r3 = *(u64 *)(r10 -8)") __msg("mark_precise: frame0: regs=r2 stack= before 3: (db) r2 = atomic64_fetch_add((u64 *)(r10 -8), r2)") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_fetch_add_dual_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r2 = 0;" ".8byte %[fetch_add_insn];" /* r2 = atomic_fetch_add(*(u64 *)(r10 - 8), r2) */ "r3 = *(u64 *)(r10 - 8);" "r4 = r2;" "r4 += r3;" "r4 &= 7;" "r5 = r10;" "r5 += r4;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(fetch_add_insn, BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r0,r3 stack= before 5: (79) r3 = *(u64 *)(r10 -8)") __msg("mark_precise: frame0: regs=r0 stack= before 4: (db) r0 = atomic64_cmpxchg((u64 *)(r10 -8), r0, r2)") __msg("mark_precise: frame0: regs= stack=-8 before 3: (b7) r2 = 0") __msg("mark_precise: frame0: regs= stack=-8 before 2: (b7) r0 = 8") __msg("mark_precise: frame0: regs= stack=-8 before 1: (7b) *(u64 *)(r10 -8) = r1") __msg("mark_precise: frame0: regs=r1 stack= before 0: (b7) r1 = 8") __naked int bpf_atomic_cmpxchg_dual_precision(void) { asm volatile ( "r1 = 8;" "*(u64 *)(r10 - 8) = r1;" "r0 = 8;" "r2 = 0;" ".8byte %[cmpxchg_insn];" /* r0 = atomic_cmpxchg(*(u64 *)(r10 - 8), r0, r2) */ "r3 = *(u64 *)(r10 - 8);" "r4 = r0;" "r4 += r3;" "r4 &= 7;" "r5 = r10;" "r5 += r4;" /* mark_precise */ "r0 = 0;" "exit;" : : __imm_insn(cmpxchg_insn, BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_10, BPF_REG_2, -8)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r1 stack= before 10: (57) r1 &= 7") __msg("mark_precise: frame0: regs=r1 stack= before 9: (db) r1 = atomic64_fetch_add((u64 *)(r0 +0), r1)") __not_msg("falling back to forcing all scalars precise") __naked int bpf_atomic_fetch_add_map_precision(void) { asm volatile ( "r1 = 0;" "*(u64 *)(r10 - 8) = r1;" "r2 = r10;" "r2 += -8;" "r1 = %[precision_map] ll;" "call %[bpf_map_lookup_elem];" "if r0 == 0 goto 1f;" "r1 = 0;" ".8byte %[fetch_add_insn];" /* r1 = atomic_fetch_add(*(u64 *)(r0 + 0), r1) */ "r1 &= 7;" "r2 = r10;" "r2 += r1;" /* mark_precise */ "1: r0 = 0;" "exit;" : : __imm_addr(precision_map), __imm(bpf_map_lookup_elem), __imm_insn(fetch_add_insn, BPF_ATOMIC_OP(BPF_DW, BPF_ADD | BPF_FETCH, BPF_REG_0, BPF_REG_1, 0)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r0 stack= before 12: (57) r0 &= 7") __msg("mark_precise: frame0: regs=r0 stack= before 11: (db) r0 = atomic64_cmpxchg((u64 *)(r6 +0), r0, r1)") __not_msg("falling back to forcing all scalars precise") __naked int bpf_atomic_cmpxchg_map_precision(void) { asm volatile ( "r1 = 0;" "*(u64 *)(r10 - 8) = r1;" "r2 = r10;" "r2 += -8;" "r1 = %[precision_map] ll;" "call %[bpf_map_lookup_elem];" "if r0 == 0 goto 1f;" "r6 = r0;" "r0 = 0;" "r1 = 0;" ".8byte %[cmpxchg_insn];" /* r0 = atomic_cmpxchg(*(u64 *)(r6 + 0), r0, r1) */ "r0 &= 7;" "r2 = r10;" "r2 += r0;" /* mark_precise */ "1: r0 = 0;" "exit;" : : __imm_addr(precision_map), __imm(bpf_map_lookup_elem), __imm_insn(cmpxchg_insn, BPF_ATOMIC_OP(BPF_DW, BPF_CMPXCHG, BPF_REG_6, BPF_REG_1, 0)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r1 stack= before 10: (57) r1 &= 7") __msg("mark_precise: frame0: regs=r1 stack= before 9: (c3) r1 = atomic_fetch_add((u32 *)(r0 +0), r1)") __not_msg("falling back to forcing all scalars precise") __naked int bpf_atomic_fetch_add_32bit_precision(void) { asm volatile ( "r1 = 0;" "*(u64 *)(r10 - 8) = r1;" "r2 = r10;" "r2 += -8;" "r1 = %[precision_map] ll;" "call %[bpf_map_lookup_elem];" "if r0 == 0 goto 1f;" "r1 = 0;" ".8byte %[fetch_add_insn];" /* r1 = atomic_fetch_add(*(u32 *)(r0 + 0), r1) */ "r1 &= 7;" "r2 = r10;" "r2 += r1;" /* mark_precise */ "1: r0 = 0;" "exit;" : : __imm_addr(precision_map), __imm(bpf_map_lookup_elem), __imm_insn(fetch_add_insn, BPF_ATOMIC_OP(BPF_W, BPF_ADD | BPF_FETCH, BPF_REG_0, BPF_REG_1, 0)) : __clobber_all); } SEC("?raw_tp") __success __log_level(2) __msg("mark_precise: frame0: regs=r0 stack= before 12: (57) r0 &= 7") __msg("mark_precise: frame0: regs=r0 stack= before 11: (c3) r0 = atomic_cmpxchg((u32 *)(r6 +0), r0, r1)") __not_msg("falling back to forcing all scalars precise") __naked int bpf_atomic_cmpxchg_32bit_precision(void) { asm volatile ( "r1 = 0;" "*(u64 *)(r10 - 8) = r1;" "r2 = r10;" "r2 += -8;" "r1 = %[precision_map] ll;" "call %[bpf_map_lookup_elem];" "if r0 == 0 goto 1f;" "r6 = r0;" "r0 = 0;" "r1 = 0;" ".8byte %[cmpxchg_insn];" /* r0 = atomic_cmpxchg(*(u32 *)(r6 + 0), r0, r1) */ "r0 &= 7;" "r2 = r10;" "r2 += r0;" /* mark_precise */ "1: r0 = 0;" "exit;" : : __imm_addr(precision_map), __imm(bpf_map_lookup_elem), __imm_insn(cmpxchg_insn, BPF_ATOMIC_OP(BPF_W, BPF_CMPXCHG, BPF_REG_6, BPF_REG_1, 0)) : __clobber_all); } char _license[] SEC("license") = "GPL";