| /linux/fs/xfs/scrub/ |
| H A D | xfarray.c | 764 if (si->stack_depth >= si->max_stack_depth - 1) { in xfarray_qsort_push() 765 ASSERT(si->stack_depth < si->max_stack_depth - 1); in xfarray_qsort_push() 770 si->stack_depth + 2); in xfarray_qsort_push() 772 si_lo[si->stack_depth + 1] = lo + 1; in xfarray_qsort_push() 773 si_hi[si->stack_depth + 1] = si_hi[si->stack_depth]; in xfarray_qsort_push() 774 si_hi[si->stack_depth++] = lo - 1; in xfarray_qsort_push() 780 if (si_hi[si->stack_depth] - si_lo[si->stack_depth] > in xfarray_qsort_push() 781 si_hi[si->stack_depth - 1] - si_lo[si->stack_depth - 1]) { in xfarray_qsort_push() 782 swap(si_lo[si->stack_depth], si_lo[si->stack_depth - 1]); in xfarray_qsort_push() 783 swap(si_hi[si->stack_depth], si_hi[si->stack_depth - 1]); in xfarray_qsort_push() [all …]
|
| H A D | xfarray.h | 122 int8_t stack_depth; member
|
| H A D | trace.h | 1113 __field(int, stack_depth) 1120 __entry->stack_depth = si->stack_depth; 1128 __entry->stack_depth,
|
| /linux/lib/ |
| H A D | test_bpf.c | 84 int stack_depth; /* for eBPF only, since tests don't call verifier */ member 452 self->stack_depth = 40; in __bpf_fill_stxdw() 8159 .stack_depth = 8, 8179 .stack_depth = 8, 8196 .stack_depth = 0, 8212 .stack_depth = 0, 8228 .stack_depth = 0, 8248 .stack_depth = 8, 8268 .stack_depth = 8, 8285 .stack_depth = 0, [all …]
|
| /linux/arch/x86/net/ |
| H A D | bpf_jit_comp.c | 506 static void emit_prologue(u8 **pprog, u8 *ip, u32 stack_depth, bool ebpf_from_cbpf, in emit_prologue() argument 552 if (stack_depth) in emit_prologue() 553 EMIT3_off32(0x48, 0x81, 0xEC, round_up(stack_depth, 8)); in emit_prologue() 728 u32 stack_depth, u8 *ip, in emit_bpf_tail_call_indirect() argument 731 int tcc_ptr_off = BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack_depth); in emit_bpf_tail_call_indirect() 793 if (stack_depth) in emit_bpf_tail_call_indirect() 795 round_up(stack_depth, 8)); in emit_bpf_tail_call_indirect() 817 bool *callee_regs_used, u32 stack_depth, in emit_bpf_tail_call_direct() argument 820 int tcc_ptr_off = BPF_TAIL_CALL_CNT_PTR_STACK_OFF(stack_depth); in emit_bpf_tail_call_direct() 860 if (stack_depth) in emit_bpf_tail_call_direct() [all …]
|
| H A D | bpf_jit_comp32.c | 181 #define _STACK_SIZE (stack_depth + SCRATCH_SIZE) 1200 static void emit_prologue(u8 **pprog, u32 stack_depth) in emit_prologue() argument 1245 static void emit_epilogue(u8 **pprog, u32 stack_depth) in emit_epilogue() argument 1670 emit_prologue(&prog, bpf_prog->aux->stack_depth); in do_jit() 2472 emit_epilogue(&prog, bpf_prog->aux->stack_depth); in do_jit()
|
| /linux/drivers/firewire/ |
| H A D | core-topology.c | 107 int phy_id, stack_depth; in build_tree() local 114 stack_depth = 0; in build_tree() 168 if (child_port_count > stack_depth) { in build_tree() 236 stack_depth += 1 - child_port_count; in build_tree()
|
| /linux/drivers/net/ethernet/netronome/nfp/bpf/ |
| H A D | verifier.c | 714 frame_depths[frame] = nfp_prog->subprog[idx].stack_depth; in nfp_bpf_get_stack_usage() 782 nfp_prog->subprog[i].stack_depth = info[i].stack_depth; in nfp_bpf_finalize() 788 nfp_prog->subprog[i].stack_depth += REG_WIDTH; in nfp_bpf_finalize() 791 nfp_prog->subprog[i].stack_depth += BPF_REG_SIZE * 4; in nfp_bpf_finalize()
|
| H A D | main.h | 484 u16 stack_depth; member
|
| H A D | jit.c | 3264 u32 ret_tgt, stack_depth, offset_br; in bpf_to_bpf_call() local 3267 stack_depth = round_up(nfp_prog->stack_frame_depth, STACK_FRAME_ALIGN); in bpf_to_bpf_call() 3271 if (stack_depth) { in bpf_to_bpf_call() 3272 tmp_reg = ur_load_imm_any(nfp_prog, stack_depth, in bpf_to_bpf_call() 3331 if (stack_depth) { in bpf_to_bpf_call() 3332 tmp_reg = ur_load_imm_any(nfp_prog, stack_depth, in bpf_to_bpf_call() 3649 unsigned int depth = nfp_prog->subprog[meta->subprog_idx].stack_depth; in nfp_start_subprog() 3837 depth = nfp_prog->subprog[0].stack_depth; in nfp_translate()
|
| /linux/include/linux/ |
| H A D | kcsan-checks.h | 138 int stack_depth; member
|
| H A D | bpf.h | 1664 u32 stack_depth; member 2923 void bpf_patch_call_args(struct bpf_insn *insn, u32 stack_depth);
|
| /linux/kernel/kcsan/ |
| H A D | core.c | 437 reorder_access->stack_depth = get_kcsan_stack_depth(); in set_reorder_access() 1122 if (get_kcsan_stack_depth() <= reorder_access->stack_depth) { in __tsan_func_exit() 1133 reorder_access->stack_depth = INT_MIN; in __tsan_func_exit()
|
| /linux/kernel/bpf/ |
| H A D | core.c | 2389 void bpf_patch_call_args(struct bpf_insn *insn, u32 stack_depth) in bpf_patch_call_args() argument 2391 stack_depth = max_t(u32, stack_depth, 1); in bpf_patch_call_args() 2393 insn->imm = interpreters_args[(round_up(stack_depth, 32) / 32) - 1] - in bpf_patch_call_args() 2516 u32 stack_depth = max_t(u32, fp->aux->stack_depth, 1); in bpf_prog_select_interpreter() local 2517 u32 idx = (round_up(stack_depth, 32) / 32) - 1; in bpf_prog_select_interpreter()
|
| H A D | verifier.c | 1508 if (env->subprog_info[state->subprogno].stack_depth < size) in grow_stack_state() 1509 env->subprog_info[state->subprogno].stack_depth = size; in grow_stack_state() 6723 static int round_up_stack_depth(struct bpf_verifier_env *env, int stack_depth) in round_up_stack_depth() argument 6726 return round_up(stack_depth, 16); in round_up_stack_depth() 6731 return round_up(max_t(u32, stack_depth, 1), 32); in round_up_stack_depth() 6781 subprog_depth = round_up_stack_depth(env, subprog[idx].stack_depth); in check_max_stack_depth_subprog() 6894 depth -= round_up_stack_depth(env, subprog[idx].stack_depth); in check_max_stack_depth_subprog() 6954 return env->subprog_info[subprog].stack_depth; in get_callee_stack_depth() 22539 -(subprogs[0].stack_depth + 8)); in convert_ctx_accesses() 22546 subprogs[0].stack_depth += 8; in convert_ctx_accesses() [all …]
|
| /linux/arch/sparc/net/ |
| H A D | bpf_jit_comp_64.c | 803 u32 stack_depth; in build_prologue() local 805 stack_depth = prog->aux->stack_depth; in build_prologue() 806 stack_needed += round_up(stack_depth, 16); in build_prologue()
|
| /linux/fs/overlayfs/ |
| H A D | super.c | 412 struct ovl_fs *ofs, int *stack_depth) in ovl_lower_dir() argument 421 *stack_depth = max(*stack_depth, path->mnt->mnt_sb->s_stack_depth); in ovl_lower_dir()
|
| /linux/Documentation/bpf/ |
| H A D | drgn.rst | 132 .stack_depth = (u32)8,
|
| /linux/arch/arm64/net/ |
| H A D | bpf_jit_comp.c | 599 ctx->stack_size = round_up(prog->aux->stack_depth, 16); in build_prologue() 2056 priv_stack_alloc_sz = round_up(prog->aux->stack_depth, 16) + in bpf_int_jit_compile() 3193 priv_stack_alloc_sz = round_up(prog->aux->stack_depth, 16) + in bpf_jit_free()
|
| /linux/security/ |
| H A D | Kconfig.hardening | 145 in the /proc file system. In particular, /proc/<pid>/stack_depth
|
| /linux/arch/mips/net/ |
| H A D | bpf_jit_comp64.c | 587 locals = ALIGN(ctx->program->aux->stack_depth, MIPS_STACK_ALIGNMENT); in build_prologue()
|
| /linux/arch/arc/net/ |
| H A D | bpf_jit_core.c | 264 ctx->frame_size = ctx->prog->aux->stack_depth; in analyze_reg_usage()
|
| /linux/arch/powerpc/net/ |
| H A D | bpf_jit_comp.c | 206 cgctx.stack_size = round_up(fp->aux->stack_depth, 16); in bpf_int_jit_compile()
|
| /linux/arch/parisc/net/ |
| H A D | bpf_jit_comp64.c | 1113 bpf_stack_adjust = ctx->prog->aux->stack_depth; in bpf_jit_build_prologue()
|
| /linux/arch/riscv/net/ |
| H A D | bpf_jit_comp32.c | 1309 round_up(ctx->prog->aux->stack_depth, STACK_ALIGN); in bpf_jit_build_prologue()
|