/linux/tools/testing/selftests/bpf/progs/ |
H A D | verifier_spin_lock.c | 42 call %[bpf_spin_unlock]; \ in spin_lock_test1_success() 48 __imm(bpf_spin_unlock), in spin_lock_test1_success() 75 call %[bpf_spin_unlock]; \ in lock_test2_direct_ld_st() 81 __imm(bpf_spin_unlock), in lock_test2_direct_ld_st() 109 call %[bpf_spin_unlock]; \ in __flag() 115 __imm(bpf_spin_unlock), in __flag() 143 call %[bpf_spin_unlock]; \ in __flag() 149 __imm(bpf_spin_unlock), in __flag() 176 call %[bpf_spin_unlock]; \ in call_within_a_locked_region() 183 __imm(bpf_spin_unlock), in call_within_a_locked_region() [all...] |
H A D | rbtree_fail.c | 57 bpf_spin_unlock(&glock); in rbtree_api_nolock_remove() 94 bpf_spin_unlock(&glock); in rbtree_api_remove_unadded_node() 126 bpf_spin_unlock(&glock); in rbtree_api_remove_no_drop() 132 bpf_spin_unlock(&glock); in rbtree_api_remove_no_drop() 151 bpf_spin_unlock(&glock); in rbtree_api_add_to_multiple_trees() 168 bpf_spin_unlock(&glock); in rbtree_api_use_unchecked_remove_retval() 173 bpf_spin_unlock(&glock); in rbtree_api_use_unchecked_remove_retval() 177 bpf_spin_unlock(&glock); in rbtree_api_use_unchecked_remove_retval() 193 bpf_spin_unlock(&glock); in rbtree_api_add_release_unlock_escape() 201 bpf_spin_unlock( in rbtree_api_add_release_unlock_escape() [all...] |
H A D | linked_list.c | 39 bpf_spin_unlock(lock); in list_push_pop() 48 bpf_spin_unlock(lock); in list_push_pop() 59 bpf_spin_unlock(lock); in list_push_pop() 64 bpf_spin_unlock(lock); in list_push_pop() 76 bpf_spin_unlock(lock); in list_push_pop() 79 bpf_spin_unlock(lock); in list_push_pop() 91 bpf_spin_unlock(lock); in list_push_pop() 99 bpf_spin_unlock(lock); in list_push_pop() 137 bpf_spin_unlock(lock); in list_push_pop_multiple() 143 bpf_spin_unlock(loc in list_push_pop_multiple() [all...] |
H A D | rbtree_search.c | 72 bpf_spin_unlock(&glock0); in rbtree_search() 76 bpf_spin_unlock(&glock1); in rbtree_search() 84 bpf_spin_unlock(&glock0); in rbtree_search() 100 bpf_spin_unlock(&glock0); in rbtree_search() 110 bpf_spin_unlock(&glock0); in rbtree_search() 125 bpf_spin_unlock(&glock1); in rbtree_search() 148 bpf_spin_unlock(&glock0); \ 165 bpf_spin_unlock(&glock0); \ 170 bpf_spin_unlock(&glock0); \ 180 bpf_spin_unlock( [all...] |
H A D | rbtree.c | 67 bpf_spin_unlock(lock); in __add_three() 76 bpf_spin_unlock(lock); in __add_three() 112 bpf_spin_unlock(&glock); in rbtree_add_and_remove() 157 bpf_spin_unlock(&glock); in rbtree_add_and_remove_array() 220 bpf_spin_unlock(&glock); in rbtree_first_and_remove() 228 bpf_spin_unlock(&glock); in rbtree_first_and_remove() 240 bpf_spin_unlock(&glock); in rbtree_first_and_remove() 246 bpf_spin_unlock(&glock); in rbtree_first_and_remove() 271 bpf_spin_unlock(&glock); in rbtree_api_release_aliasing() 301 bpf_spin_unlock( in rbtree_api_release_aliasing() [all...] |
H A D | test_spin_lock_fail.c | 109 bpf_spin_unlock(B); \ 138 bpf_spin_unlock(&f2->lock); in lock_id_mismatch_mapval_mapval() 166 bpf_spin_unlock(&f2->lock); in lock_id_mismatch_innermapval_innermapval1() 193 bpf_spin_unlock(&f2->lock); in lock_id_mismatch_innermapval_innermapval2() 232 bpf_spin_unlock(&lockA); in lock_global_subprog_call1() 244 bpf_spin_unlock(&lockA); in lock_global_subprog_call2() 289 bpf_spin_unlock(&lockA); in lock_global_sleepable_helper_subprog() 301 bpf_spin_unlock(&lockA); in lock_global_sleepable_kfunc_subprog() 313 bpf_spin_unlock(&lockA); in lock_global_sleepable_subprog_indirect()
|
H A D | test_spin_lock.c | 77 bpf_spin_unlock(&val->lock); in bpf_spin_lock_test() 91 bpf_spin_unlock(&q->lock); in bpf_spin_lock_test() 99 bpf_spin_unlock(&cls->lock); in bpf_spin_lock_test() 133 bpf_spin_unlock(&lockA); in static_subprog_unlock() 145 bpf_spin_unlock(&lockA); in lock_static_subprog_call() 155 bpf_spin_unlock(&lockA); in lock_static_subprog_lock()
|
H A D | linked_list_peek.c | 33 bpf_spin_unlock(&glock); in list_peek() 39 bpf_spin_unlock(&glock); in list_peek() 50 bpf_spin_unlock(&glock); in list_peek() 80 bpf_spin_unlock(&glock); in list_peek() 98 bpf_spin_unlock(&glock); \
|
H A D | bpf_qdisc_fq.c | 172 bpf_spin_unlock(lock); in fq_flows_add_head() 182 bpf_spin_unlock(lock); in fq_flows_add_tail() 192 bpf_spin_unlock(lock); in fq_flows_remove_front() 205 bpf_spin_unlock(lock); in fq_flows_is_empty() 208 bpf_spin_unlock(lock); in fq_flows_is_empty() 379 bpf_spin_unlock(&flow->lock); in BPF_PROG() 402 bpf_spin_unlock(&fq_delayed_lock); in fq_unset_throttled_flows() 409 bpf_spin_unlock(&fq_delayed_lock); in fq_unset_throttled_flows() 415 bpf_spin_unlock(&fq_delayed_lock); in fq_unset_throttled_flows() 436 bpf_spin_unlock( in fq_flow_set_throttled() [all...] |
H A D | linked_list_fail.c | 370 bpf_spin_unlock(&glock); in use_after_unlock() 403 bpf_spin_unlock(&glock); in list_double_add() 430 bpf_spin_unlock(&glock); in no_node_value_type() 445 bpf_spin_unlock(&glock); in incorrect_value_type() 460 bpf_spin_unlock(&glock); in incorrect_node_var_off() 475 bpf_spin_unlock(&glock); in incorrect_node_off1() 490 bpf_spin_unlock(&glock); in incorrect_node_off2() 520 bpf_spin_unlock(&glock); in incorrect_head_var_off1() 535 bpf_spin_unlock(&glock); in incorrect_head_var_off2() 557 bpf_spin_unlock( in incorrect_head_off1() [all...] |
H A D | refcounted_kptr_fail.c | 48 bpf_spin_unlock(&glock); in rbtree_refcounted_node_ref_escapes() 91 bpf_spin_unlock(&glock); in rbtree_refcounted_node_ref_escapes_owning_input() 115 bpf_spin_unlock(&glock); in BPF_PROG()
|
H A D | uptr_update_failure.c | 32 bpf_spin_unlock(&ptr->lock); in not_used() 37 bpf_spin_unlock(&ptr->lock); in not_used()
|
H A D | bpf_qdisc_fifo.c | 42 bpf_spin_unlock(&q_fifo_lock); in BPF_PROG() 60 bpf_spin_unlock(&q_fifo_lock); in BPF_PROG() 98 bpf_spin_unlock(&q_fifo_lock); in BPF_PROG()
|
H A D | test_map_lock.c | 48 bpf_spin_unlock(&val->lock); in bpf_map_lock_test() 57 bpf_spin_unlock(&q->lock); in bpf_map_lock_test()
|
H A D | rbtree_btf_fail__wrong_node_type.c | 34 bpf_spin_unlock(&glock); in rbtree_api_add__wrong_node_type()
|
H A D | freplace_attach_probe.c | 35 bpf_spin_unlock(&val->lock); in new_handle_kprobe()
|
H A D | normal_map_btf.c | 52 bpf_spin_unlock(&value->lock); in add_to_list_in_array()
|
H A D | rbtree_btf_fail__add_wrong_type.c | 48 bpf_spin_unlock(&glock); in rbtree_api_add__add_wrong_type()
|
H A D | map_in_map_btf.c | 69 bpf_spin_unlock(&value->lock); in add_to_list_in_inner_array()
|
H A D | test_helper_restricted.c | 55 bpf_spin_unlock(&lock->l); in spin_lock_work()
|
/linux/drivers/hid/bpf/progs/ |
H A D | hid_bpf_async.h | 123 bpf_spin_unlock(&elem->lock); in hid_bpf_async_find_empty_key() 127 bpf_spin_unlock(&elem->lock); in hid_bpf_async_find_empty_key() 185 bpf_spin_unlock(&elem->lock); in hid_bpf_async_delayed_call() 189 bpf_spin_unlock(&elem->lock); in hid_bpf_async_delayed_call()
|
/linux/tools/sched_ext/ |
H A D | scx_flatcg.bpf.c | 240 bpf_spin_unlock(&cgv_tree_lock); in cgrp_refresh_hweight() 303 bpf_spin_unlock(&cgv_tree_lock); in cgrp_enqueued() 494 bpf_spin_unlock(&cgv_tree_lock); in update_active_weight_sums() 604 bpf_spin_unlock(&cgv_tree_lock); in BPF_STRUCT_OPS() 621 bpf_spin_unlock(&cgv_tree_lock); in try_pick_next_cgroup() 628 bpf_spin_unlock(&cgv_tree_lock); in try_pick_next_cgroup() 687 bpf_spin_unlock(&cgv_tree_lock); in try_pick_next_cgroup() 710 bpf_spin_unlock(&cgv_tree_lock); in try_pick_next_cgroup() 773 bpf_spin_unlock(&cgv_tree_lock); in BPF_STRUCT_OPS()
|
/linux/samples/bpf/ |
H A D | hbm_edt_kern.c | 103 bpf_spin_unlock(&qdp->lock); in _hbm_out_cg()
|
H A D | hbm_out_kern.c | 111 bpf_spin_unlock(&qdp->lock); in _hbm_out_cg()
|
/linux/Documentation/bpf/ |
H A D | graph_ds_impl.rst | 96 bpf_spin_unlock(&lock); 131 bpf_spin_unlock(&lock); 225 bpf_spin_unlock(&lock); 256 of ``bpf_spin_unlock`` and doing similar invalidation after
|