| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_ref_tracking.c | 121 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in reference_tracking_leak_potential_reference() 136 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in potential_reference_to_sock_common_1() 154 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in leak_potential_reference_on_stack() 174 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in potential_reference_on_stack_2() 189 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in reference_tracking_zero_potential_reference() 204 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in potential_reference_to_sock_common_2() 221 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in copy_and_zero_potential_references() 360 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in tracking_release_reference_without_check() 379 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in to_sock_common_without_check() 397 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in reference_tracking_release_reference() [all …]
|
| H A D | test_sk_lookup_kern.c | 21 static struct bpf_sock_tuple *get_tuple(void *data, __u64 nh_off, in get_tuple() 25 struct bpf_sock_tuple *result; in get_tuple() 37 result = (struct bpf_sock_tuple *)&iph->saddr; in get_tuple() 46 result = (struct bpf_sock_tuple *)&ip6h->saddr; in get_tuple() 61 struct bpf_sock_tuple *tuple; in sk_lookup_success() 84 struct bpf_sock_tuple tuple = {}; in sk_lookup_success_simple() 96 struct bpf_sock_tuple tuple = {}; in err_use_after_free() 111 struct bpf_sock_tuple tuple = {}; in err_modify_sk_pointer() 125 struct bpf_sock_tuple tuple = {}; in err_modify_sk_or_null_pointer() 138 struct bpf_sock_tuple tuple = {}; in err_no_release() [all …]
|
| H A D | test_bpf_nf_fail.c | 18 struct nf_conn *bpf_skb_ct_alloc(struct __sk_buff *, struct bpf_sock_tuple *, u32, 20 struct nf_conn *bpf_skb_ct_lookup(struct __sk_buff *, struct bpf_sock_tuple *, u32, 22 struct nf_conn *bpf_xdp_ct_alloc(struct xdp_md *, struct bpf_sock_tuple *, u32, 24 struct nf_conn *bpf_xdp_ct_lookup(struct xdp_md *, struct bpf_sock_tuple *, u32, 37 struct bpf_sock_tuple tup = {}; in alloc_release() 51 struct bpf_sock_tuple tup = {}; in insert_insert() 68 struct bpf_sock_tuple tup = {}; in lookup_insert() 82 struct bpf_sock_tuple tup = {}; in write_not_allowlisted_field() 96 struct bpf_sock_tuple tup = {}; in set_timeout_after_insert() 113 struct bpf_sock_tuple tu in set_status_after_insert() [all...] |
| H A D | test_sk_assign.c | 54 static inline struct bpf_sock_tuple * 59 struct bpf_sock_tuple *result; in get_tuple() 79 result = (struct bpf_sock_tuple *)&iph->saddr; in get_tuple() 88 result = (struct bpf_sock_tuple *)&ip6h->saddr; in get_tuple() 90 return (struct bpf_sock_tuple *)data; in get_tuple() 102 handle_udp(struct __sk_buff *skb, struct bpf_sock_tuple *tuple, bool ipv4) in handle_udp() 133 handle_tcp(struct __sk_buff *skb, struct bpf_sock_tuple *tuple, bool ipv4) in handle_tcp() 174 struct bpf_sock_tuple *tuple; in bpf_sk_assign_test()
|
| H A D | test_bpf_nf.c | 71 struct nf_conn *bpf_xdp_ct_alloc(struct xdp_md *, struct bpf_sock_tuple *, u32, 73 struct nf_conn *bpf_xdp_ct_lookup(struct xdp_md *, struct bpf_sock_tuple *, u32, 75 struct nf_conn *bpf_skb_ct_alloc(struct __sk_buff *, struct bpf_sock_tuple *, u32, 77 struct nf_conn *bpf_skb_ct_lookup(struct __sk_buff *, struct bpf_sock_tuple *, u32, 89 nf_ct_test(struct nf_conn *(*lookup_fn)(void *, struct bpf_sock_tuple *, u32, 91 struct nf_conn *(*alloc_fn)(void *, struct bpf_sock_tuple *, u32, in nf_ct_test() 96 struct bpf_sock_tuple bpf_tuple; in nf_ct_test() 238 nf_ct_opts_new_test(struct nf_conn *(*lookup_fn)(void *, struct bpf_sock_tuple *, u32, in nf_ct_test() 240 struct nf_conn *(*alloc_fn)(void *, struct bpf_sock_tuple *, u32, in nf_ct_test() 245 struct bpf_sock_tuple bpf_tupl argument 90 nf_ct_test(struct nf_conn * (* lookup_fn)(void *,struct bpf_sock_tuple *,u32,struct bpf_ct_opts___local *,u32),struct nf_conn * (* alloc_fn)(void *,struct bpf_sock_tuple *,u32,struct bpf_ct_opts___local *,u32),void * ctx) nf_ct_test() argument [all...] |
| H A D | test_cls_redirect_dynptr.c | 88 offsetofend(struct bpf_sock_tuple, ipv4.dport) - 89 offsetof(struct bpf_sock_tuple, ipv4.sport) - 1, 93 offsetofend(struct bpf_sock_tuple, ipv6.dport) - 94 offsetof(struct bpf_sock_tuple, ipv6.sport) - 1, 511 static uint64_t fill_tuple(struct bpf_sock_tuple *tuple, void *iph, in fill_tuple() 540 static verdict_t classify_tcp(struct __sk_buff *skb, struct bpf_sock_tuple *tuple, in classify_tcp() 572 static verdict_t classify_udp(struct __sk_buff *skb, struct bpf_sock_tuple *tuple, uint64_t tuplen) in classify_udp() 589 static verdict_t classify_icmp(struct __sk_buff *skb, uint8_t proto, struct bpf_sock_tuple *tuple, in classify_icmp() 641 struct bpf_sock_tuple tuple; in process_icmpv4() 657 struct bpf_sock_tuple tuple; in process_icmpv6() [all …]
|
| H A D | test_cls_redirect.c | 92 offsetofend(struct bpf_sock_tuple, ipv4.dport) - 93 offsetof(struct bpf_sock_tuple, ipv4.sport) - 1, 97 offsetofend(struct bpf_sock_tuple, ipv6.dport) - 98 offsetof(struct bpf_sock_tuple, ipv6.sport) - 1, 619 static INLINING uint64_t fill_tuple(struct bpf_sock_tuple *tuple, void *iph, in fill_tuple() 649 struct bpf_sock_tuple *tuple, uint64_t tuplen, in classify_tcp() 682 struct bpf_sock_tuple *tuple, uint64_t tuplen) in classify_udp() 700 struct bpf_sock_tuple *tuple, uint64_t tuplen, in classify_icmp() 750 struct bpf_sock_tuple tuple; in process_icmpv4() 802 struct bpf_sock_tuple tuple; in process_icmpv6() [all …]
|
| H A D | test_btf_skc_cls_ingress.c | 68 struct bpf_sock_tuple *tuple = NULL; in handle_ip_tcp() 93 tuple = (struct bpf_sock_tuple *)&ip4h->saddr; in handle_ip_tcp() 110 tuple = (struct bpf_sock_tuple *)&ip6h->saddr; in handle_ip_tcp()
|
| H A D | vrf_socket_lookup.c | 21 struct bpf_sock_tuple *tp; in socket_lookup() 36 tp = (struct bpf_sock_tuple *)&iph->saddr; in socket_lookup()
|
| H A D | verifier_unpriv.c | 10 /* struct bpf_sock_tuple tuple = {} */ \ 434 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in pointers_stx_ctx_and_sock() 470 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in different_pointers_stx_leak_sock() 507 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in stx_sock_and_ctx_read() 547 __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) in stx_sock_and_ctx_write()
|
| H A D | cgroup_skb_sk_lookup_kern.c | 29 static inline void set_tuple(struct bpf_sock_tuple *tuple, in set_tuple() 44 struct bpf_sock_tuple tuple; in is_allowed_peer_cg()
|
| H A D | connect6_prog.c | 30 struct bpf_sock_tuple tuple = {}; in connect_v6_prog()
|
| H A D | connect4_prog.c | 146 struct bpf_sock_tuple tuple = {}; in connect_v4_prog()
|
| H A D | xdp_synproxy_kern.c | 99 struct bpf_sock_tuple *bpf_tuple, 105 struct bpf_sock_tuple *bpf_tuple, 436 struct bpf_sock_tuple tup = {}; in tcp_lookup()
|
| H A D | test_tcp_custom_syncookie.c | 509 struct bpf_sock_tuple tuple; in tcp_handle_ack()
|
| /linux/net/netfilter/ |
| H A D | nf_conntrack_bpf.c | 68 static int bpf_nf_ct_tuple_parse(struct bpf_sock_tuple *bpf_tuple, in bpf_nf_ct_tuple_parse() 109 __bpf_nf_ct_alloc_entry(struct net *net, struct bpf_sock_tuple *bpf_tuple, in __bpf_nf_ct_alloc_entry() 172 struct bpf_sock_tuple *bpf_tuple, in __bpf_nf_ct_lookup() 292 bpf_xdp_ct_alloc(struct xdp_md *xdp_ctx, struct bpf_sock_tuple *bpf_tuple, in bpf_xdp_ct_alloc() 325 bpf_xdp_ct_lookup(struct xdp_md *xdp_ctx, struct bpf_sock_tuple *bpf_tuple, in bpf_xdp_ct_lookup() 357 bpf_skb_ct_alloc(struct __sk_buff *skb_ctx, struct bpf_sock_tuple *bpf_tuple, in bpf_skb_ct_alloc() 391 bpf_skb_ct_lookup(struct __sk_buff *skb_ctx, struct bpf_sock_tuple *bpf_tuple, in bpf_skb_ct_lookup()
|
| /linux/net/core/ |
| H A D | filter.c | 6874 static struct sock *sk_lookup(struct net *net, struct bpf_sock_tuple *tuple, in sk_lookup() 6923 __bpf_skc_lookup(struct sk_buff *skb, struct bpf_sock_tuple *tuple, u32 len, in __bpf_skc_lookup() 6964 __bpf_sk_lookup(struct sk_buff *skb, struct bpf_sock_tuple *tuple, u32 len, in __bpf_sk_lookup() 6993 bpf_skc_lookup(struct sk_buff *skb, struct bpf_sock_tuple *tuple, u32 len, in bpf_skc_lookup() 7012 bpf_sk_lookup(struct sk_buff *skb, struct bpf_sock_tuple *tuple, u32 len, in bpf_sk_lookup() 7039 struct bpf_sock_tuple *, tuple, u32, len, u64, netns_id, u64, flags) in BPF_CALL_5() argument 7058 struct bpf_sock_tuple *, tuple, u32, len, u64, netns_id, u64, flags) in BPF_CALL_5() argument 7077 struct bpf_sock_tuple *, tuple, u32, len, u64, netns_id, u64, flags) in BPF_CALL_5() argument 7096 struct bpf_sock_tuple *, tuple, u32, len, u64, netns_id, u64, flags) in BPF_CALL_5() argument 7120 struct bpf_sock_tuple *, tuple, u32, len, u64, netns_id, u64, flags) in BPF_CALL_5() argument [all …]
|
| /linux/include/uapi/linux/ |
| H A D | bpf.h | 3777 * struct bpf_sock *bpf_sk_lookup_tcp(void *ctx, struct bpf_sock_tuple *tuple, u32 tuple_size, u64 netns, u64 flags) 3814 * struct bpf_sock *bpf_sk_lookup_udp(void *ctx, struct bpf_sock_tuple *tuple, u32 tuple_size, u64 netns, u64 flags) 4012 * struct bpf_sock *bpf_skc_lookup_tcp(void *ctx, struct bpf_sock_tuple *tuple, u32 tuple_size, u64 netns, u64 flags) 6505 struct bpf_sock_tuple { 6477 struct bpf_sock_tuple { global() struct 6478 __anon1e8efe9c530abpf_sock_tuple global() argument
|
| /linux/tools/include/uapi/linux/ |
| H A D | bpf.h | 3777 * struct bpf_sock *bpf_sk_lookup_tcp(void *ctx, struct bpf_sock_tuple *tuple, u32 tuple_size, u64 netns, u64 flags) 3814 * struct bpf_sock *bpf_sk_lookup_udp(void *ctx, struct bpf_sock_tuple *tuple, u32 tuple_size, u64 netns, u64 flags) 4012 * struct bpf_sock *bpf_skc_lookup_tcp(void *ctx, struct bpf_sock_tuple *tuple, u32 tuple_size, u64 netns, u64 flags) 6505 struct bpf_sock_tuple { 6477 struct bpf_sock_tuple { global() struct 6478 __anonc2086e7c530abpf_sock_tuple global() argument
|
| /linux/tools/testing/selftests/bpf/ |
| H A D | test_verifier.c | 474 BPF_MOV64_IMM(BPF_REG_3, sizeof(struct bpf_sock_tuple)), \
|