| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | verifier_log.c | 36 static int load_prog(struct bpf_prog_load_opts *opts, bool expect_load_error) in load_prog() argument 41 "GPL", insns, insn_cnt, opts); in load_prog() 49 LIBBPF_OPTS(bpf_prog_load_opts, opts); in verif_log_subtest() 77 opts.log_buf = logs.reference; in verif_log_subtest() 78 opts.log_size = sizeof(logs.reference); in verif_log_subtest() 79 opts.log_level = log_level | 8 /* BPF_LOG_FIXED */; in verif_log_subtest() 80 load_prog(&opts, expect_load_error); in verif_log_subtest() 100 opts.log_buf = logs.buf; in verif_log_subtest() 101 opts.log_level = log_level | 8; /* fixed-length log */ in verif_log_subtest() 102 opts in verif_log_subtest() 255 load_btf(struct bpf_btf_load_opts * opts,bool expect_err) load_btf() argument [all...] |
| H A D | kprobe_multi_test.c | 68 static void test_link_api(struct bpf_link_create_opts *opts) in test_link_api() 79 link1_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_KPROBE_MULTI, opts); in test_link_api() 83 opts->kprobe_multi.flags = BPF_F_KPROBE_MULTI_RETURN; in test_link_api() 85 link2_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_KPROBE_MULTI, opts); in test_link_api() 107 LIBBPF_OPTS(bpf_link_create_opts, opts); in test_link_api_addrs() 119 opts.kprobe_multi.addrs = (const unsigned long*) addrs; in test_link_api_addrs() 120 opts.kprobe_multi.cnt = ARRAY_SIZE(addrs); in test_link_api_addrs() 121 test_link_api(&opts); in test_link_api_addrs() 126 LIBBPF_OPTS(bpf_link_create_opts, opts); in test_link_api_syms() 138 opts in test_link_api_syms() 67 test_link_api(struct bpf_link_create_opts * opts) test_link_api() argument 143 test_attach_api(const char * pattern,struct bpf_kprobe_multi_opts * opts) test_attach_api() argument 428 do_bench_test(struct kprobe_multi_empty * skel,struct bpf_kprobe_multi_opts * opts) do_bench_test() argument [all...] |
| H A D | local_kptr_stash.c | 11 LIBBPF_OPTS(bpf_test_run_opts, opts, in test_local_kptr_stash_simple() 23 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_rb_nodes), &opts); in test_local_kptr_stash_simple() 25 ASSERT_OK(opts.retval, "local_kptr_stash_add_nodes retval"); in test_local_kptr_stash_simple() 32 LIBBPF_OPTS(bpf_test_run_opts, opts, in test_local_kptr_stash_plain() 44 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_plain), &opts); in test_local_kptr_stash_plain() 46 ASSERT_OK(opts.retval, "local_kptr_stash_add_plain retval"); in test_local_kptr_stash_plain() 53 LIBBPF_OPTS(bpf_test_run_opts, opts, in test_local_kptr_stash_local_with_root() 65 ret = bpf_prog_test_run_opts(bpf_program__fd(skel->progs.stash_local_with_root), &opts); in test_local_kptr_stash_local_with_root() 67 ASSERT_OK(opts.retval, "local_kptr_stash_add_local_with_root retval"); in test_local_kptr_stash_local_with_root() 74 LIBBPF_OPTS(bpf_test_run_opts, opts, in test_local_kptr_stash_unstash() [all …]
|
| H A D | bpf_cookie.c | 25 DECLARE_LIBBPF_OPTS(bpf_kprobe_opts, opts); in kprobe_subtest() 30 opts.bpf_cookie = 0x1; in kprobe_subtest() 31 opts.retprobe = false; in kprobe_subtest() 33 SYS_NANOSLEEP_KPROBE_NAME, &opts); in kprobe_subtest() 37 opts.bpf_cookie = 0x2; in kprobe_subtest() 38 opts.retprobe = false; in kprobe_subtest() 40 SYS_NANOSLEEP_KPROBE_NAME, &opts); in kprobe_subtest() 45 opts.bpf_cookie = 0x10; in kprobe_subtest() 46 opts.retprobe = true; in kprobe_subtest() 48 SYS_NANOSLEEP_KPROBE_NAME, &opts); in kprobe_subtest() [all...] |
| H A D | log_buf.c | 38 LIBBPF_OPTS(bpf_object_open_opts, opts); in obj_load_log_buf() 52 opts.kernel_log_buf = obj_log_buf; in obj_load_log_buf() 53 opts.kernel_log_size = log_buf_sz; in obj_load_log_buf() 54 opts.kernel_log_level = 4; /* for BTF this will turn into 1 */ in obj_load_log_buf() 59 skel = test_log_buf__open_opts(&opts); in obj_load_log_buf() 106 opts.kernel_log_buf = NULL; /* let everything through into print callback */ in obj_load_log_buf() 107 opts.kernel_log_size = 0; in obj_load_log_buf() 108 opts.kernel_log_level = 1; in obj_load_log_buf() 110 skel = test_log_buf__open_opts(&opts); in obj_load_log_buf() 156 LIBBPF_OPTS(bpf_prog_load_opts, opts); in bpf_prog_load_log_buf() [all …]
|
| H A D | raw_tp_test_run.c | 16 LIBBPF_OPTS(bpf_test_run_opts, opts, in test_raw_tp_test_run() 46 opts.ctx_in = args; in test_raw_tp_test_run() 47 opts.ctx_size_in = sizeof(__u64); in test_raw_tp_test_run() 49 err = bpf_prog_test_run_opts(prog_fd, &opts); in test_raw_tp_test_run() 52 opts.ctx_size_in = sizeof(args); in test_raw_tp_test_run() 53 err = bpf_prog_test_run_opts(prog_fd, &opts); in test_raw_tp_test_run() 55 ASSERT_EQ(opts.retval, expected_retval, "check_retval"); in test_raw_tp_test_run() 61 opts.cpu = i; in test_raw_tp_test_run() 62 opts.retval = 0; in test_raw_tp_test_run() 63 err = bpf_prog_test_run_opts(prog_fd, &opts); in test_raw_tp_test_run() [all …]
|
| /linux/tools/objtool/ |
| H A D | builtin-check.c | 23 struct opts opts; variable 38 opts.dump_orc = true; in parse_dump() 57 opts.hack_jump_label = true; in parse_hacks() 62 opts.hack_noinstr = true; in parse_hacks() 67 opts.hack_skylake = true; in parse_hacks() 76 OPT_BOOLEAN(0, "checksum", &opts.checksum, "generate per-function checksums"), 77 …OPT_BOOLEAN(0, "cfi", &opts.cfi, "annotate kernel control flow integrity (kCFI) function preambl… 78 OPT_STRING_OPTARG('d', "disas", &opts.disas, "function-pattern", "disassemble functions", "*"), 80 OPT_BOOLEAN('i', "ibt", &opts.ibt, "validate and annotate IBT"), 81 OPT_BOOLEAN('m', "mcount", &opts.mcount, "annotate mcount/fentry calls for ftrace"), [all …]
|
| /linux/drivers/usb/gadget/function/ |
| H A D | u_ether_configfs.h | 19 struct f_##_f_##_opts *opts = to_f_##_f_##_opts(item); \ 21 usb_put_function_instance(&opts->func_inst); \ 32 struct f_##_f_##_opts *opts = to_f_##_f_##_opts(item); \ 35 mutex_lock(&opts->lock); \ 36 result = gether_get_dev_addr(opts->net, page, PAGE_SIZE); \ 37 mutex_unlock(&opts->lock); \ 45 struct f_##_f_##_opts *opts = to_f_##_f_##_opts(item); \ 48 mutex_lock(&opts->lock); \ 49 if (opts->refcnt) { \ 50 mutex_unlock(&opts->lock); \ [all …]
|
| H A D | f_uac2.c | 790 static void setup_headers(struct f_uac2_opts *opts, in setup_headers() argument 832 if (EPIN_EN(opts)) in setup_headers() 834 if (EPOUT_EN(opts)) { in setup_headers() 838 if (FUOUT_EN(opts)) in setup_headers() 842 if (EPIN_EN(opts)) { in setup_headers() 845 if (FUIN_EN(opts)) in setup_headers() 851 if (EPOUT_EN(opts)) in setup_headers() 854 if (FUOUT_EN(opts) || FUIN_EN(opts)) { in setup_headers() 860 if (EPOUT_EN(opts)) { in setup_headers() 871 if (EPOUT_FBACK_IN_EN(opts)) { in setup_headers() [all …]
|
| H A D | f_uac1.c | 491 struct f_uac1_opts *opts = g_audio_to_uac1_opts(audio); in in_rq_cur() local 499 if ((FUIN_EN(opts) && (entity_id == USB_IN_FU_ID)) || in in_rq_cur() 500 (FUOUT_EN(opts) && (entity_id == USB_OUT_FU_ID))) { in in_rq_cur() 503 if (FUIN_EN(opts) && (entity_id == USB_IN_FU_ID)) in in_rq_cur() 542 struct f_uac1_opts *opts = g_audio_to_uac1_opts(audio); in in_rq_min() local 550 if ((FUIN_EN(opts) && (entity_id == USB_IN_FU_ID)) || in in_rq_min() 551 (FUOUT_EN(opts) && (entity_id == USB_OUT_FU_ID))) { in in_rq_min() 554 if (FUIN_EN(opts) && (entity_id == USB_IN_FU_ID)) in in_rq_min() 562 min_db = opts->p_volume_min; in in_rq_min() 564 min_db = opts->c_volume_min; in in_rq_min() [all …]
|
| H A D | uvc_configfs.c | 192 struct f_uvc_opts *opts; \ 200 opts = to_f_uvc_opts(opts_item); \ 202 mutex_lock(&opts->lock); \ 204 mutex_unlock(&opts->lock); \ 215 struct f_uvc_opts *opts; \ 224 opts = to_f_uvc_opts(opts_item); \ 226 mutex_lock(&opts->lock); \ 227 if (ch->linked || opts->refcnt) { \ 243 mutex_unlock(&opts->lock); \ 310 struct f_uvc_opts *opts; \ [all …]
|
| /linux/drivers/nvme/host/ |
| H A D | fabrics.c | 134 if (ctrl->opts->mask & NVMF_OPT_TRADDR) in nvmf_get_address() 135 len += scnprintf(buf, size, "traddr=%s", ctrl->opts->traddr); in nvmf_get_address() 136 if (ctrl->opts->mask & NVMF_OPT_TRSVCID) in nvmf_get_address() 138 (len) ? "," : "", ctrl->opts->trsvcid); in nvmf_get_address() 139 if (ctrl->opts->mask & NVMF_OPT_HOST_TRADDR) in nvmf_get_address() 141 (len) ? "," : "", ctrl->opts->host_traddr); in nvmf_get_address() 142 if (ctrl->opts->mask & NVMF_OPT_HOST_IFACE) in nvmf_get_address() 144 (len) ? "," : "", ctrl->opts->host_iface); in nvmf_get_address() 403 uuid_copy(&data->hostid, &ctrl->opts->host->id); in nvmf_connect_data_prep() 405 strscpy(data->subsysnqn, ctrl->opts->subsysnqn, NVMF_NQN_SIZE); in nvmf_connect_data_prep() [all …]
|
| /linux/tools/lib/bpf/ |
| H A D | bpf.c | 173 const struct bpf_map_create_opts *opts) in bpf_map_create() argument 183 if (!OPTS_VALID(opts, bpf_map_create_opts)) in bpf_map_create() 193 attr.btf_fd = OPTS_GET(opts, btf_fd, 0); in bpf_map_create() 194 attr.btf_key_type_id = OPTS_GET(opts, btf_key_type_id, 0); in bpf_map_create() 195 attr.btf_value_type_id = OPTS_GET(opts, btf_value_type_id, 0); in bpf_map_create() 196 attr.btf_vmlinux_value_type_id = OPTS_GET(opts, btf_vmlinux_value_type_id, 0); in bpf_map_create() 197 attr.value_type_btf_obj_fd = OPTS_GET(opts, value_type_btf_obj_fd, 0); in bpf_map_create() 199 attr.inner_map_fd = OPTS_GET(opts, inner_map_fd, 0); in bpf_map_create() 200 attr.map_flags = OPTS_GET(opts, map_flags, 0); in bpf_map_create() 201 attr.map_extra = OPTS_GET(opts, map_extr in bpf_map_create() 241 bpf_prog_load(enum bpf_prog_type prog_type,const char * prog_name,const char * license,const struct bpf_insn * insns,size_t insn_cnt,struct bpf_prog_load_opts * opts) bpf_prog_load() argument 529 bpf_map_batch_common(int cmd,int fd,void * in_batch,void * out_batch,void * keys,void * values,__u32 * count,const struct bpf_map_batch_opts * opts) bpf_map_batch_common() argument 555 bpf_map_delete_batch(int fd,const void * keys,__u32 * count,const struct bpf_map_batch_opts * opts) bpf_map_delete_batch() argument 563 bpf_map_lookup_batch(int fd,void * in_batch,void * out_batch,void * keys,void * values,__u32 * count,const struct bpf_map_batch_opts * opts) bpf_map_lookup_batch() argument 571 bpf_map_lookup_and_delete_batch(int fd,void * in_batch,void * out_batch,void * keys,void * values,__u32 * count,const struct bpf_map_batch_opts * opts) bpf_map_lookup_and_delete_batch() argument 579 bpf_map_update_batch(int fd,const void * keys,const void * values,__u32 * count,const struct bpf_map_batch_opts * opts) bpf_map_update_batch() argument 585 bpf_obj_pin_opts(int fd,const char * pathname,const struct bpf_obj_pin_opts * opts) bpf_obj_pin_opts() argument 614 bpf_obj_get_opts(const char * pathname,const struct bpf_obj_get_opts * opts) bpf_obj_get_opts() argument 643 bpf_prog_attach_opts(int prog_fd,int target,enum bpf_attach_type type,const struct bpf_prog_attach_opts * opts) bpf_prog_attach_opts() argument 681 bpf_prog_detach_opts(int prog_fd,int target,enum bpf_attach_type type,const struct bpf_prog_detach_opts * opts) bpf_prog_detach_opts() argument 729 bpf_link_create(int prog_fd,int target_fd,enum bpf_attach_type attach_type,const struct bpf_link_create_opts * opts) bpf_link_create() argument 940 bpf_link_update(int link_fd,int new_prog_fd,const struct bpf_link_update_opts * opts) bpf_link_update() argument 979 bpf_prog_query_opts(int target,enum bpf_attach_type type,struct bpf_prog_query_opts * opts) bpf_prog_query_opts() argument 1026 bpf_prog_test_run_opts(int prog_fd,struct bpf_test_run_opts * opts) bpf_prog_test_run_opts() argument 1098 bpf_prog_get_fd_by_id_opts(__u32 id,const struct bpf_get_fd_by_id_opts * opts) bpf_prog_get_fd_by_id_opts() argument 1121 bpf_map_get_fd_by_id_opts(__u32 id,const struct bpf_get_fd_by_id_opts * opts) bpf_map_get_fd_by_id_opts() argument 1144 bpf_btf_get_fd_by_id_opts(__u32 id,const struct bpf_get_fd_by_id_opts * opts) bpf_btf_get_fd_by_id_opts() argument 1168 bpf_link_get_fd_by_id_opts(__u32 id,const struct bpf_get_fd_by_id_opts * opts) bpf_link_get_fd_by_id_opts() argument 1227 bpf_raw_tracepoint_open_opts(int prog_fd,struct bpf_raw_tp_opts * opts) bpf_raw_tracepoint_open_opts() argument 1252 bpf_btf_load(const void * btf_data,size_t btf_size,struct bpf_btf_load_opts * opts) bpf_btf_load() argument 1346 bpf_prog_bind_map(int prog_fd,int map_fd,const struct bpf_prog_bind_opts * opts) bpf_prog_bind_map() argument 1364 bpf_token_create(int bpffs_fd,struct bpf_token_create_opts * opts) bpf_token_create() argument 1382 bpf_prog_stream_read(int prog_fd,__u32 stream_id,void * buf,__u32 buf_len,struct bpf_prog_stream_read_opts * opts) bpf_prog_stream_read() argument [all...] |
| H A D | libbpf_probes.c | 107 LIBBPF_OPTS(bpf_prog_load_opts, opts, in probe_prog_load() 118 opts.expected_attach_type = BPF_CGROUP_INET4_CONNECT; in probe_prog_load() 121 opts.expected_attach_type = BPF_CGROUP_GETSOCKOPT; in probe_prog_load() 124 opts.expected_attach_type = BPF_SK_LOOKUP; in probe_prog_load() 127 opts.kern_version = get_kernel_version(); in probe_prog_load() 130 opts.expected_attach_type = BPF_LIRC_MODE2; in probe_prog_load() 134 opts.log_buf = buf; in probe_prog_load() 135 opts.log_size = sizeof(buf); in probe_prog_load() 136 opts.log_level = 1; in probe_prog_load() 138 opts in probe_prog_load() 205 libbpf_probe_bpf_prog_type(enum bpf_prog_type prog_type,const void * opts) libbpf_probe_bpf_prog_type() argument 409 libbpf_probe_bpf_map_type(enum bpf_map_type map_type,const void * opts) libbpf_probe_bpf_map_type() argument 421 libbpf_probe_bpf_helper(enum bpf_prog_type prog_type,enum bpf_func_id helper_id,const void * opts) libbpf_probe_bpf_helper() argument [all...] |
| /linux/tools/testing/vsock/ |
| H A D | vsock_test.c | 38 static void test_stream_connection_reset(const struct test_opts *opts) in test_stream_connection_reset() argument 46 .svm_port = opts->peer_port, in test_stream_connection_reset() 47 .svm_cid = opts->peer_cid, in test_stream_connection_reset() 74 static void test_stream_bind_only_client(const struct test_opts *opts) in test_stream_bind_only_client() argument 82 .svm_port = opts->peer_port, in test_stream_bind_only_client() 83 .svm_cid = opts->peer_cid, in test_stream_bind_only_client() 116 static void test_stream_bind_only_server(const struct test_opts *opts) in test_stream_bind_only_server() argument 120 fd = vsock_bind(VMADDR_CID_ANY, opts->peer_port, SOCK_STREAM); in test_stream_bind_only_server() 131 static void test_stream_client_close_client(const struct test_opts *opts) in test_stream_client_close_client() argument 135 fd = vsock_stream_connect(opts->peer_cid, opts->peer_port); in test_stream_client_close_client() [all …]
|
| /linux/net/mptcp/ |
| H A D | options.c | 402 unsigned int *size, struct mptcp_out_options *opts) in mptcp_syn_options() argument 421 opts->suboptions = OPTION_MPTCP_MPC_SYN; in mptcp_syn_options() 422 opts->csum_reqd = mptcp_is_checksum_enabled(sock_net(sk)); in mptcp_syn_options() 423 opts->allow_join_id0 = mptcp_allow_join_id0(sock_net(sk)); in mptcp_syn_options() 429 opts->suboptions = OPTION_MPTCP_MPJ_SYN; in mptcp_syn_options() 430 opts->join_id = subflow->local_id; in mptcp_syn_options() 431 opts->token = subflow->remote_token; in mptcp_syn_options() 432 opts->nonce = subflow->local_nonce; in mptcp_syn_options() 433 opts->backup = subflow->request_bkup; in mptcp_syn_options() 452 struct mptcp_out_options *opts) in mptcp_established_options_mp() argument 565 mptcp_established_options_dss(struct sock * sk,struct sk_buff * skb,bool snd_data_fin_enable,unsigned int * size,struct mptcp_out_options * opts) mptcp_established_options_dss() argument 657 mptcp_established_options_add_addr(struct sock * sk,struct sk_buff * skb,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options_add_addr() argument 721 mptcp_established_options_rm_addr(struct sock * sk,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options_rm_addr() argument 751 mptcp_established_options_mp_prio(struct sock * sk,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options_mp_prio() argument 777 mptcp_established_options_rst(struct sock * sk,struct sk_buff * skb,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options_rst() argument 796 mptcp_established_options_fastclose(struct sock * sk,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options_fastclose() argument 819 mptcp_established_options_mp_fail(struct sock * sk,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options_mp_fail() argument 841 mptcp_established_options(struct sock * sk,struct sk_buff * skb,unsigned int * size,unsigned int remaining,struct mptcp_out_options * opts) mptcp_established_options() argument 914 mptcp_synack_options(const struct request_sock * req,unsigned int * size,struct mptcp_out_options * opts) mptcp_synack_options() argument 1416 mptcp_write_options(struct tcphdr * th,__be32 * ptr,struct tcp_sock * tp,struct mptcp_out_options * opts) mptcp_write_options() argument [all...] |
| /linux/net/netfilter/ |
| H A D | nf_synproxy_core.c | 30 const struct tcphdr *th, struct synproxy_options *opts) in synproxy_parse_options() argument 42 opts->options = 0; in synproxy_parse_options() 65 opts->mss_option = get_unaligned_be16(ptr); in synproxy_parse_options() 66 opts->options |= NF_SYNPROXY_OPT_MSS; in synproxy_parse_options() 71 opts->wscale = *ptr; in synproxy_parse_options() 72 if (opts->wscale > TCP_MAX_WSCALE) in synproxy_parse_options() 73 opts->wscale = TCP_MAX_WSCALE; in synproxy_parse_options() 74 opts->options |= NF_SYNPROXY_OPT_WSCALE; in synproxy_parse_options() 79 opts->tsval = get_unaligned_be32(ptr); in synproxy_parse_options() 80 opts->tsecr = get_unaligned_be32(ptr + 4); in synproxy_parse_options() [all …]
|
| H A D | nf_conntrack_bpf.c | 110 u32 tuple_len, struct bpf_ct_opts *opts, u32 opts_len, in __bpf_nf_ct_alloc_entry() argument 121 if (opts->reserved[0] || opts->reserved[1] || opts->reserved[2]) in __bpf_nf_ct_alloc_entry() 124 if (opts->ct_zone_id) in __bpf_nf_ct_alloc_entry() 128 if (unlikely(opts->netns_id < BPF_F_CURRENT_NETNS)) in __bpf_nf_ct_alloc_entry() 131 err = bpf_nf_ct_tuple_parse(bpf_tuple, tuple_len, opts->l4proto, in __bpf_nf_ct_alloc_entry() 136 err = bpf_nf_ct_tuple_parse(bpf_tuple, tuple_len, opts->l4proto, in __bpf_nf_ct_alloc_entry() 141 if (opts->netns_id >= 0) { in __bpf_nf_ct_alloc_entry() 142 net = get_net_ns_by_id(net, opts->netns_id); in __bpf_nf_ct_alloc_entry() 148 if (opts->ct_zone_dir == 0) in __bpf_nf_ct_alloc_entry() 149 opts->ct_zone_dir = NF_CT_DEFAULT_ZONE_DIR; in __bpf_nf_ct_alloc_entry() [all …]
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | test_bpf_nf_fail.c | 36 struct bpf_ct_opts___local opts = {}; in alloc_release() 40 ct = bpf_skb_ct_alloc(ctx, &tup, sizeof(tup.ipv4), &opts, sizeof(opts)); in alloc_release() 50 struct bpf_ct_opts___local opts = {}; in insert_insert() 54 ct = bpf_skb_ct_alloc(ctx, &tup, sizeof(tup.ipv4), &opts, sizeof(opts)); in insert_insert() 67 struct bpf_ct_opts___local opts = {}; in lookup_insert() 71 ct = bpf_skb_ct_lookup(ctx, &tup, sizeof(tup.ipv4), &opts, sizeof(opts)); in lookup_insert() 81 struct bpf_ct_opts___local opts in write_not_allowlisted_field() 31 struct bpf_ct_opts___local opts = {}; alloc_release() local 45 struct bpf_ct_opts___local opts = {}; insert_insert() local 62 struct bpf_ct_opts___local opts = {}; lookup_insert() local 76 struct bpf_ct_opts___local opts = {}; write_not_allowlisted_field() local 90 struct bpf_ct_opts___local opts = {}; set_timeout_after_insert() local 107 struct bpf_ct_opts___local opts = {}; set_status_after_insert() local 124 struct bpf_ct_opts___local opts = {}; change_timeout_after_alloc() local 138 struct bpf_ct_opts___local opts = {}; change_status_after_alloc() local [all...] |
| /linux/tools/testing/selftests/net/netfilter/ |
| H A D | nf_queue.c | 32 static struct options opts; variable 93 if (opts.verbose > 0) in queue_cb() 102 if (opts.verbose > 0) { in queue_cb() 116 if (opts.count_packets) in queue_cb() 230 queue_num = opts.queue_num; in open_queue() 240 flags = opts.gso_enabled ? NFQA_CFG_F_GSO : 0; in open_queue() 242 if (opts.failopen) in open_queue() 253 tv.tv_sec = opts.timeout; in open_queue() 254 if (opts.timeout && setsockopt(mnl_socket_get_fd(nl), in open_queue() 316 if (opts.bogus_verdict && errno == ENOENT) in mainloop() [all …]
|
| /linux/tools/perf/util/ |
| H A D | record.c | 95 void evlist__config(struct evlist *evlist, struct record_opts *opts, struct callchain_param *callch… in evlist__config() argument 100 bool sample_id = opts->sample_id; in evlist__config() 103 opts->no_inherit = true; in evlist__config() 108 evsel__config(evsel, opts, callchain); in evlist__config() 117 if (opts->full_auxtrace || opts->sample_identifier) { in evlist__config() 150 static int record_opts__config_freq(struct record_opts *opts) in record_opts__config_freq() argument 152 bool user_freq = opts->user_freq != UINT_MAX; in record_opts__config_freq() 153 bool user_interval = opts->user_interval != ULLONG_MAX; in record_opts__config_freq() 162 opts->default_interval = opts->user_interval; in record_opts__config_freq() 164 opts->freq = opts->user_freq; in record_opts__config_freq() [all …]
|
| /linux/tools/perf/arch/x86/util/ |
| H A D | intel-bts.c | 107 struct record_opts *opts) in intel_bts_recording_options() argument 116 if (opts->auxtrace_sample_mode) { in intel_bts_recording_options() 122 btsr->snapshot_mode = opts->auxtrace_snapshot_mode; in intel_bts_recording_options() 134 opts->full_auxtrace = true; in intel_bts_recording_options() 138 if (opts->auxtrace_snapshot_mode && !opts->full_auxtrace) { in intel_bts_recording_options() 143 if (!opts->full_auxtrace) in intel_bts_recording_options() 146 if (opts->full_auxtrace && !perf_cpu_map__is_any_cpu_or_is_empty(cpus)) { in intel_bts_recording_options() 152 if (opts->auxtrace_snapshot_mode) { in intel_bts_recording_options() 153 if (!opts->auxtrace_snapshot_size && !opts->auxtrace_mmap_pages) { in intel_bts_recording_options() 155 opts->auxtrace_mmap_pages = MiB(4) / page_size; in intel_bts_recording_options() [all …]
|
| /linux/drivers/net/ppp/ |
| H A D | ppp_mppe.h | 34 #define MPPE_OPTS_TO_CI(opts, ci) \ argument 39 if (opts & MPPE_OPT_STATEFUL) \ 48 if (opts & MPPE_OPT_128) \ 50 if (opts & MPPE_OPT_40) \ 56 #define MPPE_CI_TO_OPTS(ci, opts) \ argument 60 opts = 0; \ 64 opts |= MPPE_OPT_STATEFUL; \ 68 opts |= MPPE_OPT_128; \ 70 opts |= MPPE_OPT_40; \ 74 opts |= MPPE_OPT_56; \ [all …]
|
| /linux/net/xfrm/ |
| H A D | xfrm_state_bpf.c | 59 * @opts - Options for lookup (documented above) 65 bpf_xdp_get_xfrm_state(struct xdp_md *ctx, struct bpf_xfrm_state_opts *opts, u32 opts__sz) in bpf_xdp_get_xfrm_state() argument 71 if (opts__sz < sizeof(opts->error)) in bpf_xdp_get_xfrm_state() 75 opts->error = -EINVAL; in bpf_xdp_get_xfrm_state() 79 if (unlikely(opts->netns_id < BPF_F_CURRENT_NETNS)) { in bpf_xdp_get_xfrm_state() 80 opts->error = -EINVAL; in bpf_xdp_get_xfrm_state() 84 if (opts->netns_id >= 0) { in bpf_xdp_get_xfrm_state() 85 net = get_net_ns_by_id(net, opts->netns_id); in bpf_xdp_get_xfrm_state() 87 opts->error = -ENONET; in bpf_xdp_get_xfrm_state() 92 x = xfrm_state_lookup(net, opts in bpf_xdp_get_xfrm_state() [all...] |
| /linux/tools/perf/arch/arm64/util/ |
| H A D | arm-spe.c | 190 arm_spe_snapshot_resolve_auxtrace_defaults(struct record_opts *opts, in arm_spe_snapshot_resolve_auxtrace_defaults() argument 207 if (!opts->auxtrace_snapshot_size && !opts->auxtrace_mmap_pages) { in arm_spe_snapshot_resolve_auxtrace_defaults() 209 opts->auxtrace_mmap_pages = MiB(4) / page_size; in arm_spe_snapshot_resolve_auxtrace_defaults() 211 opts->auxtrace_mmap_pages = KiB(128) / page_size; in arm_spe_snapshot_resolve_auxtrace_defaults() 212 if (opts->mmap_pages == UINT_MAX) in arm_spe_snapshot_resolve_auxtrace_defaults() 213 opts->mmap_pages = KiB(256) / page_size; in arm_spe_snapshot_resolve_auxtrace_defaults() 215 } else if (!opts->auxtrace_mmap_pages && !privileged && opts->mmap_pages == UINT_MAX) { in arm_spe_snapshot_resolve_auxtrace_defaults() 216 opts->mmap_pages = KiB(256) / page_size; in arm_spe_snapshot_resolve_auxtrace_defaults() 223 if (!opts->auxtrace_snapshot_size) in arm_spe_snapshot_resolve_auxtrace_defaults() 224 opts->auxtrace_snapshot_size = opts->auxtrace_mmap_pages * (size_t)page_size; in arm_spe_snapshot_resolve_auxtrace_defaults() [all …]
|