Home
last modified time | relevance | path

Searched refs:max_entries (Results 1 – 25 of 459) sorted by relevance

12345678910>>...19

/linux/tools/testing/selftests/bpf/map_tests/
H A Darray_map_batch_ops.c15 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument
25 for (i = 0; i < max_entries; i++) { in map_batch_update()
36 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update()
40 static void map_batch_verify(int *visited, __u32 max_entries, int *keys, in map_batch_verify() argument
46 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify()
47 for (i = 0; i < max_entries; i++) { in map_batch_verify()
64 for (i = 0; i < max_entries; i++) { in map_batch_verify()
74 const __u32 max_entries = 10; in __test_map_lookup_and_update_batch() local
84 "array_map", sizeof(int), sizeof(__s64), max_entries, NULL); in __test_map_lookup_and_update_batch()
92 keys = calloc(max_entries, sizeo in __test_map_lookup_and_update_batch()
[all...]
H A Dmap_in_map_batch_ops.c126 int err, max_entries = OUTER_MAP_ENTRIES - !!has_holes; in fetch_and_validate() local
132 fetched_keys = calloc(max_entries, value_size); in fetch_and_validate()
133 fetched_values = calloc(max_entries, value_size); in fetch_and_validate()
172 CHECK((total_fetched != max_entries), in fetch_and_validate()
174 "total_fetched(%d) and max_entries(%d) error: (%d):%s\n", in fetch_and_validate()
175 total_fetched, max_entries, errno, strerror(errno)); in fetch_and_validate()
191 __u32 max_entries = OUTER_MAP_ENTRIES - !!has_holes; in _map_in_map_batch_ops() local
209 for (map_index = 0; map_index < max_entries; map_index++) in _map_in_map_batch_ops()
216 * max_entries == OUTER_MAP_ENTRIES - 1 if it is true. Say in _map_in_map_batch_ops()
217 * max_entries i in _map_in_map_batch_ops()
[all...]
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_btf_map_in_map.c8 __uint(max_entries, 1);
16 __uint(max_entries, 2);
23 __uint(max_entries, 3);
29 /* changing max_entries to 2 will fail during load
31 __uint(max_entries, 1);
47 __uint(max_entries, 3);
56 __uint(max_entries, 5);
63 __uint(max_entries, 3);
69 __uint(max_entries, 1);
83 __uint(max_entries,
[all...]
H A Dmap_ptr_kern.c34 __u32 max_entries; member
39 __u32 value_size, __u32 max_entries) in check_bpf_map_fields() argument
44 VERIFY(map->max_entries == max_entries); in check_bpf_map_fields()
56 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr()
63 __u32 key_size, __u32 value_size, __u32 max_entries) in check() argument
67 max_entries)); in check()
101 __uint(max_entries, MAX_ENTRIES);
142 __uint(max_entries, MAX_ENTRIE
[all...]
H A Dlsm.c16 __uint(max_entries, 1);
23 __uint(max_entries, 1);
30 __uint(max_entries, 1);
37 __uint(max_entries, 1);
44 __uint(max_entries, 1);
51 __uint(max_entries, 1);
58 __uint(max_entries, 1);
65 __uint(max_entries, 1);
75 __uint(max_entries, 1);
H A Dtest_unpriv_bpf_disabled.c17 __uint(max_entries, 1);
24 __uint(max_entries, 1);
31 __uint(max_entries, 1);
38 __uint(max_entries, 1);
51 __uint(max_entries, 1 << 12);
56 __uint(max_entries, 1);
H A Dverifier_map_ptr_mixing.c17 __uint(max_entries, 1);
24 __uint(max_entries, 1);
31 __uint(max_entries, 1);
36 __uint(max_entries, 1);
49 __uint(max_entries, 4);
62 __uint(max_entries, 8);
H A Dsockmap_verdict_prog.c8 __uint(max_entries, 20);
15 __uint(max_entries, 20);
22 __uint(max_entries, 20);
29 __uint(max_entries, 20);
H A Dtest_stacktrace_build_id.c13 __uint(max_entries, 1);
20 __uint(max_entries, 16384);
29 __uint(max_entries, 128);
37 __uint(max_entries, 128);
H A Dsyscall.c22 int max_entries; member
41 __uint(max_entries, 1);
48 __uint(max_entries, 1);
55 __uint(max_entries, 1);
132 map_create_attr.max_entries = ctx->max_entries; in load_prog()
182 attr->max_entries = 1; in update_outer_map()
H A Dtest_sockmap_listen.c12 __uint(max_entries, 2);
19 __uint(max_entries, 2);
26 __uint(max_entries, 2);
33 __uint(max_entries, 2);
40 __uint(max_entries, 1);
H A Dtest_stacktrace_map.c13 __uint(max_entries, 1);
20 __uint(max_entries, 16384);
29 __uint(max_entries, 16384);
36 __uint(max_entries, 16384);
H A Dtest_sockmap_drop_prog.c7 __uint(max_entries, 20);
14 __uint(max_entries, 20);
21 __uint(max_entries, 20);
H A Dupdate_map_in_htab.c10 __uint(max_entries, 1);
17 __uint(max_entries, 2);
26 __uint(max_entries, 2);
H A Dtest_pinning.c8 __uint(max_entries, 1);
16 __uint(max_entries, 1);
23 __uint(max_entries, 1);
H A Dtest_sockmap_kern.h34 __uint(max_entries, 20);
41 __uint(max_entries, 20);
48 __uint(max_entries, 20);
55 __uint(max_entries, 1);
62 __uint(max_entries, 1);
69 __uint(max_entries, 6);
76 __uint(max_entries, 1);
83 __uint(max_entries, 3);
90 __uint(max_entries, 20);
/linux/tools/testing/selftests/bpf/prog_tests/
H A Dfor_each.c15 int i, err, max_entries; in test_hash_map() local
31 max_entries = bpf_map__max_entries(skel->maps.hashmap); in test_hash_map()
32 for (i = 0; i < max_entries; i++) { in test_hash_map()
62 ASSERT_EQ(skel->bss->hashmap_elems, max_entries, "hashmap_elems"); in test_hash_map()
81 __u32 key, num_cpus, max_entries; in test_array_map() local
98 max_entries = bpf_map__max_entries(skel->maps.arraymap); in test_array_map()
99 for (i = 0; i < max_entries; i++) { in test_array_map()
103 if (i != max_entries - 1) in test_array_map()
152 __u32 key, max_entries; in test_multi_maps() local
166 max_entries in test_multi_maps()
210 int max_entries, i, err; test_hash_modify() local
[all...]
H A Dbtf.c72 __u32 max_entries; member
140 .max_entries = 4,
195 .max_entries = 4,
220 .max_entries = 4,
261 .max_entries = 4,
306 .max_entries = 1,
328 .max_entries = 1,
350 .max_entries = 1,
372 .max_entries = 1,
397 .max_entries
[all...]
/linux/kernel/bpf/
H A Darraymap.c26 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu()
37 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu()
58 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check()
87 u32 elem_size, index_mask, max_entries; in array_map_alloc() local
94 max_entries = attr->max_entries; in array_map_alloc()
96 /* On 32 bit archs roundup_pow_of_two() with max_entries that has in array_map_alloc()
100 mask64 = fls_long(max_entries - 1); in array_map_alloc()
109 max_entries = index_mask + 1; in array_map_alloc()
111 if (max_entries < att in array_map_alloc()
[all...]
/linux/samples/bpf/
H A Dmap_perf_test.bpf.c21 __uint(max_entries, MAX_ENTRIES);
28 __uint(max_entries, 10000);
35 __uint(max_entries, 10000);
43 __uint(max_entries, MAX_ENTRIES);
50 __uint(max_entries, MAX_NR_CPUS);
62 __uint(max_entries, MAX_ENTRIES);
69 __uint(max_entries, MAX_ENTRIES);
77 __uint(max_entries, MAX_ENTRIES);
85 __uint(max_entries, 10000);
93 __uint(max_entries, MAX_ENTRIE
[all...]
H A Dtest_map_in_map.bpf.c25 __uint(max_entries, MAX_NR_PORTS);
33 __uint(max_entries, 1);
41 __uint(max_entries, 1);
49 __uint(max_entries, 1);
55 __uint(max_entries, MAX_NR_PORTS);
63 __uint(max_entries, 1);
71 __uint(max_entries, 1);
/linux/drivers/md/persistent-data/
H A Ddm-btree-remove.c76 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift()
99 if (nr_left + shift > le32_to_cpu(left->header.max_entries)) { in node_copy()
111 if (shift > le32_to_cpu(right->header.max_entries)) { in node_copy()
152 return le32_to_cpu(n->header.max_entries) / 3; in merge_threshold()
197 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local
198 uint32_t r_max_entries = le32_to_cpu(right->header.max_entries); in shift()
200 if (max_entries != r_max_entries) { in shift()
201 DMERR("node max_entries mismatch"); in shift()
205 if (nr_left - count > max_entries) { in shift()
316 uint32_t max_entries = le32_to_cpu(left->header.max_entries); delete_center_node() local
357 uint32_t max_entries = le32_to_cpu(left->header.max_entries); redistribute3() local
[all...]
/linux/net/xdp/
H A Dxskmap.c70 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc()
76 size = struct_size(m, xsk_map, attr->max_entries); in xsk_map_alloc()
92 return struct_size(m, xsk_map, map->max_entries) + in xsk_map_mem_usage()
110 if (index >= m->map.max_entries) { in xsk_map_get_next_key()
115 if (index == m->map.max_entries - 1) in xsk_map_get_next_key()
127 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); in xsk_map_gen_lookup()
145 if (key >= map->max_entries) in __xsk_map_lookup_elem()
174 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem()
229 if (k >= map->max_entries) in xsk_map_delete_elem()
262 return meta0->max_entries in xsk_map_meta_equal()
[all...]
/linux/fs/ext4/
H A Dmigrate.c97 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_ind_extent_range() local
104 for (i = 0; i < max_entries; i++) { in update_ind_extent_range()
126 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_dind_extent_range() local
133 for (i = 0; i < max_entries; i++) { in update_dind_extent_range()
141 lb->curr_block += max_entries; in update_dind_extent_range()
156 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_tind_extent_range() local
163 for (i = 0; i < max_entries; i++) { in update_tind_extent_range()
171 lb->curr_block += max_entries * max_entries; in update_tind_extent_range()
186 unsigned long max_entries in free_dind_blocks() local
226 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; free_tind_blocks() local
418 unsigned long max_entries; ext4_ext_migrate() local
[all...]
/linux/lib/
H A Dfw_table.c135 * @max_entries: how many entries can we process?
154 int proc_num, unsigned int max_entries) in acpi_parse_entries_array() argument
180 if (!max_entries || count < max_entries) in acpi_parse_entries_array()
203 if (max_entries && count > max_entries) { in acpi_parse_entries_array()
205 id, proc->id, count - max_entries, count); in acpi_parse_entries_array()

12345678910>>...19