Home
last modified time | relevance | path

Searched refs:max_entries (Results 1 – 25 of 288) sorted by relevance

12345678910>>...12

/Linux-v5.15/tools/testing/selftests/bpf/map_tests/
Dhtab_map_batch_ops.c13 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument
27 for (i = 0; i < max_entries; i++) { in map_batch_update()
36 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update()
40 static void map_batch_verify(int *visited, __u32 max_entries, in map_batch_verify() argument
50 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify()
51 for (i = 0; i < max_entries; i++) { in map_batch_verify()
70 for (i = 0; i < max_entries; i++) { in map_batch_verify()
81 const __u32 max_entries = 10; in __test_map_lookup_and_delete_batch() local
82 value pcpu_values[max_entries]; in __test_map_lookup_and_delete_batch()
98 xattr.max_entries = max_entries; in __test_map_lookup_and_delete_batch()
[all …]
Dlpm_trie_map_batch_ops.c21 static void map_batch_update(int map_fd, __u32 max_entries, in map_batch_update() argument
32 for (i = 0; i < max_entries; i++) { in map_batch_update()
39 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update()
43 static void map_batch_verify(int *visited, __u32 max_entries, in map_batch_verify() argument
50 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify()
51 for (i = 0; i < max_entries; i++) { in map_batch_verify()
59 for (i = 0; i < max_entries; i++) { in map_batch_verify()
77 const __u32 max_entries = 10; in test_lpm_trie_map_batch_ops() local
85 xattr.max_entries = max_entries; in test_lpm_trie_map_batch_ops()
90 keys = malloc(max_entries * sizeof(struct test_lpm_key)); in test_lpm_trie_map_batch_ops()
[all …]
Darray_map_batch_ops.c14 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument
24 for (i = 0; i < max_entries; i++) { in map_batch_update()
35 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update()
39 static void map_batch_verify(int *visited, __u32 max_entries, int *keys, in map_batch_verify() argument
45 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify()
46 for (i = 0; i < max_entries; i++) { in map_batch_verify()
63 for (i = 0; i < max_entries; i++) { in map_batch_verify()
80 const __u32 max_entries = 10; in __test_map_lookup_and_update_batch() local
89 xattr.max_entries = max_entries; in __test_map_lookup_and_update_batch()
98 keys = calloc(max_entries, sizeof(*keys)); in __test_map_lookup_and_update_batch()
[all …]
/Linux-v5.15/tools/testing/selftests/bpf/progs/
Dtest_btf_map_in_map.c8 __uint(max_entries, 1);
16 __uint(max_entries, 2);
23 __uint(max_entries, 3);
31 __uint(max_entries, 1);
47 __uint(max_entries, 3);
56 __uint(max_entries, 5);
63 __uint(max_entries, 3);
69 __uint(max_entries, 1);
83 __uint(max_entries, 5);
99 __uint(max_entries, 1);
[all …]
Dmap_ptr_kern.c34 __u32 max_entries; member
39 __u32 value_size, __u32 max_entries) in check_bpf_map_fields() argument
44 VERIFY(map->max_entries == max_entries); in check_bpf_map_fields()
56 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr()
63 __u32 key_size, __u32 value_size, __u32 max_entries) in check() argument
67 max_entries)); in check()
101 __uint(max_entries, MAX_ENTRIES);
137 __uint(max_entries, MAX_ENTRIES);
152 for (i = 0; i < array->map.max_entries && i < LOOP_BOUND; ++i) { in check_array()
169 __uint(max_entries, MAX_ENTRIES);
[all …]
Dlsm.c14 __uint(max_entries, 1);
21 __uint(max_entries, 1);
28 __uint(max_entries, 1);
35 __uint(max_entries, 1);
42 __uint(max_entries, 1);
49 __uint(max_entries, 1);
56 __uint(max_entries, 1);
63 __uint(max_entries, 1);
73 __uint(max_entries, 1);
Dsockmap_verdict_prog.c9 __uint(max_entries, 20);
16 __uint(max_entries, 20);
23 __uint(max_entries, 20);
30 __uint(max_entries, 20);
Dtest_stacktrace_build_id.c13 __uint(max_entries, 1);
20 __uint(max_entries, 16384);
29 __uint(max_entries, 128);
37 __uint(max_entries, 128);
Dtest_stacktrace_map.c13 __uint(max_entries, 1);
20 __uint(max_entries, 16384);
29 __uint(max_entries, 16384);
36 __uint(max_entries, 16384);
Dtest_pinning.c10 __uint(max_entries, 1);
18 __uint(max_entries, 1);
25 __uint(max_entries, 1);
Dtest_sockmap_kern.h33 __uint(max_entries, 20);
40 __uint(max_entries, 20);
47 __uint(max_entries, 20);
54 __uint(max_entries, 1);
61 __uint(max_entries, 1);
68 __uint(max_entries, 6);
75 __uint(max_entries, 1);
82 __uint(max_entries, 3);
89 __uint(max_entries, 20);
Dtest_sockmap_update.c8 __uint(max_entries, 1);
15 __uint(max_entries, 1);
22 __uint(max_entries, 1);
Dbpf_iter_sockmap.c13 __uint(max_entries, 64);
20 __uint(max_entries, 64);
27 __uint(max_entries, 64);
/Linux-v5.15/drivers/md/persistent-data/
Ddm-array.c25 __le32 max_entries; member
47 bh_le->csum = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_prepare_for_write()
66 csum_disk = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_check()
151 uint32_t max_entries, in alloc_ablock() argument
161 (*ab)->max_entries = cpu_to_le32(max_entries); in alloc_ablock()
179 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in fill_ablock()
202 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in trim_ablock()
348 uint32_t max_entries, in insert_new_ablock() argument
356 r = alloc_ablock(info, size_of_block, max_entries, &block, &ab); in insert_new_ablock()
369 unsigned max_entries, const void *value, in insert_full_ablocks() argument
[all …]
Ddm-btree-remove.c72 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift()
90 BUG_ON(nr_left + shift > le32_to_cpu(left->header.max_entries)); in node_copy()
98 BUG_ON(shift > le32_to_cpu(right->header.max_entries)); in node_copy()
133 return le32_to_cpu(n->header.max_entries) / 3; in merge_threshold()
177 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local
178 uint32_t r_max_entries = le32_to_cpu(right->header.max_entries); in shift()
180 BUG_ON(max_entries != r_max_entries); in shift()
181 BUG_ON(nr_left - count > max_entries); in shift()
182 BUG_ON(nr_right + count > max_entries); in shift()
274 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in delete_center_node() local
[all …]
/Linux-v5.15/tools/testing/selftests/bpf/prog_tests/
Dbtf.c75 __u32 max_entries; member
143 .max_entries = 4,
198 .max_entries = 4,
223 .max_entries = 4,
264 .max_entries = 4,
309 .max_entries = 1,
331 .max_entries = 1,
353 .max_entries = 1,
375 .max_entries = 1,
400 .max_entries = 1,
[all …]
Dfor_each.c12 int i, err, hashmap_fd, max_entries, percpu_map_fd; in test_hash_map() local
23 max_entries = bpf_map__max_entries(skel->maps.hashmap); in test_hash_map()
24 for (i = 0; i < max_entries; i++) { in test_hash_map()
53 ASSERT_EQ(skel->bss->hashmap_elems, max_entries, "hashmap_elems"); in test_hash_map()
72 __u32 key, num_cpus, max_entries, retval; in test_array_map() local
84 max_entries = bpf_map__max_entries(skel->maps.arraymap); in test_array_map()
85 for (i = 0; i < max_entries; i++) { in test_array_map()
89 if (i != max_entries - 1) in test_array_map()
Dtailcalls.c40 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1()
56 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1()
72 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1()
93 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1()
94 j = bpf_map__def(prog_array)->max_entries - 1 - i; in test_tailcall_1()
110 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1()
111 j = bpf_map__def(prog_array)->max_entries - 1 - i; in test_tailcall_1()
128 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1()
178 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_2()
343 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_4()
[all …]
/Linux-v5.15/kernel/bpf/
Darraymap.c25 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu()
36 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu()
57 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check()
85 u32 elem_size, index_mask, max_entries; in array_map_alloc() local
92 max_entries = attr->max_entries; in array_map_alloc()
98 mask64 = fls_long(max_entries - 1); in array_map_alloc()
107 max_entries = index_mask + 1; in array_map_alloc()
109 if (max_entries < attr->max_entries) in array_map_alloc()
115 array_size += (u64) max_entries * sizeof(void *); in array_map_alloc()
122 array_size += PAGE_ALIGN((u64) max_entries * elem_size); in array_map_alloc()
[all …]
/Linux-v5.15/samples/bpf/
Dmap_perf_test_kern.c23 __uint(max_entries, MAX_ENTRIES);
30 __uint(max_entries, 10000);
37 __uint(max_entries, 10000);
45 __uint(max_entries, MAX_ENTRIES);
52 __uint(max_entries, MAX_NR_CPUS);
64 __uint(max_entries, MAX_ENTRIES);
71 __uint(max_entries, MAX_ENTRIES);
79 __uint(max_entries, MAX_ENTRIES);
87 __uint(max_entries, 10000);
95 __uint(max_entries, MAX_ENTRIES);
[all …]
Dtest_map_in_map_kern.c25 __uint(max_entries, MAX_NR_PORTS);
33 __uint(max_entries, 1);
41 __uint(max_entries, 1);
49 __uint(max_entries, 1);
55 __uint(max_entries, MAX_NR_PORTS);
63 __uint(max_entries, 1);
71 __uint(max_entries, 1);
/Linux-v5.15/tools/lib/perf/
Dcpumap.c109 int max_entries = 0; in perf_cpu_map__read() local
125 if (new_max >= max_entries) { in perf_cpu_map__read()
126 max_entries = new_max + MAX_NR_CPUS / 2; in perf_cpu_map__read()
127 tmp = realloc(tmp_cpus, max_entries * sizeof(int)); in perf_cpu_map__read()
136 if (nr_cpus == max_entries) { in perf_cpu_map__read()
137 max_entries += MAX_NR_CPUS; in perf_cpu_map__read()
138 tmp = realloc(tmp_cpus, max_entries * sizeof(int)); in perf_cpu_map__read()
183 int max_entries = 0; in perf_cpu_map__new() local
226 if (nr_cpus == max_entries) { in perf_cpu_map__new()
227 max_entries += MAX_NR_CPUS; in perf_cpu_map__new()
[all …]
/Linux-v5.15/fs/ext4/
Dmigrate.c97 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_ind_extent_range() local
104 for (i = 0; i < max_entries; i++) { in update_ind_extent_range()
126 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_dind_extent_range() local
133 for (i = 0; i < max_entries; i++) { in update_dind_extent_range()
141 lb->curr_block += max_entries; in update_dind_extent_range()
156 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_tind_extent_range() local
163 for (i = 0; i < max_entries; i++) { in update_tind_extent_range()
171 lb->curr_block += max_entries * max_entries; in update_tind_extent_range()
186 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in free_dind_blocks() local
194 for (i = 0; i < max_entries; i++) { in free_dind_blocks()
[all …]
/Linux-v5.15/net/xdp/
Dxskmap.c68 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc()
74 size = struct_size(m, xsk_map, attr->max_entries); in xsk_map_alloc()
100 if (index >= m->map.max_entries) { in xsk_map_get_next_key()
105 if (index == m->map.max_entries - 1) in xsk_map_get_next_key()
117 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); in xsk_map_gen_lookup()
135 if (key >= map->max_entries) in __xsk_map_lookup_elem()
164 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem()
219 if (k >= map->max_entries) in xsk_map_delete_elem()
252 return meta0->max_entries == meta1->max_entries && in xsk_map_meta_equal()
/Linux-v5.15/tools/lib/bpf/
Dlibbpf_probes.c204 int key_size, value_size, max_entries, map_flags; in bpf_probe_map_type() local
211 max_entries = 1; in bpf_probe_map_type()
227 max_entries = 0; in bpf_probe_map_type()
239 max_entries = 0; in bpf_probe_map_type()
248 max_entries = 4096; in bpf_probe_map_type()
294 attr.max_entries = max_entries; in bpf_probe_map_type()

12345678910>>...12