/Linux-v5.10/tools/testing/selftests/bpf/map_tests/ |
D | htab_map_batch_ops.c | 13 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument 27 for (i = 0; i < max_entries; i++) { in map_batch_update() 36 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 40 static void map_batch_verify(int *visited, __u32 max_entries, in map_batch_verify() argument 50 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify() 51 for (i = 0; i < max_entries; i++) { in map_batch_verify() 70 for (i = 0; i < max_entries; i++) { in map_batch_verify() 81 const __u32 max_entries = 10; in __test_map_lookup_and_delete_batch() local 82 value pcpu_values[max_entries]; in __test_map_lookup_and_delete_batch() 98 xattr.max_entries = max_entries; in __test_map_lookup_and_delete_batch() [all …]
|
D | array_map_batch_ops.c | 12 static void map_batch_update(int map_fd, __u32 max_entries, int *keys, in map_batch_update() argument 21 for (i = 0; i < max_entries; i++) { in map_batch_update() 26 err = bpf_map_update_batch(map_fd, keys, values, &max_entries, &opts); in map_batch_update() 30 static void map_batch_verify(int *visited, __u32 max_entries, in map_batch_verify() argument 35 memset(visited, 0, max_entries * sizeof(*visited)); in map_batch_verify() 36 for (i = 0; i < max_entries; i++) { in map_batch_verify() 41 for (i = 0; i < max_entries; i++) { in map_batch_verify() 57 const __u32 max_entries = 10; in test_array_map_batch_ops() local 66 xattr.max_entries = max_entries; in test_array_map_batch_ops() 71 keys = malloc(max_entries * sizeof(int)); in test_array_map_batch_ops() [all …]
|
/Linux-v5.10/tools/testing/selftests/bpf/progs/ |
D | test_btf_map_in_map.c | 8 __uint(max_entries, 1); 16 __uint(max_entries, 2); 23 __uint(max_entries, 3); 31 __uint(max_entries, 1); 47 __uint(max_entries, 3); 56 __uint(max_entries, 5); 63 __uint(max_entries, 3); 69 __uint(max_entries, 1); 83 __uint(max_entries, 5); 99 __uint(max_entries, 1); [all …]
|
D | map_ptr_kern.c | 37 __u32 max_entries; member 43 __u32 value_size, __u32 max_entries) in check_bpf_map_fields() argument 48 VERIFY(map->max_entries == max_entries); in check_bpf_map_fields() 61 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr() 69 __u32 key_size, __u32 value_size, __u32 max_entries) in check() argument 73 max_entries)); in check() 107 __uint(max_entries, MAX_ENTRIES); 143 __uint(max_entries, MAX_ENTRIES); 158 for (i = 0; i < array->map.max_entries && i < LOOP_BOUND; ++i) { in check_array() 175 __uint(max_entries, MAX_ENTRIES); [all …]
|
D | sockmap_verdict_prog.c | 9 __uint(max_entries, 20); 16 __uint(max_entries, 20); 23 __uint(max_entries, 20); 30 __uint(max_entries, 20);
|
D | test_stacktrace_build_id.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 128); 37 __uint(max_entries, 128);
|
D | test_stacktrace_map.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 16384); 36 __uint(max_entries, 16384);
|
D | test_pinning.c | 10 __uint(max_entries, 1); 18 __uint(max_entries, 1); 25 __uint(max_entries, 1);
|
D | test_sockmap_kern.h | 33 __uint(max_entries, 20); 40 __uint(max_entries, 20); 47 __uint(max_entries, 20); 54 __uint(max_entries, 1); 61 __uint(max_entries, 1); 68 __uint(max_entries, 6); 75 __uint(max_entries, 1); 82 __uint(max_entries, 3); 89 __uint(max_entries, 20);
|
D | test_sockmap_update.c | 8 __uint(max_entries, 1); 15 __uint(max_entries, 1); 22 __uint(max_entries, 1);
|
D | bpf_iter_sockmap.c | 13 __uint(max_entries, 64); 20 __uint(max_entries, 64); 27 __uint(max_entries, 64);
|
/Linux-v5.10/drivers/md/persistent-data/ |
D | dm-array.c | 25 __le32 max_entries; member 47 bh_le->csum = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_prepare_for_write() 66 csum_disk = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_check() 153 uint32_t max_entries, in alloc_ablock() argument 163 (*ab)->max_entries = cpu_to_le32(max_entries); in alloc_ablock() 182 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in fill_ablock() 206 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in trim_ablock() 352 uint32_t max_entries, in insert_new_ablock() argument 360 r = alloc_ablock(info, size_of_block, max_entries, &block, &ab); in insert_new_ablock() 373 unsigned max_entries, const void *value, in insert_full_ablocks() argument [all …]
|
D | dm-btree-remove.c | 72 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift() 90 BUG_ON(nr_left + shift > le32_to_cpu(left->header.max_entries)); in node_copy() 98 BUG_ON(shift > le32_to_cpu(right->header.max_entries)); in node_copy() 133 return le32_to_cpu(n->header.max_entries) / 3; in merge_threshold() 177 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local 178 uint32_t r_max_entries = le32_to_cpu(right->header.max_entries); in shift() 180 BUG_ON(max_entries != r_max_entries); in shift() 181 BUG_ON(nr_left - count > max_entries); in shift() 182 BUG_ON(nr_right + count > max_entries); in shift() 274 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in delete_center_node() local [all …]
|
/Linux-v5.10/tools/testing/selftests/bpf/prog_tests/ |
D | btf.c | 75 __u32 max_entries; member 143 .max_entries = 4, 198 .max_entries = 4, 223 .max_entries = 4, 264 .max_entries = 4, 309 .max_entries = 1, 331 .max_entries = 1, 353 .max_entries = 1, 375 .max_entries = 1, 400 .max_entries = 1, [all …]
|
D | tailcalls.c | 40 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 56 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 72 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 93 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 94 j = bpf_map__def(prog_array)->max_entries - 1 - i; in test_tailcall_1() 110 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 111 j = bpf_map__def(prog_array)->max_entries - 1 - i; in test_tailcall_1() 128 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_1() 178 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_2() 343 for (i = 0; i < bpf_map__def(prog_array)->max_entries; i++) { in test_tailcall_4() [all …]
|
/Linux-v5.10/samples/bpf/ |
D | map_perf_test_kern.c | 23 __uint(max_entries, MAX_ENTRIES); 30 __uint(max_entries, 10000); 37 __uint(max_entries, 10000); 45 __uint(max_entries, MAX_ENTRIES); 52 __uint(max_entries, MAX_NR_CPUS); 64 __uint(max_entries, MAX_ENTRIES); 71 __uint(max_entries, MAX_ENTRIES); 79 __uint(max_entries, MAX_ENTRIES); 87 __uint(max_entries, 10000); 95 __uint(max_entries, MAX_ENTRIES); [all …]
|
D | test_map_in_map_kern.c | 25 __uint(max_entries, MAX_NR_PORTS); 33 __uint(max_entries, 1); 41 __uint(max_entries, 1); 49 __uint(max_entries, 1); 55 __uint(max_entries, MAX_NR_PORTS); 63 __uint(max_entries, 1); 71 __uint(max_entries, 1);
|
D | xdp_monitor_kern.c | 13 __uint(max_entries, 2); 22 __uint(max_entries, XDP_UNKNOWN + 1); 136 __uint(max_entries, MAX_CPUS); 143 __uint(max_entries, 1); 217 __uint(max_entries, 1);
|
D | xdp_router_ipv4_kern.c | 49 __uint(max_entries, 50); 58 __uint(max_entries, 256); 66 __uint(max_entries, 50); 74 __uint(max_entries, 50); 81 __uint(max_entries, 100);
|
/Linux-v5.10/kernel/bpf/ |
D | arraymap.c | 25 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu() 36 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu() 57 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check() 85 u32 elem_size, index_mask, max_entries; in array_map_alloc() local 93 max_entries = attr->max_entries; in array_map_alloc() 99 mask64 = fls_long(max_entries - 1); in array_map_alloc() 108 max_entries = index_mask + 1; in array_map_alloc() 110 if (max_entries < attr->max_entries) in array_map_alloc() 116 array_size += (u64) max_entries * sizeof(void *); in array_map_alloc() 123 array_size += PAGE_ALIGN((u64) max_entries * elem_size); in array_map_alloc() [all …]
|
D | reuseport_array.c | 57 if (unlikely(index >= array->map.max_entries)) in reuseport_array_lookup_elem() 71 if (index >= map->max_entries) in reuseport_array_delete_elem() 128 for (i = 0; i < map->max_entries; i++) { in reuseport_array_free() 162 array_size += (u64)attr->max_entries * sizeof(struct sock *); in reuseport_array_alloc() 263 if (index >= map->max_entries) in bpf_fd_reuseport_array_update_elem() 340 if (index >= array->map.max_entries) { in reuseport_array_get_next_key() 345 if (index == array->map.max_entries - 1) in reuseport_array_get_next_key()
|
/Linux-v5.10/tools/lib/perf/ |
D | cpumap.c | 104 int max_entries = 0; in perf_cpu_map__read() local 120 if (new_max >= max_entries) { in perf_cpu_map__read() 121 max_entries = new_max + MAX_NR_CPUS / 2; in perf_cpu_map__read() 122 tmp = realloc(tmp_cpus, max_entries * sizeof(int)); in perf_cpu_map__read() 131 if (nr_cpus == max_entries) { in perf_cpu_map__read() 132 max_entries += MAX_NR_CPUS; in perf_cpu_map__read() 133 tmp = realloc(tmp_cpus, max_entries * sizeof(int)); in perf_cpu_map__read() 178 int max_entries = 0; in perf_cpu_map__new() local 221 if (nr_cpus == max_entries) { in perf_cpu_map__new() 222 max_entries += MAX_NR_CPUS; in perf_cpu_map__new() [all …]
|
/Linux-v5.10/fs/ext4/ |
D | migrate.c | 97 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_ind_extent_range() local 104 for (i = 0; i < max_entries; i++) { in update_ind_extent_range() 126 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_dind_extent_range() local 133 for (i = 0; i < max_entries; i++) { in update_dind_extent_range() 141 lb->curr_block += max_entries; in update_dind_extent_range() 156 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_tind_extent_range() local 163 for (i = 0; i < max_entries; i++) { in update_tind_extent_range() 171 lb->curr_block += max_entries * max_entries; in update_tind_extent_range() 186 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in free_dind_blocks() local 194 for (i = 0; i < max_entries; i++) { in free_dind_blocks() [all …]
|
/Linux-v5.10/net/xdp/ |
D | xskmap.c | 84 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc() 90 size = struct_size(m, xsk_map, attr->max_entries); in xsk_map_alloc() 124 if (index >= m->map.max_entries) { in xsk_map_get_next_key() 129 if (index == m->map.max_entries - 1) in xsk_map_get_next_key() 141 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); in xsk_map_gen_lookup() 174 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem() 228 if (k >= map->max_entries) in xsk_map_delete_elem() 255 return meta0->max_entries == meta1->max_entries && in xsk_map_meta_equal()
|
/Linux-v5.10/tools/lib/bpf/ |
D | libbpf_probes.c | 202 int key_size, value_size, max_entries, map_flags; in bpf_probe_map_type() local 209 max_entries = 1; in bpf_probe_map_type() 225 max_entries = 0; in bpf_probe_map_type() 236 max_entries = 0; in bpf_probe_map_type() 245 max_entries = 4096; in bpf_probe_map_type() 291 attr.max_entries = max_entries; in bpf_probe_map_type()
|