| /Linux-v5.4/drivers/md/persistent-data/ |
| D | dm-array.c | 25 __le32 max_entries; member 47 bh_le->csum = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_prepare_for_write() 66 csum_disk = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_check() 153 uint32_t max_entries, in alloc_ablock() argument 163 (*ab)->max_entries = cpu_to_le32(max_entries); in alloc_ablock() 182 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in fill_ablock() 206 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in trim_ablock() 352 uint32_t max_entries, in insert_new_ablock() argument 360 r = alloc_ablock(info, size_of_block, max_entries, &block, &ab); in insert_new_ablock() 373 unsigned max_entries, const void *value, in insert_full_ablocks() argument [all …]
|
| D | dm-btree-remove.c | 72 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift() 90 BUG_ON(nr_left + shift > le32_to_cpu(left->header.max_entries)); in node_copy() 98 BUG_ON(shift > le32_to_cpu(right->header.max_entries)); in node_copy() 133 return le32_to_cpu(n->header.max_entries) / 3; in merge_threshold() 177 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local 178 uint32_t r_max_entries = le32_to_cpu(right->header.max_entries); in shift() 180 BUG_ON(max_entries != r_max_entries); in shift() 181 BUG_ON(nr_left - count > max_entries); in shift() 182 BUG_ON(nr_right + count > max_entries); in shift() 268 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in delete_center_node() local [all …]
|
| /Linux-v5.4/kernel/bpf/ |
| D | arraymap.c | 23 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu() 34 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu() 55 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check() 75 u32 elem_size, index_mask, max_entries; in array_map_alloc() local 83 max_entries = attr->max_entries; in array_map_alloc() 89 mask64 = fls_long(max_entries - 1); in array_map_alloc() 98 max_entries = index_mask + 1; in array_map_alloc() 100 if (max_entries < attr->max_entries) in array_map_alloc() 106 array_size += (u64) max_entries * sizeof(void *); in array_map_alloc() 108 array_size += (u64) max_entries * elem_size; in array_map_alloc() [all …]
|
| D | xskmap.c | 90 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc() 102 cost = (u64)m->map.max_entries * sizeof(struct xdp_sock *); in xsk_map_alloc() 119 m->xsk_map = bpf_map_area_alloc(m->map.max_entries * in xsk_map_alloc() 152 if (index >= m->map.max_entries) { in xsk_map_get_next_key() 157 if (index == m->map.max_entries - 1) in xsk_map_get_next_key() 168 if (key >= map->max_entries) in __xsk_map_lookup_elem() 227 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem() 286 if (k >= map->max_entries) in xsk_map_delete_elem()
|
| D | map_in_map.c | 57 inner_map_meta->max_entries = inner_map->max_entries; in bpf_map_meta_alloc() 85 meta0->max_entries == meta1->max_entries; in bpf_map_meta_equal()
|
| D | reuseport_array.c | 54 if (unlikely(index >= array->map.max_entries)) in reuseport_array_lookup_elem() 68 if (index >= map->max_entries) in reuseport_array_delete_elem() 127 for (i = 0; i < map->max_entries; i++) { in reuseport_array_free() 161 array_size += (u64)attr->max_entries * sizeof(struct sock *); in reuseport_array_alloc() 261 if (index >= map->max_entries) in bpf_fd_reuseport_array_update_elem() 341 if (index >= array->map.max_entries) { in reuseport_array_get_next_key() 346 if (index == array->map.max_entries - 1) in reuseport_array_get_next_key()
|
| D | cpumap.c | 91 if (attr->max_entries == 0 || attr->key_size != 4 || in cpu_map_alloc() 102 if (cmap->map.max_entries > NR_CPUS) { in cpu_map_alloc() 108 cost = (u64) cmap->map.max_entries * sizeof(struct bpf_cpu_map_entry *); in cpu_map_alloc() 126 cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries * in cpu_map_alloc() 460 if (key_cpu >= map->max_entries) in cpu_map_delete_elem() 481 if (unlikely(key_cpu >= cmap->map.max_entries)) in cpu_map_update_elem() 540 for (i = 0; i < cmap->map.max_entries; i++) { in cpu_map_free() 560 if (key >= map->max_entries) in __cpu_map_lookup_elem() 581 if (index >= cmap->map.max_entries) { in cpu_map_get_next_key() 586 if (index == cmap->map.max_entries - 1) in cpu_map_get_next_key()
|
| /Linux-v5.4/tools/perf/lib/ |
| D | cpumap.c | 90 int max_entries = 0; in perf_cpu_map__read() local 106 if (new_max >= max_entries) { in perf_cpu_map__read() 107 max_entries = new_max + MAX_NR_CPUS / 2; in perf_cpu_map__read() 108 tmp = realloc(tmp_cpus, max_entries * sizeof(int)); in perf_cpu_map__read() 117 if (nr_cpus == max_entries) { in perf_cpu_map__read() 118 max_entries += MAX_NR_CPUS; in perf_cpu_map__read() 119 tmp = realloc(tmp_cpus, max_entries * sizeof(int)); in perf_cpu_map__read() 164 int max_entries = 0; in perf_cpu_map__new() local 207 if (nr_cpus == max_entries) { in perf_cpu_map__new() 208 max_entries += MAX_NR_CPUS; in perf_cpu_map__new() [all …]
|
| /Linux-v5.4/samples/bpf/ |
| D | map_perf_test_kern.c | 20 .max_entries = MAX_ENTRIES, 27 .max_entries = 10000, 34 .max_entries = 10000, 42 .max_entries = MAX_ENTRIES, 50 .max_entries = MAX_NR_CPUS, 57 .max_entries = MAX_ENTRIES, 64 .max_entries = MAX_ENTRIES, 72 .max_entries = MAX_ENTRIES, 80 .max_entries = 10000, 88 .max_entries = MAX_ENTRIES, [all …]
|
| D | test_map_in_map_kern.c | 22 .max_entries = MAX_NR_PORTS, 30 .max_entries = 1, 38 .max_entries = 1, 46 .max_entries = 1, 54 .max_entries = MAX_NR_PORTS, 62 .max_entries = 1, 70 .max_entries = 1,
|
| D | xdp_monitor_kern.c | 13 .max_entries = 2, 22 .max_entries = XDP_UNKNOWN + 1, 136 .max_entries = MAX_CPUS, 143 .max_entries = 1, 217 .max_entries = 1,
|
| D | xdp_router_ipv4_kern.c | 49 .max_entries = 50, 58 .max_entries = 256, 66 .max_entries = 50, 74 .max_entries = 50, 81 .max_entries = 100,
|
| D | tracex6_kern.c | 10 .max_entries = 64, 16 .max_entries = 64, 22 .max_entries = 64,
|
| D | spintest_kern.c | 18 .max_entries = 1024, 24 .max_entries = 1024, 31 .max_entries = 10000,
|
| /Linux-v5.4/tools/testing/selftests/bpf/ |
| D | test_btf.c | 113 __u32 max_entries; member 181 .max_entries = 4, 236 .max_entries = 4, 261 .max_entries = 4, 302 .max_entries = 4, 347 .max_entries = 1, 369 .max_entries = 1, 391 .max_entries = 1, 413 .max_entries = 1, 438 .max_entries = 1, [all …]
|
| /Linux-v5.4/fs/ext4/ |
| D | migrate.c | 117 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_ind_extent_range() local 124 for (i = 0; i < max_entries; i++) { in update_ind_extent_range() 146 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_dind_extent_range() local 153 for (i = 0; i < max_entries; i++) { in update_dind_extent_range() 161 lb->curr_block += max_entries; in update_dind_extent_range() 176 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_tind_extent_range() local 183 for (i = 0; i < max_entries; i++) { in update_tind_extent_range() 191 lb->curr_block += max_entries * max_entries; in update_tind_extent_range() 225 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in free_dind_blocks() local 232 for (i = 0; i < max_entries; i++) { in free_dind_blocks() [all …]
|
| /Linux-v5.4/tools/testing/selftests/bpf/progs/ |
| D | sockmap_verdict_prog.c | 9 __uint(max_entries, 20); 16 __uint(max_entries, 20); 23 __uint(max_entries, 20); 30 __uint(max_entries, 20);
|
| D | test_stacktrace_build_id.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 128); 37 __uint(max_entries, 128);
|
| D | test_stacktrace_map.c | 13 __uint(max_entries, 1); 20 __uint(max_entries, 16384); 29 __uint(max_entries, 16384); 36 __uint(max_entries, 16384);
|
| D | test_select_reuseport_kern.c | 26 __uint(max_entries, 1); 33 __uint(max_entries, NR_RESULTS); 40 __uint(max_entries, 1); 47 __uint(max_entries, 1); 54 __uint(max_entries, 1);
|
| D | pyperf.h | 115 __uint(max_entries, 1); 122 __uint(max_entries, 1); 129 __uint(max_entries, 1); 136 __uint(max_entries, 1); 143 __uint(max_entries, 32); 150 __uint(max_entries, 1000);
|
| /Linux-v5.4/tools/lib/bpf/ |
| D | bpf.h | 45 __u32 max_entries; member 58 int max_entries, __u32 map_flags, int node); 61 int max_entries, __u32 map_flags); 63 int value_size, int max_entries, __u32 map_flags); 66 int inner_map_fd, int max_entries, 70 int inner_map_fd, int max_entries,
|
| D | bpf.c | 88 attr.max_entries = create_attr->max_entries; in bpf_create_map_xattr() 104 int key_size, int value_size, int max_entries, in bpf_create_map_node() argument 114 map_attr.max_entries = max_entries; in bpf_create_map_node() 124 int value_size, int max_entries, __u32 map_flags) in bpf_create_map() argument 132 map_attr.max_entries = max_entries; in bpf_create_map() 138 int key_size, int value_size, int max_entries, in bpf_create_map_name() argument 148 map_attr.max_entries = max_entries; in bpf_create_map_name() 154 int key_size, int inner_map_fd, int max_entries, in bpf_create_map_in_map_node() argument 165 attr.max_entries = max_entries; in bpf_create_map_in_map_node() 180 int key_size, int inner_map_fd, int max_entries, in bpf_create_map_in_map() argument [all …]
|
| D | libbpf_probes.c | 195 int key_size, value_size, max_entries, map_flags; in bpf_probe_map_type() local 202 max_entries = 1; in bpf_probe_map_type() 218 max_entries = 0; in bpf_probe_map_type() 228 max_entries = 0; in bpf_probe_map_type() 277 attr.max_entries = max_entries; in bpf_probe_map_type()
|
| /Linux-v5.4/kernel/ |
| D | stacktrace.c | 277 .max_entries = size, in stack_trace_save() 301 .max_entries = size, in stack_trace_save_tsk() 324 .max_entries = size, in stack_trace_save_regs() 350 .max_entries = size, in stack_trace_save_tsk_reliable() 370 .max_entries = size, in stack_trace_save_user()
|