/Linux-v4.19/drivers/md/persistent-data/ |
D | dm-array.c | 25 __le32 max_entries; member 47 bh_le->csum = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_prepare_for_write() 66 csum_disk = cpu_to_le32(dm_bm_checksum(&bh_le->max_entries, in array_block_check() 153 uint32_t max_entries, in alloc_ablock() argument 163 (*ab)->max_entries = cpu_to_le32(max_entries); in alloc_ablock() 182 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in fill_ablock() 206 BUG_ON(new_nr > le32_to_cpu(ab->max_entries)); in trim_ablock() 352 uint32_t max_entries, in insert_new_ablock() argument 360 r = alloc_ablock(info, size_of_block, max_entries, &block, &ab); in insert_new_ablock() 373 unsigned max_entries, const void *value, in insert_full_ablocks() argument [all …]
|
D | dm-btree-remove.c | 72 BUG_ON(nr_entries + shift > le32_to_cpu(n->header.max_entries)); in node_shift() 90 BUG_ON(nr_left + shift > le32_to_cpu(left->header.max_entries)); in node_copy() 98 BUG_ON(shift > le32_to_cpu(right->header.max_entries)); in node_copy() 133 return le32_to_cpu(n->header.max_entries) / 3; in merge_threshold() 177 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in shift() local 178 uint32_t r_max_entries = le32_to_cpu(right->header.max_entries); in shift() 180 BUG_ON(max_entries != r_max_entries); in shift() 181 BUG_ON(nr_left - count > max_entries); in shift() 182 BUG_ON(nr_right + count > max_entries); in shift() 268 uint32_t max_entries = le32_to_cpu(left->header.max_entries); in delete_center_node() local [all …]
|
/Linux-v4.19/kernel/bpf/ |
D | arraymap.c | 31 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu() 42 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu() 63 if (attr->max_entries == 0 || attr->key_size != 4 || in array_map_alloc_check() 82 u32 elem_size, index_mask, max_entries; in array_map_alloc() local 89 max_entries = attr->max_entries; in array_map_alloc() 95 mask64 = fls_long(max_entries - 1); in array_map_alloc() 104 max_entries = index_mask + 1; in array_map_alloc() 106 if (max_entries < attr->max_entries) in array_map_alloc() 112 array_size += (u64) max_entries * sizeof(void *); in array_map_alloc() 114 array_size += (u64) max_entries * elem_size; in array_map_alloc() [all …]
|
D | xskmap.c | 27 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc() 38 cost = (u64)m->map.max_entries * sizeof(struct xdp_sock *); in xsk_map_alloc() 59 m->xsk_map = bpf_map_area_alloc(m->map.max_entries * in xsk_map_alloc() 81 for (i = 0; i < map->max_entries; i++) { in xsk_map_free() 102 if (index >= m->map.max_entries) { in xsk_map_get_next_key() 107 if (index == m->map.max_entries - 1) in xsk_map_get_next_key() 118 if (key >= map->max_entries) in __xsk_map_lookup_elem() 171 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem() 208 if (k >= map->max_entries) in xsk_map_delete_elem()
|
D | cpumap.c | 76 return BITS_TO_LONGS(attr->max_entries) * sizeof(unsigned long); in cpu_map_bitmap_size() 90 if (attr->max_entries == 0 || attr->key_size != 4 || in cpu_map_alloc() 101 if (cmap->map.max_entries > NR_CPUS) { in cpu_map_alloc() 107 cost = (u64) cmap->map.max_entries * sizeof(struct bpf_cpu_map_entry *); in cpu_map_alloc() 127 cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries * in cpu_map_alloc() 422 if (key_cpu >= map->max_entries) in cpu_map_delete_elem() 443 if (unlikely(key_cpu >= cmap->map.max_entries)) in cpu_map_update_elem() 494 while (!bitmap_empty(bitmap, cmap->map.max_entries)) in cpu_map_free() 501 for (i = 0; i < cmap->map.max_entries; i++) { in cpu_map_free() 521 if (key >= map->max_entries) in __cpu_map_lookup_elem() [all …]
|
D | devmap.c | 85 return BITS_TO_LONGS((u64) attr->max_entries) * sizeof(unsigned long); in dev_map_bitmap_size() 98 if (attr->max_entries == 0 || attr->key_size != 4 || in dev_map_alloc() 109 cost = (u64) dtab->map.max_entries * sizeof(struct bpf_dtab_netdev *); in dev_map_alloc() 130 dtab->netdev_map = bpf_map_area_alloc(dtab->map.max_entries * in dev_map_alloc() 175 while (!bitmap_empty(bitmap, dtab->map.max_entries)) in dev_map_free() 179 for (i = 0; i < dtab->map.max_entries; i++) { in dev_map_free() 201 if (index >= dtab->map.max_entries) { in dev_map_get_next_key() 206 if (index == dtab->map.max_entries - 1) in dev_map_get_next_key() 281 for_each_set_bit(bit, bitmap, map->max_entries) { in __dev_map_flush() 307 if (key >= map->max_entries) in __dev_map_lookup_elem() [all …]
|
D | map_in_map.c | 49 inner_map_meta->max_entries = inner_map->max_entries; in bpf_map_meta_alloc() 68 meta0->max_entries == meta1->max_entries; in bpf_map_meta_equal()
|
/Linux-v4.19/samples/bpf/ |
D | map_perf_test_kern.c | 20 .max_entries = MAX_ENTRIES, 27 .max_entries = 10000, 34 .max_entries = 10000, 42 .max_entries = MAX_ENTRIES, 50 .max_entries = MAX_NR_CPUS, 57 .max_entries = MAX_ENTRIES, 64 .max_entries = MAX_ENTRIES, 72 .max_entries = MAX_ENTRIES, 80 .max_entries = 10000, 88 .max_entries = MAX_ENTRIES, [all …]
|
D | test_map_in_map_kern.c | 22 .max_entries = MAX_NR_PORTS, 30 .max_entries = 1, 38 .max_entries = 1, 46 .max_entries = 1, 54 .max_entries = MAX_NR_PORTS, 62 .max_entries = 1, 70 .max_entries = 1,
|
D | xdpsock_kern.c | 12 .max_entries = 1, 19 .max_entries = 4, 26 .max_entries = 1,
|
D | xdp_router_ipv4_kern.c | 49 .max_entries = 50, 58 .max_entries = 256, 66 .max_entries = 50, 74 .max_entries = 50, 81 .max_entries = 100,
|
/Linux-v4.19/tools/testing/selftests/bpf/ |
D | test_btf.c | 128 __u32 max_entries; member 191 .max_entries = 4, 246 .max_entries = 4, 272 .max_entries = 4, 304 .max_entries = 4, 338 .max_entries = 4, 372 .max_entries = 4, 414 .max_entries = 4, 450 .max_entries = 4, 479 .max_entries = 4, [all …]
|
D | test_stacktrace_map.c | 15 .max_entries = 1, 22 .max_entries = 16384, 29 .max_entries = 16384, 36 .max_entries = 16384,
|
D | sockmap_verdict_prog.c | 19 .max_entries = 20, 26 .max_entries = 20, 33 .max_entries = 20, 40 .max_entries = 20,
|
D | test_stacktrace_build_id.c | 15 .max_entries = 1, 22 .max_entries = 16384, 30 .max_entries = 128, 39 .max_entries = 128,
|
D | test_sockmap_kern.h | 42 .max_entries = 20, 49 .max_entries = 20, 56 .max_entries = 20, 63 .max_entries = 1 70 .max_entries = 1 77 .max_entries = 2 84 .max_entries = 1 91 .max_entries = 1
|
D | test_select_reuseport_kern.c | 28 .max_entries = 1, 35 .max_entries = NR_RESULTS, 42 .max_entries = 1, 49 .max_entries = 1, 56 .max_entries = 1,
|
/Linux-v4.19/fs/ext4/ |
D | migrate.c | 117 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_ind_extent_range() local 124 for (i = 0; i < max_entries; i++) { in update_ind_extent_range() 146 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_dind_extent_range() local 153 for (i = 0; i < max_entries; i++) { in update_dind_extent_range() 161 lb->curr_block += max_entries; in update_dind_extent_range() 176 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in update_tind_extent_range() local 183 for (i = 0; i < max_entries; i++) { in update_tind_extent_range() 191 lb->curr_block += max_entries * max_entries; in update_tind_extent_range() 225 unsigned long max_entries = inode->i_sb->s_blocksize >> 2; in free_dind_blocks() local 232 for (i = 0; i < max_entries; i++) { in free_dind_blocks() [all …]
|
/Linux-v4.19/tools/lib/bpf/ |
D | bpf.h | 36 __u32 max_entries; member 47 int key_size, int value_size, int max_entries, 50 int key_size, int value_size, int max_entries, 53 int max_entries, __u32 map_flags); 55 int key_size, int inner_map_fd, int max_entries, 58 int key_size, int inner_map_fd, int max_entries,
|
D | bpf.c | 86 attr.max_entries = create_attr->max_entries; in bpf_create_map_xattr() 101 int key_size, int value_size, int max_entries, in bpf_create_map_node() argument 111 map_attr.max_entries = max_entries; in bpf_create_map_node() 121 int value_size, int max_entries, __u32 map_flags) in bpf_create_map() argument 129 map_attr.max_entries = max_entries; in bpf_create_map() 135 int key_size, int value_size, int max_entries, in bpf_create_map_name() argument 145 map_attr.max_entries = max_entries; in bpf_create_map_name() 151 int key_size, int inner_map_fd, int max_entries, in bpf_create_map_in_map_node() argument 163 attr.max_entries = max_entries; in bpf_create_map_in_map_node() 176 int key_size, int inner_map_fd, int max_entries, in bpf_create_map_in_map() argument [all …]
|
/Linux-v4.19/arch/sh/kernel/ |
D | stacktrace.c | 41 if (trace->nr_entries < trace->max_entries) in save_stack_address() 55 if (trace->nr_entries < trace->max_entries) in save_stack_trace() 76 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 90 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk()
|
/Linux-v4.19/arch/x86/kernel/ |
D | stacktrace.c | 26 if (trace->nr_entries >= trace->max_entries) in save_stack_address() 50 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 136 if (trace->nr_entries < trace->max_entries) in __save_stack_trace_reliable() 197 if (trace->nr_entries < trace->max_entries) in __save_stack_trace_user() 200 while (trace->nr_entries < trace->max_entries) { in __save_stack_trace_user() 227 if (trace->nr_entries < trace->max_entries) in save_stack_trace_user()
|
/Linux-v4.19/arch/s390/kernel/ |
D | stacktrace.c | 25 if (trace->nr_entries < trace->max_entries) { in __save_address() 48 if (trace->nr_entries < trace->max_entries) in save_stack_trace() 61 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk() 72 if (trace->nr_entries < trace->max_entries) in save_stack_trace_regs()
|
/Linux-v4.19/arch/arm/kernel/ |
D | stacktrace.c | 87 if (trace->nr_entries >= trace->max_entries) in save_trace() 97 return trace->nr_entries >= trace->max_entries; in save_trace() 118 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 137 if (trace->nr_entries < trace->max_entries) in __save_stack_trace() 156 if (trace->nr_entries < trace->max_entries) in save_stack_trace_regs()
|
/Linux-v4.19/arch/parisc/kernel/ |
D | stacktrace.c | 23 while (trace->nr_entries < trace->max_entries) { in dump_trace() 39 if (trace->nr_entries < trace->max_entries) in save_stack_trace() 47 if (trace->nr_entries < trace->max_entries) in save_stack_trace_tsk()
|