Lines Matching full:map
25 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_free_percpu()
36 for (i = 0; i < array->map.max_entries; i++) { in bpf_array_alloc_percpu()
37 ptr = bpf_map_alloc_percpu(&array->map, array->elem_size, 8, in bpf_array_alloc_percpu()
128 /* allocate all map elements and zero-initialize them */ in array_map_alloc()
144 array->map.bypass_spec_v1 = bypass_spec_v1; in array_map_alloc()
146 /* copy mandatory map attributes */ in array_map_alloc()
147 bpf_map_init_from_attr(&array->map, attr); in array_map_alloc()
155 return &array->map; in array_map_alloc()
159 static void *array_map_lookup_elem(struct bpf_map *map, void *key) in array_map_lookup_elem() argument
161 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_lookup_elem()
164 if (unlikely(index >= array->map.max_entries)) in array_map_lookup_elem()
170 static int array_map_direct_value_addr(const struct bpf_map *map, u64 *imm, in array_map_direct_value_addr() argument
173 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_direct_value_addr()
175 if (map->max_entries != 1) in array_map_direct_value_addr()
177 if (off >= map->value_size) in array_map_direct_value_addr()
184 static int array_map_direct_value_meta(const struct bpf_map *map, u64 imm, in array_map_direct_value_meta() argument
187 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_direct_value_meta()
191 if (map->max_entries != 1) in array_map_direct_value_meta()
201 static int array_map_gen_lookup(struct bpf_map *map, struct bpf_insn *insn_buf) in array_map_gen_lookup() argument
203 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_gen_lookup()
205 u32 elem_size = round_up(map->value_size, 8); in array_map_gen_lookup()
210 if (map->map_flags & BPF_F_INNER_MAP) in array_map_gen_lookup()
215 if (!map->bypass_spec_v1) { in array_map_gen_lookup()
216 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 4); in array_map_gen_lookup()
219 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 3); in array_map_gen_lookup()
234 static void *percpu_array_map_lookup_elem(struct bpf_map *map, void *key) in percpu_array_map_lookup_elem() argument
236 struct bpf_array *array = container_of(map, struct bpf_array, map); in percpu_array_map_lookup_elem()
239 if (unlikely(index >= array->map.max_entries)) in percpu_array_map_lookup_elem()
245 int bpf_percpu_array_copy(struct bpf_map *map, void *key, void *value) in bpf_percpu_array_copy() argument
247 struct bpf_array *array = container_of(map, struct bpf_array, map); in bpf_percpu_array_copy()
253 if (unlikely(index >= array->map.max_entries)) in bpf_percpu_array_copy()
260 size = round_up(map->value_size, 8); in bpf_percpu_array_copy()
272 static int array_map_get_next_key(struct bpf_map *map, void *key, void *next_key) in array_map_get_next_key() argument
274 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_get_next_key()
278 if (index >= array->map.max_entries) { in array_map_get_next_key()
283 if (index == array->map.max_entries - 1) in array_map_get_next_key()
292 if (unlikely(map_value_has_timer(&arr->map))) in check_and_free_timer_in_array()
293 bpf_timer_cancel_and_free(val + arr->map.timer_off); in check_and_free_timer_in_array()
297 static int array_map_update_elem(struct bpf_map *map, void *key, void *value, in array_map_update_elem() argument
300 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_update_elem()
308 if (unlikely(index >= array->map.max_entries)) in array_map_update_elem()
317 !map_value_has_spin_lock(map))) in array_map_update_elem()
320 if (array->map.map_type == BPF_MAP_TYPE_PERCPU_ARRAY) { in array_map_update_elem()
322 value, map->value_size); in array_map_update_elem()
327 copy_map_value_locked(map, val, value, false); in array_map_update_elem()
329 copy_map_value(map, val, value); in array_map_update_elem()
335 int bpf_percpu_array_update(struct bpf_map *map, void *key, void *value, in bpf_percpu_array_update() argument
338 struct bpf_array *array = container_of(map, struct bpf_array, map); in bpf_percpu_array_update()
348 if (unlikely(index >= array->map.max_entries)) in bpf_percpu_array_update()
362 size = round_up(map->value_size, 8); in bpf_percpu_array_update()
374 static int array_map_delete_elem(struct bpf_map *map, void *key) in array_map_delete_elem() argument
384 static void array_map_free_timers(struct bpf_map *map) in array_map_free_timers() argument
386 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_free_timers()
389 if (likely(!map_value_has_timer(map))) in array_map_free_timers()
392 for (i = 0; i < array->map.max_entries; i++) in array_map_free_timers()
394 map->timer_off); in array_map_free_timers()
397 /* Called when map->refcnt goes to zero, either from workqueue or from syscall */
398 static void array_map_free(struct bpf_map *map) in array_map_free() argument
400 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_free()
402 if (array->map.map_type == BPF_MAP_TYPE_PERCPU_ARRAY) in array_map_free()
405 if (array->map.map_flags & BPF_F_MMAPABLE) in array_map_free()
411 static void array_map_seq_show_elem(struct bpf_map *map, void *key, in array_map_seq_show_elem() argument
418 value = array_map_lookup_elem(map, key); in array_map_seq_show_elem()
424 if (map->btf_key_type_id) in array_map_seq_show_elem()
426 btf_type_seq_show(map->btf, map->btf_value_type_id, value, m); in array_map_seq_show_elem()
432 static void percpu_array_map_seq_show_elem(struct bpf_map *map, void *key, in percpu_array_map_seq_show_elem() argument
435 struct bpf_array *array = container_of(map, struct bpf_array, map); in percpu_array_map_seq_show_elem()
446 btf_type_seq_show(map->btf, map->btf_value_type_id, in percpu_array_map_seq_show_elem()
455 static int array_map_check_btf(const struct bpf_map *map, in array_map_check_btf() argument
462 /* One exception for keyless BTF: .bss/.data/.rodata map */ in array_map_check_btf()
464 if (map->map_type != BPF_MAP_TYPE_ARRAY || in array_map_check_btf()
465 map->max_entries != 1) in array_map_check_btf()
487 static int array_map_mmap(struct bpf_map *map, struct vm_area_struct *vma) in array_map_mmap() argument
489 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_map_mmap()
492 if (!(map->map_flags & BPF_F_MMAPABLE)) in array_map_mmap()
496 PAGE_ALIGN((u64)array->map.max_entries * array->elem_size)) in array_map_mmap()
513 struct bpf_map *map; member
521 struct bpf_map *map = info->map; in bpf_array_map_seq_start() local
525 if (info->index >= map->max_entries) in bpf_array_map_seq_start()
530 array = container_of(map, struct bpf_array, map); in bpf_array_map_seq_start()
540 struct bpf_map *map = info->map; in bpf_array_map_seq_next() local
546 if (info->index >= map->max_entries) in bpf_array_map_seq_next()
549 array = container_of(map, struct bpf_array, map); in bpf_array_map_seq_next()
560 struct bpf_map *map = info->map; in __bpf_array_map_seq_show() local
573 ctx.map = info->map; in __bpf_array_map_seq_show()
581 size = round_up(map->value_size, 8); in __bpf_array_map_seq_show()
610 struct bpf_map *map = aux->map; in bpf_iter_init_array_map() local
614 if (map->map_type == BPF_MAP_TYPE_PERCPU_ARRAY) { in bpf_iter_init_array_map()
615 buf_size = round_up(map->value_size, 8) * num_possible_cpus(); in bpf_iter_init_array_map()
623 seq_info->map = map; in bpf_iter_init_array_map()
648 static int bpf_for_each_array_elem(struct bpf_map *map, void *callback_fn, in bpf_for_each_array_elem() argument
660 is_percpu = map->map_type == BPF_MAP_TYPE_PERCPU_ARRAY; in bpf_for_each_array_elem()
661 array = container_of(map, struct bpf_array, map); in bpf_for_each_array_elem()
664 for (i = 0; i < map->max_entries; i++) { in bpf_for_each_array_elem()
671 ret = BPF_CAST_CALL(callback_fn)((u64)(long)map, in bpf_for_each_array_elem()
733 /* only file descriptors can be stored in this type of map */ in fd_array_map_alloc_check()
742 static void fd_array_map_free(struct bpf_map *map) in fd_array_map_free() argument
744 struct bpf_array *array = container_of(map, struct bpf_array, map); in fd_array_map_free()
748 for (i = 0; i < array->map.max_entries; i++) in fd_array_map_free()
754 static void *fd_array_map_lookup_elem(struct bpf_map *map, void *key) in fd_array_map_lookup_elem() argument
760 int bpf_fd_array_map_lookup_elem(struct bpf_map *map, void *key, u32 *value) in bpf_fd_array_map_lookup_elem() argument
765 if (!map->ops->map_fd_sys_lookup_elem) in bpf_fd_array_map_lookup_elem()
769 elem = array_map_lookup_elem(map, key); in bpf_fd_array_map_lookup_elem()
771 *value = map->ops->map_fd_sys_lookup_elem(ptr); in bpf_fd_array_map_lookup_elem()
780 int bpf_fd_array_map_update_elem(struct bpf_map *map, struct file *map_file, in bpf_fd_array_map_update_elem() argument
783 struct bpf_array *array = container_of(map, struct bpf_array, map); in bpf_fd_array_map_update_elem()
790 if (index >= array->map.max_entries) in bpf_fd_array_map_update_elem()
794 new_ptr = map->ops->map_fd_get_ptr(map, map_file, ufd); in bpf_fd_array_map_update_elem()
798 if (map->ops->map_poke_run) { in bpf_fd_array_map_update_elem()
801 map->ops->map_poke_run(map, index, old_ptr, new_ptr); in bpf_fd_array_map_update_elem()
808 map->ops->map_fd_put_ptr(old_ptr); in bpf_fd_array_map_update_elem()
812 static int fd_array_map_delete_elem(struct bpf_map *map, void *key) in fd_array_map_delete_elem() argument
814 struct bpf_array *array = container_of(map, struct bpf_array, map); in fd_array_map_delete_elem()
818 if (index >= array->map.max_entries) in fd_array_map_delete_elem()
821 if (map->ops->map_poke_run) { in fd_array_map_delete_elem()
824 map->ops->map_poke_run(map, index, old_ptr, NULL); in fd_array_map_delete_elem()
831 map->ops->map_fd_put_ptr(old_ptr); in fd_array_map_delete_elem()
838 static void *prog_fd_array_get_ptr(struct bpf_map *map, in prog_fd_array_get_ptr() argument
841 struct bpf_array *array = container_of(map, struct bpf_array, map); in prog_fd_array_get_ptr()
865 /* decrement refcnt of all bpf_progs that are stored in this map */
866 static void bpf_fd_array_map_clear(struct bpf_map *map) in bpf_fd_array_map_clear() argument
868 struct bpf_array *array = container_of(map, struct bpf_array, map); in bpf_fd_array_map_clear()
871 for (i = 0; i < array->map.max_entries; i++) in bpf_fd_array_map_clear()
872 fd_array_map_delete_elem(map, &i); in bpf_fd_array_map_clear()
875 static void prog_array_map_seq_show_elem(struct bpf_map *map, void *key, in prog_array_map_seq_show_elem() argument
883 elem = array_map_lookup_elem(map, key); in prog_array_map_seq_show_elem()
889 btf_type_seq_show(map->btf, map->btf_value_type_id, in prog_array_map_seq_show_elem()
903 static int prog_array_map_poke_track(struct bpf_map *map, in prog_array_map_poke_track() argument
910 aux = container_of(map, struct bpf_array, map)->aux; in prog_array_map_poke_track()
936 static void prog_array_map_poke_untrack(struct bpf_map *map, in prog_array_map_poke_untrack() argument
942 aux = container_of(map, struct bpf_array, map)->aux; in prog_array_map_poke_untrack()
954 static void prog_array_map_poke_run(struct bpf_map *map, u32 key, in prog_array_map_poke_run() argument
962 aux = container_of(map, struct bpf_array, map)->aux; in prog_array_map_poke_run()
1006 if (poke->tail_call.map != map || in prog_array_map_poke_run()
1049 struct bpf_map *map = container_of(work, struct bpf_array_aux, in prog_array_map_clear_deferred() local
1050 work)->map; in prog_array_map_clear_deferred()
1051 bpf_fd_array_map_clear(map); in prog_array_map_clear_deferred()
1052 bpf_map_put(map); in prog_array_map_clear_deferred()
1055 static void prog_array_map_clear(struct bpf_map *map) in prog_array_map_clear() argument
1057 struct bpf_array_aux *aux = container_of(map, struct bpf_array, in prog_array_map_clear()
1058 map)->aux; in prog_array_map_clear()
1059 bpf_map_inc(map); in prog_array_map_clear()
1066 struct bpf_map *map; in prog_array_map_alloc() local
1077 map = array_map_alloc(attr); in prog_array_map_alloc()
1078 if (IS_ERR(map)) { in prog_array_map_alloc()
1080 return map; in prog_array_map_alloc()
1083 container_of(map, struct bpf_array, map)->aux = aux; in prog_array_map_alloc()
1084 aux->map = map; in prog_array_map_alloc()
1086 return map; in prog_array_map_alloc()
1089 static void prog_array_map_free(struct bpf_map *map) in prog_array_map_free() argument
1094 aux = container_of(map, struct bpf_array, map)->aux; in prog_array_map_free()
1100 fd_array_map_free(map); in prog_array_map_free()
1157 static void *perf_event_fd_array_get_ptr(struct bpf_map *map, in perf_event_fd_array_get_ptr() argument
1188 static void perf_event_fd_array_release(struct bpf_map *map, in perf_event_fd_array_release() argument
1191 struct bpf_array *array = container_of(map, struct bpf_array, map); in perf_event_fd_array_release()
1195 if (map->map_flags & BPF_F_PRESERVE_ELEMS) in perf_event_fd_array_release()
1199 for (i = 0; i < array->map.max_entries; i++) { in perf_event_fd_array_release()
1202 fd_array_map_delete_elem(map, &i); in perf_event_fd_array_release()
1207 static void perf_event_fd_array_map_free(struct bpf_map *map) in perf_event_fd_array_map_free() argument
1209 if (map->map_flags & BPF_F_PRESERVE_ELEMS) in perf_event_fd_array_map_free()
1210 bpf_fd_array_map_clear(map); in perf_event_fd_array_map_free()
1211 fd_array_map_free(map); in perf_event_fd_array_map_free()
1232 static void *cgroup_fd_array_get_ptr(struct bpf_map *map, in cgroup_fd_array_get_ptr() argument
1245 static void cgroup_fd_array_free(struct bpf_map *map) in cgroup_fd_array_free() argument
1247 bpf_fd_array_map_clear(map); in cgroup_fd_array_free()
1248 fd_array_map_free(map); in cgroup_fd_array_free()
1270 struct bpf_map *map, *inner_map_meta; in array_of_map_alloc() local
1276 map = array_map_alloc(attr); in array_of_map_alloc()
1277 if (IS_ERR(map)) { in array_of_map_alloc()
1279 return map; in array_of_map_alloc()
1282 map->inner_map_meta = inner_map_meta; in array_of_map_alloc()
1284 return map; in array_of_map_alloc()
1287 static void array_of_map_free(struct bpf_map *map) in array_of_map_free() argument
1289 /* map->inner_map_meta is only accessed by syscall which in array_of_map_free()
1292 bpf_map_meta_free(map->inner_map_meta); in array_of_map_free()
1293 bpf_fd_array_map_clear(map); in array_of_map_free()
1294 fd_array_map_free(map); in array_of_map_free()
1297 static void *array_of_map_lookup_elem(struct bpf_map *map, void *key) in array_of_map_lookup_elem() argument
1299 struct bpf_map **inner_map = array_map_lookup_elem(map, key); in array_of_map_lookup_elem()
1307 static int array_of_map_gen_lookup(struct bpf_map *map, in array_of_map_gen_lookup() argument
1310 struct bpf_array *array = container_of(map, struct bpf_array, map); in array_of_map_gen_lookup()
1311 u32 elem_size = round_up(map->value_size, 8); in array_of_map_gen_lookup()
1319 if (!map->bypass_spec_v1) { in array_of_map_gen_lookup()
1320 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 6); in array_of_map_gen_lookup()
1323 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); in array_of_map_gen_lookup()