Home
last modified time | relevance | path

Searched refs:map_flags (Results 1 – 25 of 92) sorted by relevance

1234

/Linux-v5.15/tools/testing/selftests/bpf/
Dtest_lru_map.c29 static int create_map(int map_type, int map_flags, unsigned int size) in create_map() argument
34 sizeof(unsigned long long), size, map_flags); in create_map()
159 static void test_lru_sanity0(int map_type, int map_flags) in test_lru_sanity0() argument
166 map_flags); in test_lru_sanity0()
170 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_sanity0()
171 lru_map_fd = create_map(map_type, map_flags, 2 * nr_cpus); in test_lru_sanity0()
173 lru_map_fd = create_map(map_type, map_flags, 2); in test_lru_sanity0()
256 static void test_lru_sanity1(int map_type, int map_flags, unsigned int tgt_free) in test_lru_sanity1() argument
264 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_sanity1()
269 map_flags); in test_lru_sanity1()
[all …]
Dtest_maps.c35 static int map_flags; variable
43 2, map_flags); in test_hashmap()
141 2, map_flags); in test_hashmap_sizes()
163 sizeof(bpf_percpu(value, 0)), 2, map_flags); in test_hashmap_percpu()
275 max_entries, map_flags); in helper_fill_hashmap()
278 "err: %s, flags: 0x%x\n", strerror(errno), map_flags); in helper_fill_hashmap()
334 old_flags = map_flags; in test_hashmap_zero_seed()
335 map_flags |= BPF_F_ZERO_SEED; in test_hashmap_zero_seed()
357 map_flags = old_flags; in test_hashmap_zero_seed()
555 map_flags); in test_queuemap()
[all …]
/Linux-v5.15/samples/bpf/
Dtest_lru_dist.c208 static int create_map(int map_type, int map_flags, unsigned int size) in create_map() argument
213 sizeof(unsigned long long), size, map_flags); in create_map()
310 static void test_parallel_lru_dist(int map_type, int map_flags, in test_parallel_lru_dist() argument
317 map_flags); in test_parallel_lru_dist()
319 if (map_flags & BPF_F_NO_COMMON_LRU) in test_parallel_lru_dist()
320 lru_map_fd = create_map(map_type, map_flags, in test_parallel_lru_dist()
323 lru_map_fd = create_map(map_type, map_flags, in test_parallel_lru_dist()
335 static void test_lru_loss0(int map_type, int map_flags) in test_lru_loss0() argument
344 map_flags); in test_lru_loss0()
348 if (map_flags & BPF_F_NO_COMMON_LRU) in test_lru_loss0()
[all …]
Dmap_perf_test_kern.c38 __uint(map_flags, BPF_F_NO_COMMON_LRU);
46 __uint(map_flags, BPF_F_NUMA_NODE);
72 __uint(map_flags, BPF_F_NO_PREALLOC);
80 __uint(map_flags, BPF_F_NO_PREALLOC);
88 __uint(map_flags, BPF_F_NO_PREALLOC);
Dtcp_dumpstats_kern.c17 __u32 map_flags; member
22 .map_flags = BPF_F_NO_PREALLOC,
/Linux-v5.15/kernel/bpf/
Dbpf_local_storage.c248 u64 map_flags) in check_flags() argument
250 if (old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_NOEXIST) in check_flags()
254 if (!old_sdata && (map_flags & ~BPF_F_LOCK) == BPF_EXIST) in check_flags()
330 void *value, u64 map_flags) in bpf_local_storage_update() argument
339 if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST) || in bpf_local_storage_update()
341 unlikely((map_flags & BPF_F_LOCK) && in bpf_local_storage_update()
348 err = check_flags(NULL, map_flags); in bpf_local_storage_update()
366 if ((map_flags & BPF_F_LOCK) && !(map_flags & BPF_NOEXIST)) { in bpf_local_storage_update()
373 err = check_flags(old_sdata, map_flags); in bpf_local_storage_update()
397 err = check_flags(old_sdata, map_flags); in bpf_local_storage_update()
[all …]
Dreuseport_array.c201 u32 map_flags) in reuseport_array_update_check() argument
203 if (osk && map_flags == BPF_NOEXIST) in reuseport_array_update_check()
206 if (!osk && map_flags == BPF_EXIST) in reuseport_array_update_check()
242 void *value, u64 map_flags) in bpf_fd_reuseport_array_update_elem() argument
252 if (map_flags > BPF_EXIST) in bpf_fd_reuseport_array_update_elem()
282 map_flags); in bpf_fd_reuseport_array_update_elem()
298 err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags); in bpf_fd_reuseport_array_update_elem()
Dhashtab.c132 return !(htab->map.map_flags & BPF_F_NO_PREALLOC); in htab_is_prealloc()
338 htab->map.map_flags & BPF_F_NO_COMMON_LRU, in prealloc_init()
410 bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); in htab_map_alloc_check()
411 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); in htab_map_alloc_check()
412 bool zero_seed = (attr->map_flags & BPF_F_ZERO_SEED); in htab_map_alloc_check()
430 if (attr->map_flags & ~HTAB_CREATE_FLAG_MASK || in htab_map_alloc_check()
431 !bpf_map_flags_access_ok(attr->map_flags)) in htab_map_alloc_check()
473 bool percpu_lru = (attr->map_flags & BPF_F_NO_COMMON_LRU); in htab_map_alloc()
474 bool prealloc = !(attr->map_flags & BPF_F_NO_PREALLOC); in htab_map_alloc()
530 if (htab->map.map_flags & BPF_F_ZERO_SEED) in htab_map_alloc()
[all …]
Darraymap.c59 attr->map_flags & ~ARRAY_CREATE_FLAG_MASK || in array_map_alloc_check()
60 !bpf_map_flags_access_ok(attr->map_flags) || in array_map_alloc_check()
65 attr->map_flags & (BPF_F_MMAPABLE | BPF_F_INNER_MAP)) in array_map_alloc_check()
69 attr->map_flags & BPF_F_PRESERVE_ELEMS) in array_map_alloc_check()
120 if (attr->map_flags & BPF_F_MMAPABLE) { in array_map_alloc()
129 if (attr->map_flags & BPF_F_MMAPABLE) { in array_map_alloc()
210 if (map->map_flags & BPF_F_INNER_MAP) in array_map_gen_lookup()
298 u64 map_flags) in array_map_update_elem() argument
304 if (unlikely((map_flags & ~BPF_F_LOCK) > BPF_EXIST)) in array_map_update_elem()
312 if (unlikely(map_flags & BPF_NOEXIST)) in array_map_update_elem()
[all …]
Dmap_in_map.c51 inner_map_meta->map_flags = inner_map->map_flags; in bpf_map_meta_alloc()
86 meta0->map_flags == meta1->map_flags; in bpf_map_meta_equal()
Ddevmap.c121 attr->map_flags & ~DEV_CREATE_FLAG_MASK) in dev_map_init_map()
127 attr->map_flags |= BPF_F_RDONLY_PROG; in dev_map_init_map()
897 void *key, void *value, u64 map_flags) in __dev_map_update_elem() argument
904 if (unlikely(map_flags > BPF_EXIST)) in __dev_map_update_elem()
908 if (unlikely(map_flags == BPF_NOEXIST)) in __dev_map_update_elem()
937 u64 map_flags) in dev_map_update_elem() argument
940 map, key, value, map_flags); in dev_map_update_elem()
944 void *key, void *value, u64 map_flags) in __dev_map_hash_update_elem() argument
956 if (unlikely(map_flags > BPF_EXIST || !val.ifindex)) in __dev_map_hash_update_elem()
962 if (old_dev && (map_flags & BPF_NOEXIST)) in __dev_map_hash_update_elem()
[all …]
Dlocal_storage.c212 void *value, u64 map_flags) in bpf_percpu_cgroup_storage_update() argument
219 if (map_flags != BPF_ANY && map_flags != BPF_EXIST) in bpf_percpu_cgroup_storage_update()
308 if (attr->map_flags & ~LOCAL_STORAGE_CREATE_FLAG_MASK || in cgroup_storage_map_alloc()
309 !bpf_map_flags_access_ok(attr->map_flags)) in cgroup_storage_map_alloc()
/Linux-v5.15/tools/lib/bpf/
Dlibbpf_probes.c204 int key_size, value_size, max_entries, map_flags; in bpf_probe_map_type() local
212 map_flags = 0; in bpf_probe_map_type()
221 map_flags = BPF_F_NO_PREALLOC; in bpf_probe_map_type()
240 map_flags = BPF_F_NO_PREALLOC; in bpf_probe_map_type()
295 attr.map_flags = map_flags; in bpf_probe_map_type()
Dbpf.h40 __u32 map_flags; member
59 int max_entries, __u32 map_flags, int node);
62 int max_entries, __u32 map_flags);
64 int value_size, int max_entries, __u32 map_flags);
68 __u32 map_flags, int node);
72 __u32 map_flags);
Dbpf.c91 attr.map_flags = create_attr->map_flags; in bpf_create_map_xattr()
112 __u32 map_flags, int node) in bpf_create_map_node() argument
118 map_attr.map_flags = map_flags; in bpf_create_map_node()
124 map_attr.map_flags |= BPF_F_NUMA_NODE; in bpf_create_map_node()
131 int value_size, int max_entries, __u32 map_flags) in bpf_create_map() argument
136 map_attr.map_flags = map_flags; in bpf_create_map()
146 __u32 map_flags) in bpf_create_map_name() argument
152 map_attr.map_flags = map_flags; in bpf_create_map_name()
162 __u32 map_flags, int node) in bpf_create_map_in_map_node() argument
174 attr.map_flags = map_flags; in bpf_create_map_in_map_node()
[all …]
/Linux-v5.15/tools/testing/selftests/bpf/progs/
Dsockopt_inherit.c19 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE);
26 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE);
33 __uint(map_flags, BPF_F_NO_PREALLOC);
Dtest_btf_map_in_map.c46 __uint(map_flags, BPF_F_INNER_MAP);
55 __uint(map_flags, BPF_F_INNER_MAP);
68 __uint(map_flags, BPF_F_INNER_MAP);
Dtest_mmap.c12 __uint(map_flags, BPF_F_MMAPABLE | BPF_F_RDONLY_PROG);
19 __uint(map_flags, BPF_F_MMAPABLE);
Dtest_map_in_map.c11 __uint(map_flags, 0);
20 __uint(map_flags, 0);
Dtest_queue_stack_map.h16 __uint(map_flags, 0);
24 __uint(map_flags, 0);
Dtask_ls_recursion.c12 __uint(map_flags, BPF_F_NO_PREALLOC);
19 __uint(map_flags, BPF_F_NO_PREALLOC);
Dnetns_cookie_prog.c11 __uint(map_flags, BPF_F_NO_PREALLOC);
18 __uint(map_flags, BPF_F_NO_PREALLOC);
Dlocal_storage.c28 __uint(map_flags, BPF_F_NO_PREALLOC);
35 __uint(map_flags, BPF_F_NO_PREALLOC | BPF_F_CLONE);
42 __uint(map_flags, BPF_F_NO_PREALLOC);
Dtest_sk_storage_tracing.c17 __uint(map_flags, BPF_F_NO_PREALLOC);
25 __uint(map_flags, BPF_F_NO_PREALLOC);
/Linux-v5.15/net/xdp/
Dxskmap.c70 attr->map_flags & ~(BPF_F_NUMA_NODE | BPF_F_RDONLY | BPF_F_WRONLY)) in xsk_map_alloc()
152 u64 map_flags) in xsk_map_update_elem() argument
162 if (unlikely(map_flags > BPF_EXIST)) in xsk_map_update_elem()
190 } else if (old_xs && map_flags == BPF_NOEXIST) { in xsk_map_update_elem()
193 } else if (!old_xs && map_flags == BPF_EXIST) { in xsk_map_update_elem()

1234