Lines Matching refs:dtab

104 static inline struct hlist_head *dev_map_index_hash(struct bpf_dtab *dtab,  in dev_map_index_hash()  argument
107 return &dtab->dev_index_head[idx & (dtab->n_buckets - 1)]; in dev_map_index_hash()
110 static int dev_map_init_map(struct bpf_dtab *dtab, union bpf_attr *attr) in dev_map_init_map() argument
130 bpf_map_init_from_attr(&dtab->map, attr); in dev_map_init_map()
133 dtab->n_buckets = roundup_pow_of_two(dtab->map.max_entries); in dev_map_init_map()
135 if (!dtab->n_buckets) /* Overflow check */ in dev_map_init_map()
140 dtab->dev_index_head = dev_map_create_hash(dtab->n_buckets, in dev_map_init_map()
141 dtab->map.numa_node); in dev_map_init_map()
142 if (!dtab->dev_index_head) in dev_map_init_map()
145 spin_lock_init(&dtab->index_lock); in dev_map_init_map()
147 dtab->netdev_map = bpf_map_area_alloc((u64) dtab->map.max_entries * in dev_map_init_map()
149 dtab->map.numa_node); in dev_map_init_map()
150 if (!dtab->netdev_map) in dev_map_init_map()
159 struct bpf_dtab *dtab; in dev_map_alloc() local
162 dtab = bpf_map_area_alloc(sizeof(*dtab), NUMA_NO_NODE); in dev_map_alloc()
163 if (!dtab) in dev_map_alloc()
166 err = dev_map_init_map(dtab, attr); in dev_map_alloc()
168 bpf_map_area_free(dtab); in dev_map_alloc()
173 list_add_tail_rcu(&dtab->list, &dev_map_list); in dev_map_alloc()
176 return &dtab->map; in dev_map_alloc()
181 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_free() local
195 list_del_rcu(&dtab->list); in dev_map_free()
204 if (dtab->map.map_type == BPF_MAP_TYPE_DEVMAP_HASH) { in dev_map_free()
205 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_free()
210 head = dev_map_index_hash(dtab, i); in dev_map_free()
221 bpf_map_area_free(dtab->dev_index_head); in dev_map_free()
223 for (i = 0; i < dtab->map.max_entries; i++) { in dev_map_free()
226 dev = rcu_dereference_raw(dtab->netdev_map[i]); in dev_map_free()
236 bpf_map_area_free(dtab->netdev_map); in dev_map_free()
239 bpf_map_area_free(dtab); in dev_map_free()
244 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_get_next_key() local
248 if (index >= dtab->map.max_entries) { in dev_map_get_next_key()
253 if (index == dtab->map.max_entries - 1) in dev_map_get_next_key()
265 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_hash_lookup_elem() local
266 struct hlist_head *head = dev_map_index_hash(dtab, key); in __dev_map_hash_lookup_elem()
270 lockdep_is_held(&dtab->index_lock)) in __dev_map_hash_lookup_elem()
280 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_hash_get_next_key() local
303 i = idx & (dtab->n_buckets - 1); in dev_map_hash_get_next_key()
307 for (; i < dtab->n_buckets; i++) { in dev_map_hash_get_next_key()
308 head = dev_map_index_hash(dtab, i); in dev_map_hash_get_next_key()
427 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_lookup_elem() local
433 obj = rcu_dereference_check(dtab->netdev_map[key], in __dev_map_lookup_elem()
594 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_enqueue_multi() local
609 dst = rcu_dereference_check(dtab->netdev_map[i], in dev_map_enqueue_multi()
630 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_enqueue_multi()
631 head = dev_map_index_hash(dtab, i); in dev_map_enqueue_multi()
633 lockdep_is_held(&dtab->index_lock)) { in dev_map_enqueue_multi()
711 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_redirect_multi() local
727 dst = rcu_dereference_check(dtab->netdev_map[i], in dev_map_redirect_multi()
749 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_redirect_multi()
750 head = dev_map_index_hash(dtab, i); in dev_map_redirect_multi()
810 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_delete_elem() local
817 old_dev = unrcu_pointer(xchg(&dtab->netdev_map[k], NULL)); in dev_map_delete_elem()
820 atomic_dec((atomic_t *)&dtab->items); in dev_map_delete_elem()
827 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_hash_delete_elem() local
833 spin_lock_irqsave(&dtab->index_lock, flags); in dev_map_hash_delete_elem()
837 dtab->items--; in dev_map_hash_delete_elem()
842 spin_unlock_irqrestore(&dtab->index_lock, flags); in dev_map_hash_delete_elem()
848 struct bpf_dtab *dtab, in __dev_map_alloc_node() argument
855 dev = bpf_map_kmalloc_node(&dtab->map, sizeof(*dev), in __dev_map_alloc_node()
857 dtab->map.numa_node); in __dev_map_alloc_node()
871 !bpf_prog_map_compatible(&dtab->map, prog)) in __dev_map_alloc_node()
898 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_update_elem() local
905 if (unlikely(i >= dtab->map.max_entries)) in __dev_map_update_elem()
919 dev = __dev_map_alloc_node(net, dtab, &val, i); in __dev_map_update_elem()
928 old_dev = unrcu_pointer(xchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev))); in __dev_map_update_elem()
932 atomic_inc((atomic_t *)&dtab->items); in __dev_map_update_elem()
947 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in __dev_map_hash_update_elem() local
960 spin_lock_irqsave(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
966 dev = __dev_map_alloc_node(net, dtab, &val, idx); in __dev_map_hash_update_elem()
975 if (dtab->items >= dtab->map.max_entries) { in __dev_map_hash_update_elem()
976 spin_unlock_irqrestore(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
980 dtab->items++; in __dev_map_hash_update_elem()
984 dev_map_index_hash(dtab, idx)); in __dev_map_hash_update_elem()
985 spin_unlock_irqrestore(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
993 spin_unlock_irqrestore(&dtab->index_lock, flags); in __dev_map_hash_update_elem()
1020 struct bpf_dtab *dtab = container_of(map, struct bpf_dtab, map); in dev_map_mem_usage() local
1024 usage += (u64)dtab->n_buckets * sizeof(struct hlist_head); in dev_map_mem_usage()
1027 usage += atomic_read((atomic_t *)&dtab->items) * in dev_map_mem_usage()
1061 static void dev_map_hash_remove_netdev(struct bpf_dtab *dtab, in dev_map_hash_remove_netdev() argument
1067 spin_lock_irqsave(&dtab->index_lock, flags); in dev_map_hash_remove_netdev()
1068 for (i = 0; i < dtab->n_buckets; i++) { in dev_map_hash_remove_netdev()
1073 head = dev_map_index_hash(dtab, i); in dev_map_hash_remove_netdev()
1079 dtab->items--; in dev_map_hash_remove_netdev()
1084 spin_unlock_irqrestore(&dtab->index_lock, flags); in dev_map_hash_remove_netdev()
1091 struct bpf_dtab *dtab; in dev_map_notification() local
1114 list_for_each_entry_rcu(dtab, &dev_map_list, list) { in dev_map_notification()
1115 if (dtab->map.map_type == BPF_MAP_TYPE_DEVMAP_HASH) { in dev_map_notification()
1116 dev_map_hash_remove_netdev(dtab, netdev); in dev_map_notification()
1120 for (i = 0; i < dtab->map.max_entries; i++) { in dev_map_notification()
1123 dev = rcu_dereference(dtab->netdev_map[i]); in dev_map_notification()
1126 odev = unrcu_pointer(cmpxchg(&dtab->netdev_map[i], RCU_INITIALIZER(dev), NULL)); in dev_map_notification()
1130 atomic_dec((atomic_t *)&dtab->items); in dev_map_notification()