Lines Matching refs:bpf_cpu_map_entry

47 struct bpf_cpu_map_entry {  struct
67 struct bpf_cpu_map_entry **cpu_map; argument
71 static int bq_flush_to_queue(struct bpf_cpu_map_entry *rcpu,
107 cost = (u64) cmap->map.max_entries * sizeof(struct bpf_cpu_map_entry *); in cpu_map_alloc()
128 sizeof(struct bpf_cpu_map_entry *), in cpu_map_alloc()
141 static void get_cpu_map_entry(struct bpf_cpu_map_entry *rcpu) in get_cpu_map_entry()
149 struct bpf_cpu_map_entry *rcpu; in cpu_map_kthread_stop()
151 rcpu = container_of(work, struct bpf_cpu_map_entry, kthread_stop_wq); in cpu_map_kthread_stop()
162 static struct sk_buff *cpu_map_build_skb(struct bpf_cpu_map_entry *rcpu, in cpu_map_build_skb()
225 static void put_cpu_map_entry(struct bpf_cpu_map_entry *rcpu) in put_cpu_map_entry()
238 struct bpf_cpu_map_entry *rcpu = data; in cpu_map_kthread_run()
302 static struct bpf_cpu_map_entry *__cpu_map_entry_alloc(u32 qsize, u32 cpu, in __cpu_map_entry_alloc()
306 struct bpf_cpu_map_entry *rcpu; in __cpu_map_entry_alloc()
363 struct bpf_cpu_map_entry *rcpu; in __cpu_map_entry_free()
371 rcpu = container_of(rcu, struct bpf_cpu_map_entry, rcu); in __cpu_map_entry_free()
405 u32 key_cpu, struct bpf_cpu_map_entry *rcpu) in __cpu_map_entry_replace()
407 struct bpf_cpu_map_entry *old_rcpu; in __cpu_map_entry_replace()
434 struct bpf_cpu_map_entry *rcpu; in cpu_map_update_elem()
502 struct bpf_cpu_map_entry *rcpu; in cpu_map_free()
516 struct bpf_cpu_map_entry *__cpu_map_lookup_elem(struct bpf_map *map, u32 key) in __cpu_map_lookup_elem()
519 struct bpf_cpu_map_entry *rcpu; in __cpu_map_lookup_elem()
530 struct bpf_cpu_map_entry *rcpu = in cpu_map_lookup_elem()
563 static int bq_flush_to_queue(struct bpf_cpu_map_entry *rcpu, in bq_flush_to_queue()
602 static int bq_enqueue(struct bpf_cpu_map_entry *rcpu, struct xdp_frame *xdpf) in bq_enqueue()
622 int cpu_map_enqueue(struct bpf_cpu_map_entry *rcpu, struct xdp_buff *xdp, in cpu_map_enqueue()
657 struct bpf_cpu_map_entry *rcpu = READ_ONCE(cmap->cpu_map[bit]); in __cpu_map_flush()