Lines Matching refs:bq
429 struct xdp_bulk_queue *bq; in __cpu_map_entry_alloc() local
445 bq = per_cpu_ptr(rcpu->bulkq, i); in __cpu_map_entry_alloc()
446 bq->obj = rcpu; in __cpu_map_entry_alloc()
691 static void bq_flush_to_queue(struct xdp_bulk_queue *bq) in bq_flush_to_queue() argument
693 struct bpf_cpu_map_entry *rcpu = bq->obj; in bq_flush_to_queue()
699 if (unlikely(!bq->count)) in bq_flush_to_queue()
705 for (i = 0; i < bq->count; i++) { in bq_flush_to_queue()
706 struct xdp_frame *xdpf = bq->q[i]; in bq_flush_to_queue()
716 bq->count = 0; in bq_flush_to_queue()
719 __list_del_clearprev(&bq->flush_node); in bq_flush_to_queue()
731 struct xdp_bulk_queue *bq = this_cpu_ptr(rcpu->bulkq); in bq_enqueue() local
733 if (unlikely(bq->count == CPU_MAP_BULK_SIZE)) in bq_enqueue()
734 bq_flush_to_queue(bq); in bq_enqueue()
745 bq->q[bq->count++] = xdpf; in bq_enqueue()
747 if (!bq->flush_node.prev) in bq_enqueue()
748 list_add(&bq->flush_node, flush_list); in bq_enqueue()
783 struct xdp_bulk_queue *bq, *tmp; in __cpu_map_flush() local
785 list_for_each_entry_safe(bq, tmp, flush_list, flush_node) { in __cpu_map_flush()
786 bq_flush_to_queue(bq); in __cpu_map_flush()
789 wake_up_process(bq->obj->kthread); in __cpu_map_flush()