Home
last modified time | relevance | path

Searched refs:nb_entries (Results 1 – 4 of 4) sorted by relevance

/Linux-v6.1/net/xdp/
Dxsk_queue.h216 u32 cached_cons = q->cached_cons, nb_entries = 0; in xskq_cons_read_desc_batch() local
219 while (cached_cons != q->cached_prod && nb_entries < max) { in xskq_cons_read_desc_batch()
223 descs[nb_entries] = ring->desc[idx]; in xskq_cons_read_desc_batch()
224 if (unlikely(!xskq_cons_is_valid_desc(q, &descs[nb_entries], pool))) { in xskq_cons_read_desc_batch()
230 nb_entries++; in xskq_cons_read_desc_batch()
236 return nb_entries; in xskq_cons_read_desc_batch()
356 u32 nb_entries) in xskq_prod_write_addr_batch() argument
363 for (i = 0; i < nb_entries; i++) in xskq_prod_write_addr_batch()
405 static inline void xskq_prod_submit_n(struct xsk_queue *q, u32 nb_entries) in xskq_prod_submit_n() argument
407 __xskq_prod_submit(q, q->ring->producer + nb_entries); in xskq_prod_submit_n()
Dxsk_buff_pool.c542 u32 i, cached_cons, nb_entries; in xp_alloc_new_from_fq() local
549 nb_entries = max; in xp_alloc_new_from_fq()
562 nb_entries--; in xp_alloc_new_from_fq()
580 return nb_entries; in xp_alloc_new_from_fq()
583 static u32 xp_alloc_reused(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 nb_entries) in xp_alloc_reused() argument
588 nb_entries = min_t(u32, nb_entries, pool->free_list_cnt); in xp_alloc_reused()
590 i = nb_entries; in xp_alloc_reused()
598 pool->free_list_cnt -= nb_entries; in xp_alloc_reused()
600 return nb_entries; in xp_alloc_reused()
Dxsk.c296 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument
298 xskq_prod_submit_n(pool->cq, nb_entries); in xsk_tx_completed()
/Linux-v6.1/include/net/
Dxdp_sock_drv.h17 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries);
147 static inline void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument