Searched refs:nb_entries (Results 1 – 4 of 4) sorted by relevance
/Linux-v6.1/net/xdp/ |
D | xsk_queue.h | 216 u32 cached_cons = q->cached_cons, nb_entries = 0; in xskq_cons_read_desc_batch() local 219 while (cached_cons != q->cached_prod && nb_entries < max) { in xskq_cons_read_desc_batch() 223 descs[nb_entries] = ring->desc[idx]; in xskq_cons_read_desc_batch() 224 if (unlikely(!xskq_cons_is_valid_desc(q, &descs[nb_entries], pool))) { in xskq_cons_read_desc_batch() 230 nb_entries++; in xskq_cons_read_desc_batch() 236 return nb_entries; in xskq_cons_read_desc_batch() 356 u32 nb_entries) in xskq_prod_write_addr_batch() argument 363 for (i = 0; i < nb_entries; i++) in xskq_prod_write_addr_batch() 405 static inline void xskq_prod_submit_n(struct xsk_queue *q, u32 nb_entries) in xskq_prod_submit_n() argument 407 __xskq_prod_submit(q, q->ring->producer + nb_entries); in xskq_prod_submit_n()
|
D | xsk_buff_pool.c | 542 u32 i, cached_cons, nb_entries; in xp_alloc_new_from_fq() local 549 nb_entries = max; in xp_alloc_new_from_fq() 562 nb_entries--; in xp_alloc_new_from_fq() 580 return nb_entries; in xp_alloc_new_from_fq() 583 static u32 xp_alloc_reused(struct xsk_buff_pool *pool, struct xdp_buff **xdp, u32 nb_entries) in xp_alloc_reused() argument 588 nb_entries = min_t(u32, nb_entries, pool->free_list_cnt); in xp_alloc_reused() 590 i = nb_entries; in xp_alloc_reused() 598 pool->free_list_cnt -= nb_entries; in xp_alloc_reused() 600 return nb_entries; in xp_alloc_reused()
|
D | xsk.c | 296 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument 298 xskq_prod_submit_n(pool->cq, nb_entries); in xsk_tx_completed()
|
/Linux-v6.1/include/net/ |
D | xdp_sock_drv.h | 17 void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries); 147 static inline void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) in xsk_tx_completed() argument
|