Lines Matching refs:xdp_ring

2208 			      struct i40e_ring *xdp_ring);
2211 struct i40e_ring *xdp_ring) in i40e_xmit_xdp_tx_ring() argument
2218 return i40e_xmit_xdp_ring(xdpf, xdp_ring); in i40e_xmit_xdp_tx_ring()
2230 struct i40e_ring *xdp_ring; in i40e_run_xdp() local
2247 xdp_ring = rx_ring->vsi->xdp_rings[rx_ring->queue_index]; in i40e_run_xdp()
2248 result = i40e_xmit_xdp_tx_ring(xdp, xdp_ring); in i40e_run_xdp()
2290 static inline void i40e_xdp_ring_update_tail(struct i40e_ring *xdp_ring) in i40e_xdp_ring_update_tail() argument
2296 writel_relaxed(xdp_ring->next_to_use, xdp_ring->tail); in i40e_xdp_ring_update_tail()
2439 struct i40e_ring *xdp_ring = in i40e_clean_rx_irq() local
2442 i40e_xdp_ring_update_tail(xdp_ring); in i40e_clean_rx_irq()
3502 struct i40e_ring *xdp_ring) in i40e_xmit_xdp_ring() argument
3504 u16 i = xdp_ring->next_to_use; in i40e_xmit_xdp_ring()
3510 if (!unlikely(I40E_DESC_UNUSED(xdp_ring))) { in i40e_xmit_xdp_ring()
3511 xdp_ring->tx_stats.tx_busy++; in i40e_xmit_xdp_ring()
3515 dma = dma_map_single(xdp_ring->dev, xdpf->data, size, DMA_TO_DEVICE); in i40e_xmit_xdp_ring()
3516 if (dma_mapping_error(xdp_ring->dev, dma)) in i40e_xmit_xdp_ring()
3519 tx_bi = &xdp_ring->tx_bi[i]; in i40e_xmit_xdp_ring()
3528 tx_desc = I40E_TX_DESC(xdp_ring, i); in i40e_xmit_xdp_ring()
3540 if (i == xdp_ring->count) in i40e_xmit_xdp_ring()
3544 xdp_ring->next_to_use = i; in i40e_xmit_xdp_ring()
3708 struct i40e_ring *xdp_ring; in i40e_xdp_xmit() local
3721 xdp_ring = vsi->xdp_rings[queue_index]; in i40e_xdp_xmit()
3727 err = i40e_xmit_xdp_ring(xdpf, xdp_ring); in i40e_xdp_xmit()
3735 i40e_xdp_ring_update_tail(xdp_ring); in i40e_xdp_xmit()