Lines Matching refs:tx_buffer
269 struct ixgbevf_tx_buffer *tx_buffer; in ixgbevf_clean_tx_irq() local
278 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_clean_tx_irq()
283 union ixgbe_adv_tx_desc *eop_desc = tx_buffer->next_to_watch; in ixgbevf_clean_tx_irq()
297 tx_buffer->next_to_watch = NULL; in ixgbevf_clean_tx_irq()
300 total_bytes += tx_buffer->bytecount; in ixgbevf_clean_tx_irq()
301 total_packets += tx_buffer->gso_segs; in ixgbevf_clean_tx_irq()
305 page_frag_free(tx_buffer->data); in ixgbevf_clean_tx_irq()
307 napi_consume_skb(tx_buffer->skb, napi_budget); in ixgbevf_clean_tx_irq()
311 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq()
312 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq()
316 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_clean_tx_irq()
320 tx_buffer++; in ixgbevf_clean_tx_irq()
325 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_irq()
330 if (dma_unmap_len(tx_buffer, len)) { in ixgbevf_clean_tx_irq()
332 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_irq()
333 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_irq()
335 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_clean_tx_irq()
340 tx_buffer++; in ixgbevf_clean_tx_irq()
345 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_irq()
978 struct ixgbevf_tx_buffer *tx_buffer; in ixgbevf_xmit_xdp_ring() local
995 tx_buffer = &ring->tx_buffer_info[i]; in ixgbevf_xmit_xdp_ring()
997 dma_unmap_len_set(tx_buffer, len, len); in ixgbevf_xmit_xdp_ring()
998 dma_unmap_addr_set(tx_buffer, dma, dma); in ixgbevf_xmit_xdp_ring()
999 tx_buffer->data = xdp->data; in ixgbevf_xmit_xdp_ring()
1000 tx_buffer->bytecount = len; in ixgbevf_xmit_xdp_ring()
1001 tx_buffer->gso_segs = 1; in ixgbevf_xmit_xdp_ring()
1002 tx_buffer->protocol = 0; in ixgbevf_xmit_xdp_ring()
1046 tx_buffer->next_to_watch = tx_desc; in ixgbevf_xmit_xdp_ring()
2365 struct ixgbevf_tx_buffer *tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_clean_tx_ring() local
2372 page_frag_free(tx_buffer->data); in ixgbevf_clean_tx_ring()
2374 dev_kfree_skb_any(tx_buffer->skb); in ixgbevf_clean_tx_ring()
2378 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_ring()
2379 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring()
2383 eop_desc = tx_buffer->next_to_watch; in ixgbevf_clean_tx_ring()
2388 tx_buffer++; in ixgbevf_clean_tx_ring()
2393 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_ring()
2398 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_clean_tx_ring()
2400 dma_unmap_addr(tx_buffer, dma), in ixgbevf_clean_tx_ring()
2401 dma_unmap_len(tx_buffer, len), in ixgbevf_clean_tx_ring()
2406 tx_buffer++; in ixgbevf_clean_tx_ring()
2410 tx_buffer = tx_ring->tx_buffer_info; in ixgbevf_clean_tx_ring()
3912 struct ixgbevf_tx_buffer *tx_buffer; in ixgbevf_tx_map() local
3930 tx_buffer = first; in ixgbevf_tx_map()
3937 dma_unmap_len_set(tx_buffer, len, size); in ixgbevf_tx_map()
3938 dma_unmap_addr_set(tx_buffer, dma, dma); in ixgbevf_tx_map()
3979 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_tx_map()
4013 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_tx_map()
4016 while (tx_buffer != first) { in ixgbevf_tx_map()
4017 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map()
4019 dma_unmap_addr(tx_buffer, dma), in ixgbevf_tx_map()
4020 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map()
4022 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_tx_map()
4026 tx_buffer = &tx_ring->tx_buffer_info[i]; in ixgbevf_tx_map()
4029 if (dma_unmap_len(tx_buffer, len)) in ixgbevf_tx_map()
4031 dma_unmap_addr(tx_buffer, dma), in ixgbevf_tx_map()
4032 dma_unmap_len(tx_buffer, len), in ixgbevf_tx_map()
4034 dma_unmap_len_set(tx_buffer, len, 0); in ixgbevf_tx_map()
4036 dev_kfree_skb_any(tx_buffer->skb); in ixgbevf_tx_map()
4037 tx_buffer->skb = NULL; in ixgbevf_tx_map()