Lines Matching refs:tx_buf

36 	struct ice_tx_buf *tx_buf, *first;  in ice_prgm_fdir_fltr()  local
68 first = &tx_ring->tx_buf[i]; in ice_prgm_fdir_fltr()
75 tx_buf = &tx_ring->tx_buf[i]; in ice_prgm_fdir_fltr()
80 memset(tx_buf, 0, sizeof(*tx_buf)); in ice_prgm_fdir_fltr()
81 dma_unmap_len_set(tx_buf, len, ICE_FDIR_MAX_RAW_PKT_SIZE); in ice_prgm_fdir_fltr()
82 dma_unmap_addr_set(tx_buf, dma, dma); in ice_prgm_fdir_fltr()
88 tx_buf->tx_flags = ICE_TX_FLAGS_DUMMY_PKT; in ice_prgm_fdir_fltr()
89 tx_buf->raw_buf = raw_packet; in ice_prgm_fdir_fltr()
113 ice_unmap_and_free_tx_buf(struct ice_tx_ring *ring, struct ice_tx_buf *tx_buf) in ice_unmap_and_free_tx_buf() argument
115 if (tx_buf->skb) { in ice_unmap_and_free_tx_buf()
116 if (tx_buf->tx_flags & ICE_TX_FLAGS_DUMMY_PKT) in ice_unmap_and_free_tx_buf()
117 devm_kfree(ring->dev, tx_buf->raw_buf); in ice_unmap_and_free_tx_buf()
119 page_frag_free(tx_buf->raw_buf); in ice_unmap_and_free_tx_buf()
121 dev_kfree_skb_any(tx_buf->skb); in ice_unmap_and_free_tx_buf()
122 if (dma_unmap_len(tx_buf, len)) in ice_unmap_and_free_tx_buf()
124 dma_unmap_addr(tx_buf, dma), in ice_unmap_and_free_tx_buf()
125 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf()
127 } else if (dma_unmap_len(tx_buf, len)) { in ice_unmap_and_free_tx_buf()
129 dma_unmap_addr(tx_buf, dma), in ice_unmap_and_free_tx_buf()
130 dma_unmap_len(tx_buf, len), in ice_unmap_and_free_tx_buf()
134 tx_buf->next_to_watch = NULL; in ice_unmap_and_free_tx_buf()
135 tx_buf->skb = NULL; in ice_unmap_and_free_tx_buf()
136 dma_unmap_len_set(tx_buf, len, 0); in ice_unmap_and_free_tx_buf()
160 if (!tx_ring->tx_buf) in ice_clean_tx_ring()
165 ice_unmap_and_free_tx_buf(tx_ring, &tx_ring->tx_buf[i]); in ice_clean_tx_ring()
168 memset(tx_ring->tx_buf, 0, sizeof(*tx_ring->tx_buf) * tx_ring->count); in ice_clean_tx_ring()
198 devm_kfree(tx_ring->dev, tx_ring->tx_buf); in ice_free_tx_ring()
199 tx_ring->tx_buf = NULL; in ice_free_tx_ring()
224 struct ice_tx_buf *tx_buf; in ice_clean_tx_irq() local
229 tx_buf = &tx_ring->tx_buf[i]; in ice_clean_tx_irq()
236 struct ice_tx_desc *eop_desc = tx_buf->next_to_watch; in ice_clean_tx_irq()
243 prefetchw(&tx_buf->skb->users); in ice_clean_tx_irq()
247 ice_trace(clean_tx_irq, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq()
254 tx_buf->next_to_watch = NULL; in ice_clean_tx_irq()
257 total_bytes += tx_buf->bytecount; in ice_clean_tx_irq()
258 total_pkts += tx_buf->gso_segs; in ice_clean_tx_irq()
261 napi_consume_skb(tx_buf->skb, napi_budget); in ice_clean_tx_irq()
265 dma_unmap_addr(tx_buf, dma), in ice_clean_tx_irq()
266 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq()
270 tx_buf->skb = NULL; in ice_clean_tx_irq()
271 dma_unmap_len_set(tx_buf, len, 0); in ice_clean_tx_irq()
275 ice_trace(clean_tx_irq_unmap, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq()
276 tx_buf++; in ice_clean_tx_irq()
281 tx_buf = tx_ring->tx_buf; in ice_clean_tx_irq()
286 if (dma_unmap_len(tx_buf, len)) { in ice_clean_tx_irq()
288 dma_unmap_addr(tx_buf, dma), in ice_clean_tx_irq()
289 dma_unmap_len(tx_buf, len), in ice_clean_tx_irq()
291 dma_unmap_len_set(tx_buf, len, 0); in ice_clean_tx_irq()
294 ice_trace(clean_tx_irq_unmap_eop, tx_ring, tx_desc, tx_buf); in ice_clean_tx_irq()
297 tx_buf++; in ice_clean_tx_irq()
302 tx_buf = tx_ring->tx_buf; in ice_clean_tx_irq()
350 WARN_ON(tx_ring->tx_buf); in ice_setup_tx_ring()
351 tx_ring->tx_buf = in ice_setup_tx_ring()
352 devm_kcalloc(dev, sizeof(*tx_ring->tx_buf), tx_ring->count, in ice_setup_tx_ring()
354 if (!tx_ring->tx_buf) in ice_setup_tx_ring()
374 devm_kfree(dev, tx_ring->tx_buf); in ice_setup_tx_ring()
375 tx_ring->tx_buf = NULL; in ice_setup_tx_ring()
1589 struct ice_tx_buf *tx_buf; in ice_tx_map() local
1613 tx_buf = first; in ice_tx_map()
1622 dma_unmap_len_set(tx_buf, len, size); in ice_tx_map()
1623 dma_unmap_addr_set(tx_buf, dma, dma); in ice_tx_map()
1672 tx_buf = &tx_ring->tx_buf[i]; in ice_tx_map()
1714 tx_buf = &tx_ring->tx_buf[i]; in ice_tx_map()
1715 ice_unmap_and_free_tx_buf(tx_ring, tx_buf); in ice_tx_map()
1716 if (tx_buf == first) in ice_tx_map()
2317 first = &tx_ring->tx_buf[tx_ring->next_to_use]; in ice_xmit_frame_ring()
2449 struct ice_tx_buf *tx_buf; in ice_clean_ctrl_tx_irq() local
2451 tx_buf = &tx_ring->tx_buf[i]; in ice_clean_ctrl_tx_irq()
2456 struct ice_tx_desc *eop_desc = tx_buf->next_to_watch; in ice_clean_ctrl_tx_irq()
2471 tx_buf->next_to_watch = NULL; in ice_clean_ctrl_tx_irq()
2476 tx_buf++; in ice_clean_ctrl_tx_irq()
2481 tx_buf = tx_ring->tx_buf; in ice_clean_ctrl_tx_irq()
2486 if (dma_unmap_len(tx_buf, len)) in ice_clean_ctrl_tx_irq()
2488 dma_unmap_addr(tx_buf, dma), in ice_clean_ctrl_tx_irq()
2489 dma_unmap_len(tx_buf, len), in ice_clean_ctrl_tx_irq()
2491 if (tx_buf->tx_flags & ICE_TX_FLAGS_DUMMY_PKT) in ice_clean_ctrl_tx_irq()
2492 devm_kfree(tx_ring->dev, tx_buf->raw_buf); in ice_clean_ctrl_tx_irq()
2495 tx_buf->raw_buf = NULL; in ice_clean_ctrl_tx_irq()
2496 tx_buf->tx_flags = 0; in ice_clean_ctrl_tx_irq()
2497 tx_buf->next_to_watch = NULL; in ice_clean_ctrl_tx_irq()
2498 dma_unmap_len_set(tx_buf, len, 0); in ice_clean_ctrl_tx_irq()
2503 tx_buf++; in ice_clean_ctrl_tx_irq()
2508 tx_buf = tx_ring->tx_buf; in ice_clean_ctrl_tx_irq()