Lines Matching refs:tx_buff

571 	memset(tx_pool->tx_buff, 0,  in reset_one_tx_pool()
616 kfree(tx_pool->tx_buff); in release_one_tx_pool()
647 tx_pool->tx_buff = kcalloc(num_entries, in init_one_tx_pool()
650 if (!tx_pool->tx_buff) in init_one_tx_pool()
1142 struct ibmvnic_tx_buff *tx_buff; in clean_one_tx_pool() local
1146 if (!tx_pool || !tx_pool->tx_buff) in clean_one_tx_pool()
1152 tx_buff = &tx_pool->tx_buff[i]; in clean_one_tx_pool()
1153 if (tx_buff && tx_buff->skb) { in clean_one_tx_pool()
1154 dev_kfree_skb_any(tx_buff->skb); in clean_one_tx_pool()
1155 tx_buff->skb = NULL; in clean_one_tx_pool()
1411 struct ibmvnic_tx_buff *tx_buff = NULL; in ibmvnic_xmit() local
1498 tx_buff = &tx_pool->tx_buff[index]; in ibmvnic_xmit()
1499 tx_buff->skb = skb; in ibmvnic_xmit()
1500 tx_buff->data_dma[0] = data_dma_addr; in ibmvnic_xmit()
1501 tx_buff->data_len[0] = skb->len; in ibmvnic_xmit()
1502 tx_buff->index = index; in ibmvnic_xmit()
1503 tx_buff->pool_index = queue_num; in ibmvnic_xmit()
1504 tx_buff->last_frag = true; in ibmvnic_xmit()
1551 build_hdr_descs_arr(tx_buff, &num_entries, *hdrs); in ibmvnic_xmit()
1553 tx_buff->num_entries = num_entries; in ibmvnic_xmit()
1554 tx_buff->indir_arr[0] = tx_crq; in ibmvnic_xmit()
1555 tx_buff->indir_dma = dma_map_single(dev, tx_buff->indir_arr, in ibmvnic_xmit()
1556 sizeof(tx_buff->indir_arr), in ibmvnic_xmit()
1558 if (dma_mapping_error(dev, tx_buff->indir_dma)) { in ibmvnic_xmit()
1560 tx_buff->skb = NULL; in ibmvnic_xmit()
1569 (u64)tx_buff->indir_dma, in ibmvnic_xmit()
1571 dma_unmap_single(dev, tx_buff->indir_dma, in ibmvnic_xmit()
1572 sizeof(tx_buff->indir_arr), DMA_TO_DEVICE); in ibmvnic_xmit()
1574 tx_buff->num_entries = num_entries; in ibmvnic_xmit()
1582 tx_buff->skb = NULL; in ibmvnic_xmit()
2930 txbuff = &tx_pool->tx_buff[index]; in ibmvnic_complete_tx()