Lines Matching refs:tx_skb

828 	struct nv_skb_map *tx_skb;  member
1038 kfree(np->tx_skb); in free_rings()
1962 np->get_tx_ctx = np->tx_skb; in nv_init_tx()
1963 np->put_tx_ctx = np->tx_skb; in nv_init_tx()
1964 np->last_tx_ctx = &np->tx_skb[np->tx_ring_size-1]; in nv_init_tx()
1981 np->tx_skb[i].skb = NULL; in nv_init_tx()
1982 np->tx_skb[i].dma = 0; in nv_init_tx()
1983 np->tx_skb[i].dma_len = 0; in nv_init_tx()
1984 np->tx_skb[i].dma_single = 0; in nv_init_tx()
1985 np->tx_skb[i].first_tx_desc = NULL; in nv_init_tx()
1986 np->tx_skb[i].next_tx_ctx = NULL; in nv_init_tx()
2003 static void nv_unmap_txskb(struct fe_priv *np, struct nv_skb_map *tx_skb) in nv_unmap_txskb() argument
2005 if (tx_skb->dma) { in nv_unmap_txskb()
2006 if (tx_skb->dma_single) in nv_unmap_txskb()
2007 dma_unmap_single(&np->pci_dev->dev, tx_skb->dma, in nv_unmap_txskb()
2008 tx_skb->dma_len, in nv_unmap_txskb()
2011 dma_unmap_page(&np->pci_dev->dev, tx_skb->dma, in nv_unmap_txskb()
2012 tx_skb->dma_len, in nv_unmap_txskb()
2014 tx_skb->dma = 0; in nv_unmap_txskb()
2018 static int nv_release_txskb(struct fe_priv *np, struct nv_skb_map *tx_skb) in nv_release_txskb() argument
2020 nv_unmap_txskb(np, tx_skb); in nv_release_txskb()
2021 if (tx_skb->skb) { in nv_release_txskb()
2022 dev_kfree_skb_any(tx_skb->skb); in nv_release_txskb()
2023 tx_skb->skb = NULL; in nv_release_txskb()
2044 if (nv_release_txskb(np, &np->tx_skb[i])) { in nv_drain_tx()
2049 np->tx_skb[i].dma = 0; in nv_drain_tx()
2050 np->tx_skb[i].dma_len = 0; in nv_drain_tx()
2051 np->tx_skb[i].dma_single = 0; in nv_drain_tx()
2052 np->tx_skb[i].first_tx_desc = NULL; in nv_drain_tx()
2053 np->tx_skb[i].next_tx_ctx = NULL; in nv_drain_tx()
2275 np->put_tx_ctx = np->tx_skb; in nv_start_xmit()
2301 tmp_tx_ctx = np->tx_skb; in nv_start_xmit()
2321 np->put_tx_ctx = np->tx_skb; in nv_start_xmit()
2330 if (unlikely(np->put_tx_ctx == np->tx_skb)) in nv_start_xmit()
2433 np->put_tx_ctx = np->tx_skb; in nv_start_xmit_optimized()
2459 tmp_tx_ctx = np->tx_skb; in nv_start_xmit_optimized()
2479 np->put_tx_ctx = np->tx_skb; in nv_start_xmit_optimized()
2488 if (unlikely(np->put_tx_ctx == np->tx_skb)) in nv_start_xmit_optimized()
2630 np->get_tx_ctx = np->tx_skb; in nv_tx_done()
2687 np->get_tx_ctx = np->tx_skb; in nv_tx_done_optimized()
4719 np->tx_skb = (struct nv_skb_map *)tx_skbuff; in nv_set_ringparam()
4723 memset(np->tx_skb, 0, sizeof(struct nv_skb_map) * np->tx_ring_size); in nv_set_ringparam()
5101 struct sk_buff *tx_skb, *rx_skb; in nv_loopback_test() local
5139 tx_skb = netdev_alloc_skb(dev, pkt_len); in nv_loopback_test()
5140 if (!tx_skb) { in nv_loopback_test()
5144 test_dma_addr = dma_map_single(&np->pci_dev->dev, tx_skb->data, in nv_loopback_test()
5145 skb_tailroom(tx_skb), in nv_loopback_test()
5149 dev_kfree_skb_any(tx_skb); in nv_loopback_test()
5152 pkt_data = skb_put(tx_skb, pkt_len); in nv_loopback_test()
5204 (skb_end_pointer(tx_skb) - tx_skb->data), in nv_loopback_test()
5206 dev_kfree_skb_any(tx_skb); in nv_loopback_test()
5833 np->tx_skb = kcalloc(np->tx_ring_size, sizeof(struct nv_skb_map), GFP_KERNEL); in nv_probe()
5834 if (!np->rx_skb || !np->tx_skb) in nv_probe()