Lines Matching refs:tx_cb

1921 	struct ql_tx_buf_cb *tx_cb;  in ql_process_mac_tx_intr()  local
1929 tx_cb = &qdev->tx_buf[mac_rsp->transaction_id]; in ql_process_mac_tx_intr()
1940 if (tx_cb->seg_count == 0) { in ql_process_mac_tx_intr()
1949 dma_unmap_addr(&tx_cb->map[0], mapaddr), in ql_process_mac_tx_intr()
1950 dma_unmap_len(&tx_cb->map[0], maplen), in ql_process_mac_tx_intr()
1952 tx_cb->seg_count--; in ql_process_mac_tx_intr()
1953 if (tx_cb->seg_count) { in ql_process_mac_tx_intr()
1954 for (i = 1; i < tx_cb->seg_count; i++) { in ql_process_mac_tx_intr()
1956 dma_unmap_addr(&tx_cb->map[i], in ql_process_mac_tx_intr()
1958 dma_unmap_len(&tx_cb->map[i], maplen), in ql_process_mac_tx_intr()
1963 qdev->ndev->stats.tx_bytes += tx_cb->skb->len; in ql_process_mac_tx_intr()
1966 dev_kfree_skb_irq(tx_cb->skb); in ql_process_mac_tx_intr()
1967 tx_cb->skb = NULL; in ql_process_mac_tx_intr()
2308 struct ql_tx_buf_cb *tx_cb, in ql_send_map() argument
2320 seg_cnt = tx_cb->seg_count; in ql_send_map()
2338 dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map); in ql_send_map()
2339 dma_unmap_len_set(&tx_cb->map[seg], maplen, len); in ql_send_map()
2347 oal = tx_cb->oal; in ql_send_map()
2378 dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map); in ql_send_map()
2379 dma_unmap_len_set(&tx_cb->map[seg], maplen, in ql_send_map()
2400 dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map); in ql_send_map()
2401 dma_unmap_len_set(&tx_cb->map[seg], maplen, skb_frag_size(frag)); in ql_send_map()
2415 oal = tx_cb->oal; in ql_send_map()
2429 dma_unmap_addr(&tx_cb->map[seg], mapaddr), in ql_send_map()
2430 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map()
2437 dma_unmap_addr(&tx_cb->map[seg], mapaddr), in ql_send_map()
2438 dma_unmap_len(&tx_cb->map[seg], maplen), in ql_send_map()
2443 dma_unmap_addr(&tx_cb->map[0], mapaddr), in ql_send_map()
2444 dma_unmap_addr(&tx_cb->map[0], maplen), in ql_send_map()
2468 struct ql_tx_buf_cb *tx_cb; in ql3xxx_send() local
2475 tx_cb = &qdev->tx_buf[qdev->req_producer_index]; in ql3xxx_send()
2476 tx_cb->seg_count = ql_get_seg_count(qdev, in ql3xxx_send()
2478 if (tx_cb->seg_count == -1) { in ql3xxx_send()
2483 mac_iocb_ptr = tx_cb->queue_entry; in ql3xxx_send()
2490 tx_cb->skb = skb; in ql3xxx_send()
2495 if (ql_send_map(qdev, mac_iocb_ptr, tx_cb, skb) != NETDEV_TX_OK) { in ql3xxx_send()
2810 struct ql_tx_buf_cb *tx_cb; in ql_free_send_free_list() local
2813 tx_cb = &qdev->tx_buf[0]; in ql_free_send_free_list()
2815 kfree(tx_cb->oal); in ql_free_send_free_list()
2816 tx_cb->oal = NULL; in ql_free_send_free_list()
2817 tx_cb++; in ql_free_send_free_list()
2823 struct ql_tx_buf_cb *tx_cb; in ql_create_send_free_list() local
2830 tx_cb = &qdev->tx_buf[i]; in ql_create_send_free_list()
2831 tx_cb->skb = NULL; in ql_create_send_free_list()
2832 tx_cb->queue_entry = req_q_curr; in ql_create_send_free_list()
2834 tx_cb->oal = kmalloc(512, GFP_KERNEL); in ql_create_send_free_list()
2835 if (tx_cb->oal == NULL) in ql_create_send_free_list()
3627 struct ql_tx_buf_cb *tx_cb; in ql_reset_work() local
3641 tx_cb = &qdev->tx_buf[i]; in ql_reset_work()
3642 if (tx_cb->skb) { in ql_reset_work()
3646 dma_unmap_addr(&tx_cb->map[0], in ql_reset_work()
3648 dma_unmap_len(&tx_cb->map[0], maplen), in ql_reset_work()
3650 for (j = 1; j < tx_cb->seg_count; j++) { in ql_reset_work()
3652 dma_unmap_addr(&tx_cb->map[j], in ql_reset_work()
3654 dma_unmap_len(&tx_cb->map[j], in ql_reset_work()
3658 dev_kfree_skb(tx_cb->skb); in ql_reset_work()
3659 tx_cb->skb = NULL; in ql_reset_work()