Lines Matching refs:tx_info
316 struct sk_buff *skb, u32 tx_info) in mt76_dma_tx_queue_skb_raw() argument
334 mt76_dma_add_buf(dev, q, &buf, 1, tx_info, skb, NULL); in mt76_dma_tx_queue_skb_raw()
351 struct mt76_tx_info tx_info = { in mt76_dma_tx_queue_skb() local
379 tx_info.buf[n].addr = t->dma_addr; in mt76_dma_tx_queue_skb()
380 tx_info.buf[n++].len = dev->drv->txwi_size; in mt76_dma_tx_queue_skb()
381 tx_info.buf[n].addr = addr; in mt76_dma_tx_queue_skb()
382 tx_info.buf[n++].len = len; in mt76_dma_tx_queue_skb()
385 if (n == ARRAY_SIZE(tx_info.buf)) in mt76_dma_tx_queue_skb()
393 tx_info.buf[n].addr = addr; in mt76_dma_tx_queue_skb()
394 tx_info.buf[n++].len = iter->len; in mt76_dma_tx_queue_skb()
396 tx_info.nbuf = n; in mt76_dma_tx_queue_skb()
400 ret = dev->drv->tx_prepare_skb(dev, txwi, qid, wcid, sta, &tx_info); in mt76_dma_tx_queue_skb()
406 if (q->queued + (tx_info.nbuf + 1) / 2 >= q->ndesc - 1) { in mt76_dma_tx_queue_skb()
411 return mt76_dma_add_buf(dev, q, tx_info.buf, tx_info.nbuf, in mt76_dma_tx_queue_skb()
412 tx_info.info, tx_info.skb, t); in mt76_dma_tx_queue_skb()
416 dma_unmap_single(dev->dev, tx_info.buf[n].addr, in mt76_dma_tx_queue_skb()
417 tx_info.buf[n].len, DMA_TO_DEVICE); in mt76_dma_tx_queue_skb()
422 if (tx_info.skb == dev->test.tx_skb) in mt76_dma_tx_queue_skb()
426 e.skb = tx_info.skb; in mt76_dma_tx_queue_skb()