Lines Matching refs:cur_tx
367 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail()
368 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail()
370 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail()
418 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_enable_eee_mode()
2439 unsigned int entry = tx_q->cur_tx; in stmmac_xdp_xmit_zc()
2514 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_xdp_xmit_zc()
2515 entry = tx_q->cur_tx; in stmmac_xdp_xmit_zc()
2572 while ((entry != tx_q->cur_tx) && count < priv->dma_conf.dma_tx_size) { in stmmac_tx_clean()
2716 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_tx_clean()
4008 p = &tx_q->dma_entx[tx_q->cur_tx].basic; in stmmac_vlan_insert()
4010 p = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_vlan_insert()
4016 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_vlan_insert()
4044 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, in stmmac_tso_allocator()
4046 WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]); in stmmac_tso_allocator()
4049 desc = &tx_q->dma_entx[tx_q->cur_tx].basic; in stmmac_tso_allocator()
4051 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_allocator()
4089 tx_q->tx_tail_addr = tx_q->dma_tx_phy + (tx_q->cur_tx * desc_size); in stmmac_flush_tx_descriptors()
4139 first_tx = tx_q->cur_tx; in stmmac_tso_xmit()
4171 mss_desc = &tx_q->dma_entx[tx_q->cur_tx].basic; in stmmac_tso_xmit()
4173 mss_desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_xmit()
4177 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, in stmmac_tso_xmit()
4179 WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]); in stmmac_tso_xmit()
4192 first_entry = tx_q->cur_tx; in stmmac_tso_xmit()
4246 tx_q->tx_skbuff_dma[tx_q->cur_tx].buf = des; in stmmac_tso_xmit()
4247 tx_q->tx_skbuff_dma[tx_q->cur_tx].len = skb_frag_size(frag); in stmmac_tso_xmit()
4248 tx_q->tx_skbuff_dma[tx_q->cur_tx].map_as_page = true; in stmmac_tso_xmit()
4249 tx_q->tx_skbuff_dma[tx_q->cur_tx].buf_type = STMMAC_TXBUF_T_SKB; in stmmac_tso_xmit()
4252 tx_q->tx_skbuff_dma[tx_q->cur_tx].last_segment = true; in stmmac_tso_xmit()
4255 tx_q->tx_skbuff[tx_q->cur_tx] = skb; in stmmac_tso_xmit()
4256 tx_q->tx_skbuff_dma[tx_q->cur_tx].buf_type = STMMAC_TXBUF_T_SKB; in stmmac_tso_xmit()
4259 tx_packets = (tx_q->cur_tx + 1) - first_tx; in stmmac_tso_xmit()
4276 desc = &tx_q->dma_entx[tx_q->cur_tx].basic; in stmmac_tso_xmit()
4278 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_xmit()
4289 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_tso_xmit()
4337 __func__, tx_q->cur_tx, tx_q->dirty_tx, first_entry, in stmmac_tso_xmit()
4338 tx_q->cur_tx, first, nfrags); in stmmac_tso_xmit()
4385 first_tx = tx_q->cur_tx; in stmmac_xmit()
4413 entry = tx_q->cur_tx; in stmmac_xmit()
4518 tx_q->cur_tx = entry; in stmmac_xmit()
4523 __func__, tx_q->cur_tx, tx_q->dirty_tx, first_entry, in stmmac_xmit()
4749 unsigned int entry = tx_q->cur_tx; in stmmac_xdp_xmit_xdpf()
4815 tx_q->cur_tx = entry; in stmmac_xdp_xmit_xdpf()
7745 tx_q->cur_tx = 0; in stmmac_reset_tx_queue()