Lines Matching refs:txr

328 static void bnxt_txr_db_kick(struct bnxt *bp, struct bnxt_tx_ring_info *txr,  in bnxt_txr_db_kick()  argument
331 bnxt_db_write(bp, &txr->tx_db, prod); in bnxt_txr_db_kick()
332 txr->kick_pending = 0; in bnxt_txr_db_kick()
336 struct bnxt_tx_ring_info *txr, in bnxt_txr_netif_try_stop_queue() argument
347 if (bnxt_tx_avail(bp, txr) >= bp->tx_wake_thresh) { in bnxt_txr_netif_try_stop_queue()
367 struct bnxt_tx_ring_info *txr; in bnxt_start_xmit() local
379 txr = &bp->tx_ring[bp->tx_ring_map[i]]; in bnxt_start_xmit()
380 prod = txr->tx_prod; in bnxt_start_xmit()
382 free_size = bnxt_tx_avail(bp, txr); in bnxt_start_xmit()
385 if (net_ratelimit() && txr->kick_pending) in bnxt_start_xmit()
388 if (bnxt_txr_netif_try_stop_queue(bp, txr, txq)) in bnxt_start_xmit()
396 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
400 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
438 struct tx_push_buffer *tx_push_buf = txr->tx_push; in bnxt_start_xmit()
441 void __iomem *db = txr->tx_db.doorbell; in bnxt_start_xmit()
484 txbd->tx_bd_haddr = txr->data_mapping; in bnxt_start_xmit()
486 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
491 txr->tx_prod = prod; in bnxt_start_xmit()
531 &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
571 txbd = &txr->tx_desc_ring[TX_RING(prod)][TX_IDX(prod)]; in bnxt_start_xmit()
580 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
602 txr->tx_prod = prod; in bnxt_start_xmit()
605 bnxt_txr_db_kick(bp, txr, prod); in bnxt_start_xmit()
607 txr->kick_pending = 1; in bnxt_start_xmit()
611 if (unlikely(bnxt_tx_avail(bp, txr) <= MAX_SKB_FRAGS + 1)) { in bnxt_start_xmit()
613 bnxt_txr_db_kick(bp, txr, prod); in bnxt_start_xmit()
615 bnxt_txr_netif_try_stop_queue(bp, txr, txq); in bnxt_start_xmit()
626 prod = txr->tx_prod; in bnxt_start_xmit()
627 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
635 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit()
644 if (txr->kick_pending) in bnxt_start_xmit()
645 bnxt_txr_db_kick(bp, txr, txr->tx_prod); in bnxt_start_xmit()
646 txr->tx_buf_ring[txr->tx_prod].skb = NULL; in bnxt_start_xmit()
653 struct bnxt_tx_ring_info *txr = bnapi->tx_ring; in bnxt_tx_int() local
654 struct netdev_queue *txq = netdev_get_tx_queue(bp->dev, txr->txq_index); in bnxt_tx_int()
655 u16 cons = txr->tx_cons; in bnxt_tx_int()
665 tx_buf = &txr->tx_buf_ring[cons]; in bnxt_tx_int()
683 tx_buf = &txr->tx_buf_ring[cons]; in bnxt_tx_int()
707 txr->tx_cons = cons; in bnxt_tx_int()
717 bnxt_tx_avail(bp, txr) >= bp->tx_wake_thresh && in bnxt_tx_int()
718 READ_ONCE(txr->dev_state) != BNXT_DEV_STATE_CLOSING) in bnxt_tx_int()
2584 struct bnxt_tx_ring_info *txr = bnapi->tx_ring; in __bnxt_poll_work() local
2585 u16 prod = txr->tx_prod; in __bnxt_poll_work()
2590 bnxt_db_write_relaxed(bp, &txr->tx_db, prod); in __bnxt_poll_work()
2869 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_free_tx_skbs() local
2872 if (!txr->tx_buf_ring) in bnxt_free_tx_skbs()
2876 struct bnxt_sw_tx_bd *tx_buf = &txr->tx_buf_ring[j]; in bnxt_free_tx_skbs()
2918 tx_buf = &txr->tx_buf_ring[ring_idx]; in bnxt_free_tx_skbs()
3319 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_free_tx_rings() local
3322 if (txr->tx_push) { in bnxt_free_tx_rings()
3324 txr->tx_push, txr->tx_push_mapping); in bnxt_free_tx_rings()
3325 txr->tx_push = NULL; in bnxt_free_tx_rings()
3328 ring = &txr->tx_ring_struct; in bnxt_free_tx_rings()
3355 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_alloc_tx_rings() local
3359 ring = &txr->tx_ring_struct; in bnxt_alloc_tx_rings()
3365 ring->grp_idx = txr->bnapi->index; in bnxt_alloc_tx_rings()
3372 txr->tx_push = dma_alloc_coherent(&pdev->dev, in bnxt_alloc_tx_rings()
3374 &txr->tx_push_mapping, in bnxt_alloc_tx_rings()
3377 if (!txr->tx_push) in bnxt_alloc_tx_rings()
3380 mapping = txr->tx_push_mapping + in bnxt_alloc_tx_rings()
3382 txr->data_mapping = cpu_to_le64(mapping); in bnxt_alloc_tx_rings()
3386 spin_lock_init(&txr->xdp_tx_lock); in bnxt_alloc_tx_rings()
3583 struct bnxt_tx_ring_info *txr; in bnxt_init_ring_struct() local
3621 txr = bnapi->tx_ring; in bnxt_init_ring_struct()
3622 if (!txr) in bnxt_init_ring_struct()
3625 ring = &txr->tx_ring_struct; in bnxt_init_ring_struct()
3629 rmem->pg_arr = (void **)txr->tx_desc_ring; in bnxt_init_ring_struct()
3630 rmem->dma_arr = txr->tx_desc_mapping; in bnxt_init_ring_struct()
3632 rmem->vmem = (void **)&txr->tx_buf_ring; in bnxt_init_ring_struct()
3796 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_init_tx_rings() local
3797 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; in bnxt_init_tx_rings()
4436 struct bnxt_tx_ring_info *txr; in bnxt_clear_ring_indices() local
4444 txr = bnapi->tx_ring; in bnxt_clear_ring_indices()
4445 if (txr) { in bnxt_clear_ring_indices()
4446 txr->tx_prod = 0; in bnxt_clear_ring_indices()
4447 txr->tx_cons = 0; in bnxt_clear_ring_indices()
4606 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_alloc_mem() local
4609 txr->tx_ring_struct.ring_mem.flags = in bnxt_alloc_mem()
4611 txr->bnapi = bp->bnapi[j]; in bnxt_alloc_mem()
4612 bp->bnapi[j]->tx_ring = txr; in bnxt_alloc_mem()
4615 txr->txq_index = i - bp->tx_nr_rings_xdp; in bnxt_alloc_mem()
5179 static u16 bnxt_cp_ring_for_tx(struct bnxt *bp, struct bnxt_tx_ring_info *txr) in bnxt_cp_ring_for_tx() argument
5182 struct bnxt_napi *bnapi = txr->bnapi; in bnxt_cp_ring_for_tx()
5188 return bnxt_cp_ring_from_grp(bp, &txr->tx_ring_struct); in bnxt_cp_ring_for_tx()
5718 struct bnxt_tx_ring_info *txr; in hwrm_ring_alloc_send_msg() local
5720 txr = container_of(ring, struct bnxt_tx_ring_info, in hwrm_ring_alloc_send_msg()
5725 req->cmpl_ring_id = cpu_to_le16(bnxt_cp_ring_for_tx(bp, txr)); in hwrm_ring_alloc_send_msg()
5914 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_hwrm_ring_alloc() local
5919 struct bnxt_napi *bnapi = txr->bnapi; in bnxt_hwrm_ring_alloc()
5935 ring = &txr->tx_ring_struct; in bnxt_hwrm_ring_alloc()
5940 bnxt_set_db(bp, &txr->tx_db, type, map_idx, ring->fw_ring_id); in bnxt_hwrm_ring_alloc()
6041 struct bnxt_tx_ring_info *txr = &bp->tx_ring[i]; in bnxt_hwrm_ring_free() local
6042 struct bnxt_ring_struct *ring = &txr->tx_ring_struct; in bnxt_hwrm_ring_free()
6045 u32 cmpl_ring_id = bnxt_cp_ring_for_tx(bp, txr); in bnxt_hwrm_ring_free()
9424 struct bnxt_tx_ring_info *txr; in bnxt_tx_disable() local
9428 txr = &bp->tx_ring[i]; in bnxt_tx_disable()
9429 WRITE_ONCE(txr->dev_state, BNXT_DEV_STATE_CLOSING); in bnxt_tx_disable()
9443 struct bnxt_tx_ring_info *txr; in bnxt_tx_enable() local
9446 txr = &bp->tx_ring[i]; in bnxt_tx_enable()
9447 WRITE_ONCE(txr->dev_state, 0); in bnxt_tx_enable()
11468 struct bnxt_tx_ring_info *txr = bnapi->tx_ring; in bnxt_dump_tx_sw_state() local
11471 if (!txr) in bnxt_dump_tx_sw_state()
11475 i, txr->tx_ring_struct.fw_ring_id, txr->tx_prod, in bnxt_dump_tx_sw_state()
11476 txr->tx_cons); in bnxt_dump_tx_sw_state()