Lines Matching refs:rxr

686 					 struct bnxt_rx_ring_info *rxr,  in __bnxt_alloc_rx_page()  argument
692 page = page_pool_dev_alloc_pages(rxr->page_pool); in __bnxt_alloc_rx_page()
699 page_pool_recycle_direct(rxr->page_pool, page); in __bnxt_alloc_rx_page()
727 int bnxt_alloc_rx_data(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, in bnxt_alloc_rx_data() argument
730 struct rx_bd *rxbd = &rxr->rx_desc_ring[RX_RING(prod)][RX_IDX(prod)]; in bnxt_alloc_rx_data()
731 struct bnxt_sw_rx_bd *rx_buf = &rxr->rx_buf_ring[prod]; in bnxt_alloc_rx_data()
736 __bnxt_alloc_rx_page(bp, &mapping, rxr, gfp); in bnxt_alloc_rx_data()
758 void bnxt_reuse_rx_data(struct bnxt_rx_ring_info *rxr, u16 cons, void *data) in bnxt_reuse_rx_data() argument
760 u16 prod = rxr->rx_prod; in bnxt_reuse_rx_data()
764 prod_rx_buf = &rxr->rx_buf_ring[prod]; in bnxt_reuse_rx_data()
765 cons_rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_reuse_rx_data()
772 prod_bd = &rxr->rx_desc_ring[RX_RING(prod)][RX_IDX(prod)]; in bnxt_reuse_rx_data()
773 cons_bd = &rxr->rx_desc_ring[RX_RING(cons)][RX_IDX(cons)]; in bnxt_reuse_rx_data()
778 static inline u16 bnxt_find_next_agg_idx(struct bnxt_rx_ring_info *rxr, u16 idx) in bnxt_find_next_agg_idx() argument
780 u16 next, max = rxr->rx_agg_bmap_size; in bnxt_find_next_agg_idx()
782 next = find_next_zero_bit(rxr->rx_agg_bmap, max, idx); in bnxt_find_next_agg_idx()
784 next = find_first_zero_bit(rxr->rx_agg_bmap, max); in bnxt_find_next_agg_idx()
789 struct bnxt_rx_ring_info *rxr, in bnxt_alloc_rx_page() argument
793 &rxr->rx_agg_desc_ring[RX_RING(prod)][RX_IDX(prod)]; in bnxt_alloc_rx_page()
798 u16 sw_prod = rxr->rx_sw_agg_prod; in bnxt_alloc_rx_page()
802 page = rxr->rx_page; in bnxt_alloc_rx_page()
807 rxr->rx_page = page; in bnxt_alloc_rx_page()
808 rxr->rx_page_offset = 0; in bnxt_alloc_rx_page()
810 offset = rxr->rx_page_offset; in bnxt_alloc_rx_page()
811 rxr->rx_page_offset += BNXT_RX_PAGE_SIZE; in bnxt_alloc_rx_page()
812 if (rxr->rx_page_offset == PAGE_SIZE) in bnxt_alloc_rx_page()
813 rxr->rx_page = NULL; in bnxt_alloc_rx_page()
830 if (unlikely(test_bit(sw_prod, rxr->rx_agg_bmap))) in bnxt_alloc_rx_page()
831 sw_prod = bnxt_find_next_agg_idx(rxr, sw_prod); in bnxt_alloc_rx_page()
833 __set_bit(sw_prod, rxr->rx_agg_bmap); in bnxt_alloc_rx_page()
834 rx_agg_buf = &rxr->rx_agg_ring[sw_prod]; in bnxt_alloc_rx_page()
835 rxr->rx_sw_agg_prod = NEXT_RX_AGG(sw_prod); in bnxt_alloc_rx_page()
858 struct bnxt_rx_ring_info *rxr, in bnxt_get_tpa_agg_p5() argument
861 struct bnxt_tpa_info *tpa_info = &rxr->rx_tpa[agg_id]; in bnxt_get_tpa_agg_p5()
871 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in bnxt_reuse_rx_agg_bufs() local
872 u16 prod = rxr->rx_agg_prod; in bnxt_reuse_rx_agg_bufs()
873 u16 sw_prod = rxr->rx_sw_agg_prod; in bnxt_reuse_rx_agg_bufs()
888 agg = bnxt_get_tpa_agg_p5(bp, rxr, idx, start + i); in bnxt_reuse_rx_agg_bufs()
892 __clear_bit(cons, rxr->rx_agg_bmap); in bnxt_reuse_rx_agg_bufs()
894 if (unlikely(test_bit(sw_prod, rxr->rx_agg_bmap))) in bnxt_reuse_rx_agg_bufs()
895 sw_prod = bnxt_find_next_agg_idx(rxr, sw_prod); in bnxt_reuse_rx_agg_bufs()
897 __set_bit(sw_prod, rxr->rx_agg_bmap); in bnxt_reuse_rx_agg_bufs()
898 prod_rx_buf = &rxr->rx_agg_ring[sw_prod]; in bnxt_reuse_rx_agg_bufs()
899 cons_rx_buf = &rxr->rx_agg_ring[cons]; in bnxt_reuse_rx_agg_bufs()
911 prod_bd = &rxr->rx_agg_desc_ring[RX_RING(prod)][RX_IDX(prod)]; in bnxt_reuse_rx_agg_bufs()
919 rxr->rx_agg_prod = prod; in bnxt_reuse_rx_agg_bufs()
920 rxr->rx_sw_agg_prod = sw_prod; in bnxt_reuse_rx_agg_bufs()
924 struct bnxt_rx_ring_info *rxr, in bnxt_rx_page_skb() argument
933 u16 prod = rxr->rx_prod; in bnxt_rx_page_skb()
937 err = bnxt_alloc_rx_data(bp, rxr, prod, GFP_ATOMIC); in bnxt_rx_page_skb()
939 bnxt_reuse_rx_data(rxr, cons, data); in bnxt_rx_page_skb()
949 skb = napi_alloc_skb(&rxr->bnapi->napi, payload); in bnxt_rx_page_skb()
970 struct bnxt_rx_ring_info *rxr, u16 cons, in bnxt_rx_skb() argument
975 u16 prod = rxr->rx_prod; in bnxt_rx_skb()
979 err = bnxt_alloc_rx_data(bp, rxr, prod, GFP_ATOMIC); in bnxt_rx_skb()
981 bnxt_reuse_rx_data(rxr, cons, data); in bnxt_rx_skb()
1005 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in bnxt_rx_pages() local
1006 u16 prod = rxr->rx_agg_prod; in bnxt_rx_pages()
1021 agg = bnxt_get_tpa_agg_p5(bp, rxr, idx, i); in bnxt_rx_pages()
1028 cons_rx_buf = &rxr->rx_agg_ring[cons]; in bnxt_rx_pages()
1031 __clear_bit(cons, rxr->rx_agg_bmap); in bnxt_rx_pages()
1041 if (bnxt_alloc_rx_page(bp, rxr, prod, GFP_ATOMIC) != 0) { in bnxt_rx_pages()
1056 rxr->rx_agg_prod = prod; in bnxt_rx_pages()
1071 rxr->rx_agg_prod = prod; in bnxt_rx_pages()
1167 static void bnxt_sched_reset(struct bnxt *bp, struct bnxt_rx_ring_info *rxr) in bnxt_sched_reset() argument
1169 if (!rxr->bnapi->in_reset) { in bnxt_sched_reset()
1170 rxr->bnapi->in_reset = true; in bnxt_sched_reset()
1174 rxr->rx_next_cons = 0xffff; in bnxt_sched_reset()
1177 static u16 bnxt_alloc_agg_idx(struct bnxt_rx_ring_info *rxr, u16 agg_id) in bnxt_alloc_agg_idx() argument
1179 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; in bnxt_alloc_agg_idx()
1190 static void bnxt_free_agg_idx(struct bnxt_rx_ring_info *rxr, u16 idx) in bnxt_free_agg_idx() argument
1192 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; in bnxt_free_agg_idx()
1197 static u16 bnxt_lookup_agg_idx(struct bnxt_rx_ring_info *rxr, u16 agg_id) in bnxt_lookup_agg_idx() argument
1199 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; in bnxt_lookup_agg_idx()
1204 static void bnxt_tpa_start(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, in bnxt_tpa_start() argument
1216 agg_id = bnxt_alloc_agg_idx(rxr, agg_id); in bnxt_tpa_start()
1221 prod = rxr->rx_prod; in bnxt_tpa_start()
1222 cons_rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_tpa_start()
1223 prod_rx_buf = &rxr->rx_buf_ring[prod]; in bnxt_tpa_start()
1224 tpa_info = &rxr->rx_tpa[agg_id]; in bnxt_tpa_start()
1226 if (unlikely(cons != rxr->rx_next_cons || in bnxt_tpa_start()
1229 cons, rxr->rx_next_cons, in bnxt_tpa_start()
1231 bnxt_sched_reset(bp, rxr); in bnxt_tpa_start()
1244 prod_bd = &rxr->rx_desc_ring[RX_RING(prod)][RX_IDX(prod)]; in bnxt_tpa_start()
1277 rxr->rx_prod = NEXT_RX(prod); in bnxt_tpa_start()
1279 rxr->rx_next_cons = NEXT_RX(cons); in bnxt_tpa_start()
1280 cons_rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_tpa_start()
1282 bnxt_reuse_rx_data(rxr, cons, cons_rx_buf->data); in bnxt_tpa_start()
1283 rxr->rx_prod = NEXT_RX(rxr->rx_prod); in bnxt_tpa_start()
1508 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in bnxt_tpa_end() local
1528 agg_id = bnxt_lookup_agg_idx(rxr, agg_id); in bnxt_tpa_end()
1530 tpa_info = &rxr->rx_tpa[agg_id]; in bnxt_tpa_end()
1538 bnxt_free_agg_idx(rxr, agg_id); in bnxt_tpa_end()
1544 tpa_info = &rxr->rx_tpa[agg_id]; in bnxt_tpa_end()
1639 static void bnxt_tpa_agg(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, in bnxt_tpa_agg() argument
1645 agg_id = bnxt_lookup_agg_idx(rxr, agg_id); in bnxt_tpa_agg()
1646 tpa_info = &rxr->rx_tpa[agg_id]; in bnxt_tpa_agg()
1674 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in bnxt_rx_pkt() local
1695 bnxt_tpa_agg(bp, rxr, (struct rx_agg_cmp *)rxcmp); in bnxt_rx_pkt()
1707 prod = rxr->rx_prod; in bnxt_rx_pkt()
1710 bnxt_tpa_start(bp, rxr, (struct rx_tpa_start_cmp *)rxcmp, in bnxt_rx_pkt()
1734 if (unlikely(cons != rxr->rx_next_cons)) { in bnxt_rx_pkt()
1738 cons, rxr->rx_next_cons); in bnxt_rx_pkt()
1739 bnxt_sched_reset(bp, rxr); in bnxt_rx_pkt()
1742 rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_rx_pkt()
1763 bnxt_reuse_rx_data(rxr, cons, data); in bnxt_rx_pkt()
1771 bnxt_sched_reset(bp, rxr); in bnxt_rx_pkt()
1779 if (bnxt_rx_xdp(bp, rxr, cons, data, &data_ptr, &len, event)) { in bnxt_rx_pkt()
1786 bnxt_reuse_rx_data(rxr, cons, data); in bnxt_rx_pkt()
1801 skb = bp->rx_skb_func(bp, rxr, cons, data, data_ptr, dma_addr, in bnxt_rx_pkt()
1861 rxr->rx_prod = NEXT_RX(prod); in bnxt_rx_pkt()
1862 rxr->rx_next_cons = NEXT_RX(cons); in bnxt_rx_pkt()
2239 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in __bnxt_poll_work_done() local
2242 bnxt_db_write(bp, &rxr->rx_agg_db, rxr->rx_agg_prod); in __bnxt_poll_work_done()
2243 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in __bnxt_poll_work_done()
2271 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in bnxt_poll_nitroa0() local
2321 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_poll_nitroa0()
2324 bnxt_db_write(bp, &rxr->rx_agg_db, rxr->rx_agg_prod); in bnxt_poll_nitroa0()
2544 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_free_rx_skbs() local
2548 if (rxr->rx_tpa) { in bnxt_free_rx_skbs()
2551 &rxr->rx_tpa[j]; in bnxt_free_rx_skbs()
2570 struct bnxt_sw_rx_bd *rx_buf = &rxr->rx_buf_ring[j]; in bnxt_free_rx_skbs()
2584 page_pool_recycle_direct(rxr->page_pool, data); in bnxt_free_rx_skbs()
2596 &rxr->rx_agg_ring[j]; in bnxt_free_rx_skbs()
2608 __clear_bit(j, rxr->rx_agg_bmap); in bnxt_free_rx_skbs()
2612 if (rxr->rx_page) { in bnxt_free_rx_skbs()
2613 __free_page(rxr->rx_page); in bnxt_free_rx_skbs()
2614 rxr->rx_page = NULL; in bnxt_free_rx_skbs()
2616 map = rxr->rx_tpa_idx_map; in bnxt_free_rx_skbs()
2712 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_free_tpa_info() local
2714 kfree(rxr->rx_tpa_idx_map); in bnxt_free_tpa_info()
2715 rxr->rx_tpa_idx_map = NULL; in bnxt_free_tpa_info()
2716 if (rxr->rx_tpa) { in bnxt_free_tpa_info()
2717 kfree(rxr->rx_tpa[0].agg_arr); in bnxt_free_tpa_info()
2718 rxr->rx_tpa[0].agg_arr = NULL; in bnxt_free_tpa_info()
2720 kfree(rxr->rx_tpa); in bnxt_free_tpa_info()
2721 rxr->rx_tpa = NULL; in bnxt_free_tpa_info()
2738 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_alloc_tpa_info() local
2741 rxr->rx_tpa = kcalloc(bp->max_tpa, sizeof(struct bnxt_tpa_info), in bnxt_alloc_tpa_info()
2743 if (!rxr->rx_tpa) in bnxt_alloc_tpa_info()
2749 rxr->rx_tpa[0].agg_arr = agg; in bnxt_alloc_tpa_info()
2753 rxr->rx_tpa[j].agg_arr = agg + j * MAX_SKB_FRAGS; in bnxt_alloc_tpa_info()
2754 rxr->rx_tpa_idx_map = kzalloc(sizeof(*rxr->rx_tpa_idx_map), in bnxt_alloc_tpa_info()
2756 if (!rxr->rx_tpa_idx_map) in bnxt_alloc_tpa_info()
2771 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_free_rx_rings() local
2774 if (rxr->xdp_prog) in bnxt_free_rx_rings()
2775 bpf_prog_put(rxr->xdp_prog); in bnxt_free_rx_rings()
2777 if (xdp_rxq_info_is_reg(&rxr->xdp_rxq)) in bnxt_free_rx_rings()
2778 xdp_rxq_info_unreg(&rxr->xdp_rxq); in bnxt_free_rx_rings()
2780 page_pool_destroy(rxr->page_pool); in bnxt_free_rx_rings()
2781 rxr->page_pool = NULL; in bnxt_free_rx_rings()
2783 kfree(rxr->rx_agg_bmap); in bnxt_free_rx_rings()
2784 rxr->rx_agg_bmap = NULL; in bnxt_free_rx_rings()
2786 ring = &rxr->rx_ring_struct; in bnxt_free_rx_rings()
2789 ring = &rxr->rx_agg_ring_struct; in bnxt_free_rx_rings()
2795 struct bnxt_rx_ring_info *rxr) in bnxt_alloc_rx_page_pool() argument
2804 rxr->page_pool = page_pool_create(&pp); in bnxt_alloc_rx_page_pool()
2805 if (IS_ERR(rxr->page_pool)) { in bnxt_alloc_rx_page_pool()
2806 int err = PTR_ERR(rxr->page_pool); in bnxt_alloc_rx_page_pool()
2808 rxr->page_pool = NULL; in bnxt_alloc_rx_page_pool()
2825 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_alloc_rx_rings() local
2828 ring = &rxr->rx_ring_struct; in bnxt_alloc_rx_rings()
2830 rc = bnxt_alloc_rx_page_pool(bp, rxr); in bnxt_alloc_rx_rings()
2834 rc = xdp_rxq_info_reg(&rxr->xdp_rxq, bp->dev, i); in bnxt_alloc_rx_rings()
2838 rc = xdp_rxq_info_reg_mem_model(&rxr->xdp_rxq, in bnxt_alloc_rx_rings()
2840 rxr->page_pool); in bnxt_alloc_rx_rings()
2842 xdp_rxq_info_unreg(&rxr->xdp_rxq); in bnxt_alloc_rx_rings()
2854 ring = &rxr->rx_agg_ring_struct; in bnxt_alloc_rx_rings()
2860 rxr->rx_agg_bmap_size = bp->rx_agg_ring_mask + 1; in bnxt_alloc_rx_rings()
2861 mem_size = rxr->rx_agg_bmap_size / 8; in bnxt_alloc_rx_rings()
2862 rxr->rx_agg_bmap = kzalloc(mem_size, GFP_KERNEL); in bnxt_alloc_rx_rings()
2863 if (!rxr->rx_agg_bmap) in bnxt_alloc_rx_rings()
3079 struct bnxt_rx_ring_info *rxr; in bnxt_init_ring_struct() local
3095 rxr = bnapi->rx_ring; in bnxt_init_ring_struct()
3096 if (!rxr) in bnxt_init_ring_struct()
3099 ring = &rxr->rx_ring_struct; in bnxt_init_ring_struct()
3103 rmem->pg_arr = (void **)rxr->rx_desc_ring; in bnxt_init_ring_struct()
3104 rmem->dma_arr = rxr->rx_desc_mapping; in bnxt_init_ring_struct()
3106 rmem->vmem = (void **)&rxr->rx_buf_ring; in bnxt_init_ring_struct()
3108 ring = &rxr->rx_agg_ring_struct; in bnxt_init_ring_struct()
3112 rmem->pg_arr = (void **)rxr->rx_agg_desc_ring; in bnxt_init_ring_struct()
3113 rmem->dma_arr = rxr->rx_agg_desc_mapping; in bnxt_init_ring_struct()
3115 rmem->vmem = (void **)&rxr->rx_agg_ring; in bnxt_init_ring_struct()
3158 struct bnxt_rx_ring_info *rxr; in bnxt_init_one_rx_ring() local
3169 rxr = &bp->rx_ring[ring_nr]; in bnxt_init_one_rx_ring()
3170 ring = &rxr->rx_ring_struct; in bnxt_init_one_rx_ring()
3174 rxr->xdp_prog = bpf_prog_add(bp->xdp_prog, 1); in bnxt_init_one_rx_ring()
3175 if (IS_ERR(rxr->xdp_prog)) { in bnxt_init_one_rx_ring()
3176 int rc = PTR_ERR(rxr->xdp_prog); in bnxt_init_one_rx_ring()
3178 rxr->xdp_prog = NULL; in bnxt_init_one_rx_ring()
3182 prod = rxr->rx_prod; in bnxt_init_one_rx_ring()
3184 if (bnxt_alloc_rx_data(bp, rxr, prod, GFP_KERNEL) != 0) { in bnxt_init_one_rx_ring()
3191 rxr->rx_prod = prod; in bnxt_init_one_rx_ring()
3194 ring = &rxr->rx_agg_ring_struct; in bnxt_init_one_rx_ring()
3205 prod = rxr->rx_agg_prod; in bnxt_init_one_rx_ring()
3207 if (bnxt_alloc_rx_page(bp, rxr, prod, GFP_KERNEL) != 0) { in bnxt_init_one_rx_ring()
3214 rxr->rx_agg_prod = prod; in bnxt_init_one_rx_ring()
3217 if (rxr->rx_tpa) { in bnxt_init_one_rx_ring()
3227 rxr->rx_tpa[i].data = data; in bnxt_init_one_rx_ring()
3228 rxr->rx_tpa[i].data_ptr = data + bp->rx_offset; in bnxt_init_one_rx_ring()
3229 rxr->rx_tpa[i].mapping = mapping; in bnxt_init_one_rx_ring()
3851 struct bnxt_rx_ring_info *rxr; in bnxt_clear_ring_indices() local
3866 rxr = bnapi->rx_ring; in bnxt_clear_ring_indices()
3867 if (rxr) { in bnxt_clear_ring_indices()
3868 rxr->rx_prod = 0; in bnxt_clear_ring_indices()
3869 rxr->rx_agg_prod = 0; in bnxt_clear_ring_indices()
3870 rxr->rx_sw_agg_prod = 0; in bnxt_clear_ring_indices()
3871 rxr->rx_next_cons = 0; in bnxt_clear_ring_indices()
3990 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_alloc_mem() local
3993 rxr->rx_ring_struct.ring_mem.flags = in bnxt_alloc_mem()
3995 rxr->rx_agg_ring_struct.ring_mem.flags = in bnxt_alloc_mem()
3998 rxr->bnapi = bp->bnapi[i]; in bnxt_alloc_mem()
4796 static u16 bnxt_cp_ring_for_rx(struct bnxt *bp, struct bnxt_rx_ring_info *rxr) in bnxt_cp_ring_for_rx() argument
4799 struct bnxt_napi *bnapi = rxr->bnapi; in bnxt_cp_ring_for_rx()
4805 return bnxt_cp_ring_from_grp(bp, &rxr->rx_ring_struct); in bnxt_cp_ring_for_rx()
4864 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[0]; in bnxt_hwrm_vnic_set_rss_p5() local
4887 ring_id = rxr->rx_ring_struct.fw_ring_id; in bnxt_hwrm_vnic_set_rss_p5()
4889 ring_id = bnxt_cp_ring_for_rx(bp, rxr); in bnxt_hwrm_vnic_set_rss_p5()
4891 rxr++; in bnxt_hwrm_vnic_set_rss_p5()
4895 rxr = &bp->rx_ring[0]; in bnxt_hwrm_vnic_set_rss_p5()
4989 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[0]; in bnxt_hwrm_vnic_cfg() local
4992 cpu_to_le16(rxr->rx_ring_struct.fw_ring_id); in bnxt_hwrm_vnic_cfg()
4994 cpu_to_le16(bnxt_cp_ring_for_rx(bp, rxr)); in bnxt_hwrm_vnic_cfg()
5457 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_hwrm_ring_alloc() local
5458 struct bnxt_ring_struct *ring = &rxr->rx_ring_struct; in bnxt_hwrm_ring_alloc()
5459 struct bnxt_napi *bnapi = rxr->bnapi; in bnxt_hwrm_ring_alloc()
5465 bnxt_set_db(bp, &rxr->rx_db, type, map_idx, ring->fw_ring_id); in bnxt_hwrm_ring_alloc()
5468 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_hwrm_ring_alloc()
5490 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_hwrm_ring_alloc() local
5492 &rxr->rx_agg_ring_struct; in bnxt_hwrm_ring_alloc()
5500 bnxt_set_db(bp, &rxr->rx_agg_db, type, map_idx, in bnxt_hwrm_ring_alloc()
5502 bnxt_db_write(bp, &rxr->rx_agg_db, rxr->rx_agg_prod); in bnxt_hwrm_ring_alloc()
5503 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_hwrm_ring_alloc()
5564 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_hwrm_ring_free() local
5565 struct bnxt_ring_struct *ring = &rxr->rx_ring_struct; in bnxt_hwrm_ring_free()
5566 u32 grp_idx = rxr->bnapi->index; in bnxt_hwrm_ring_free()
5569 u32 cmpl_ring_id = bnxt_cp_ring_for_rx(bp, rxr); in bnxt_hwrm_ring_free()
5586 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[i]; in bnxt_hwrm_ring_free() local
5587 struct bnxt_ring_struct *ring = &rxr->rx_agg_ring_struct; in bnxt_hwrm_ring_free()
5588 u32 grp_idx = rxr->bnapi->index; in bnxt_hwrm_ring_free()
5591 u32 cmpl_ring_id = bnxt_cp_ring_for_rx(bp, rxr); in bnxt_hwrm_ring_free()
9887 struct bnxt_rx_ring_info *rxr = bnapi->rx_ring; in bnxt_dump_rx_sw_state() local
9890 if (!rxr) in bnxt_dump_rx_sw_state()
9894 i, rxr->rx_ring_struct.fw_ring_id, rxr->rx_prod, in bnxt_dump_rx_sw_state()
9895 rxr->rx_agg_ring_struct.fw_ring_id, rxr->rx_agg_prod, in bnxt_dump_rx_sw_state()
9896 rxr->rx_sw_agg_prod); in bnxt_dump_rx_sw_state()