Lines Matching refs:rx_ring
574 struct ixgbe_ring *rx_ring; in ixgbe_dump() local
713 rx_ring = adapter->rx_ring[n]; in ixgbe_dump()
715 n, rx_ring->next_to_use, rx_ring->next_to_clean); in ixgbe_dump()
770 rx_ring = adapter->rx_ring[n]; in ixgbe_dump()
772 pr_info("RX QUEUE INDEX = %d\n", rx_ring->queue_index); in ixgbe_dump()
783 for (i = 0; i < rx_ring->count; i++) { in ixgbe_dump()
786 if (i == rx_ring->next_to_use) in ixgbe_dump()
788 else if (i == rx_ring->next_to_clean) in ixgbe_dump()
793 rx_buffer_info = &rx_ring->rx_buffer_info[i]; in ixgbe_dump()
794 rx_desc = IXGBE_RX_DESC(rx_ring, i); in ixgbe_dump()
819 ixgbe_rx_bufsz(rx_ring), true); in ixgbe_dump()
1310 struct ixgbe_ring *rx_ring, in ixgbe_update_rx_dca() argument
1315 u8 reg_idx = rx_ring->reg_idx; in ixgbe_update_rx_dca()
1318 rxctrl = dca3_get_tag(rx_ring->dev, cpu); in ixgbe_update_rx_dca()
1523 static inline unsigned int ixgbe_rx_offset(struct ixgbe_ring *rx_ring) in ixgbe_rx_offset() argument
1525 return ring_uses_build_skb(rx_ring) ? IXGBE_SKB_PAD : 0; in ixgbe_rx_offset()
1528 static bool ixgbe_alloc_mapped_page(struct ixgbe_ring *rx_ring, in ixgbe_alloc_mapped_page() argument
1539 page = dev_alloc_pages(ixgbe_rx_pg_order(rx_ring)); in ixgbe_alloc_mapped_page()
1541 rx_ring->rx_stats.alloc_rx_page_failed++; in ixgbe_alloc_mapped_page()
1546 dma = dma_map_page_attrs(rx_ring->dev, page, 0, in ixgbe_alloc_mapped_page()
1547 ixgbe_rx_pg_size(rx_ring), in ixgbe_alloc_mapped_page()
1555 if (dma_mapping_error(rx_ring->dev, dma)) { in ixgbe_alloc_mapped_page()
1556 __free_pages(page, ixgbe_rx_pg_order(rx_ring)); in ixgbe_alloc_mapped_page()
1558 rx_ring->rx_stats.alloc_rx_page_failed++; in ixgbe_alloc_mapped_page()
1564 bi->page_offset = ixgbe_rx_offset(rx_ring); in ixgbe_alloc_mapped_page()
1567 rx_ring->rx_stats.alloc_rx_page++; in ixgbe_alloc_mapped_page()
1577 void ixgbe_alloc_rx_buffers(struct ixgbe_ring *rx_ring, u16 cleaned_count) in ixgbe_alloc_rx_buffers() argument
1581 u16 i = rx_ring->next_to_use; in ixgbe_alloc_rx_buffers()
1588 rx_desc = IXGBE_RX_DESC(rx_ring, i); in ixgbe_alloc_rx_buffers()
1589 bi = &rx_ring->rx_buffer_info[i]; in ixgbe_alloc_rx_buffers()
1590 i -= rx_ring->count; in ixgbe_alloc_rx_buffers()
1592 bufsz = ixgbe_rx_bufsz(rx_ring); in ixgbe_alloc_rx_buffers()
1595 if (!ixgbe_alloc_mapped_page(rx_ring, bi)) in ixgbe_alloc_rx_buffers()
1599 dma_sync_single_range_for_device(rx_ring->dev, bi->dma, in ixgbe_alloc_rx_buffers()
1613 rx_desc = IXGBE_RX_DESC(rx_ring, 0); in ixgbe_alloc_rx_buffers()
1614 bi = rx_ring->rx_buffer_info; in ixgbe_alloc_rx_buffers()
1615 i -= rx_ring->count; in ixgbe_alloc_rx_buffers()
1624 i += rx_ring->count; in ixgbe_alloc_rx_buffers()
1626 if (rx_ring->next_to_use != i) { in ixgbe_alloc_rx_buffers()
1627 rx_ring->next_to_use = i; in ixgbe_alloc_rx_buffers()
1630 rx_ring->next_to_alloc = i; in ixgbe_alloc_rx_buffers()
1638 writel(i, rx_ring->tail); in ixgbe_alloc_rx_buffers()
1653 static void ixgbe_update_rsc_stats(struct ixgbe_ring *rx_ring, in ixgbe_update_rsc_stats() argument
1660 rx_ring->rx_stats.rsc_count += IXGBE_CB(skb)->append_cnt; in ixgbe_update_rsc_stats()
1661 rx_ring->rx_stats.rsc_flush++; in ixgbe_update_rsc_stats()
1663 ixgbe_set_rsc_gso_size(rx_ring, skb); in ixgbe_update_rsc_stats()
1679 void ixgbe_process_skb_fields(struct ixgbe_ring *rx_ring, in ixgbe_process_skb_fields() argument
1683 struct net_device *dev = rx_ring->netdev; in ixgbe_process_skb_fields()
1684 u32 flags = rx_ring->q_vector->adapter->flags; in ixgbe_process_skb_fields()
1686 ixgbe_update_rsc_stats(rx_ring, skb); in ixgbe_process_skb_fields()
1688 ixgbe_rx_hash(rx_ring, rx_desc, skb); in ixgbe_process_skb_fields()
1690 ixgbe_rx_checksum(rx_ring, rx_desc, skb); in ixgbe_process_skb_fields()
1693 ixgbe_ptp_rx_hwtstamp(rx_ring, rx_desc, skb); in ixgbe_process_skb_fields()
1702 ixgbe_ipsec_rx(rx_ring, rx_desc, skb); in ixgbe_process_skb_fields()
1706 skb_record_rx_queue(skb, rx_ring->queue_index); in ixgbe_process_skb_fields()
1731 static bool ixgbe_is_non_eop(struct ixgbe_ring *rx_ring, in ixgbe_is_non_eop() argument
1735 u32 ntc = rx_ring->next_to_clean + 1; in ixgbe_is_non_eop()
1738 ntc = (ntc < rx_ring->count) ? ntc : 0; in ixgbe_is_non_eop()
1739 rx_ring->next_to_clean = ntc; in ixgbe_is_non_eop()
1741 prefetch(IXGBE_RX_DESC(rx_ring, ntc)); in ixgbe_is_non_eop()
1744 if (ring_is_rsc_enabled(rx_ring)) { in ixgbe_is_non_eop()
1766 rx_ring->rx_buffer_info[ntc].skb = skb; in ixgbe_is_non_eop()
1767 rx_ring->rx_stats.non_eop_descs++; in ixgbe_is_non_eop()
1784 static void ixgbe_pull_tail(struct ixgbe_ring *rx_ring, in ixgbe_pull_tail() argument
1824 static void ixgbe_dma_sync_frag(struct ixgbe_ring *rx_ring, in ixgbe_dma_sync_frag() argument
1827 if (ring_uses_build_skb(rx_ring)) { in ixgbe_dma_sync_frag()
1830 dma_sync_single_range_for_cpu(rx_ring->dev, in ixgbe_dma_sync_frag()
1838 dma_sync_single_range_for_cpu(rx_ring->dev, in ixgbe_dma_sync_frag()
1847 dma_unmap_page_attrs(rx_ring->dev, IXGBE_CB(skb)->dma, in ixgbe_dma_sync_frag()
1848 ixgbe_rx_pg_size(rx_ring), in ixgbe_dma_sync_frag()
1876 bool ixgbe_cleanup_headers(struct ixgbe_ring *rx_ring, in ixgbe_cleanup_headers() argument
1880 struct net_device *netdev = rx_ring->netdev; in ixgbe_cleanup_headers()
1899 ixgbe_pull_tail(rx_ring, skb); in ixgbe_cleanup_headers()
1903 if (ixgbe_rx_is_fcoe(rx_ring, rx_desc)) in ixgbe_cleanup_headers()
1921 static void ixgbe_reuse_rx_page(struct ixgbe_ring *rx_ring, in ixgbe_reuse_rx_page() argument
1925 u16 nta = rx_ring->next_to_alloc; in ixgbe_reuse_rx_page()
1927 new_buff = &rx_ring->rx_buffer_info[nta]; in ixgbe_reuse_rx_page()
1931 rx_ring->next_to_alloc = (nta < rx_ring->count) ? nta : 0; in ixgbe_reuse_rx_page()
2001 static void ixgbe_add_rx_frag(struct ixgbe_ring *rx_ring, in ixgbe_add_rx_frag() argument
2007 unsigned int truesize = ixgbe_rx_pg_size(rx_ring) / 2; in ixgbe_add_rx_frag()
2009 unsigned int truesize = ring_uses_build_skb(rx_ring) ? in ixgbe_add_rx_frag()
2022 static struct ixgbe_rx_buffer *ixgbe_get_rx_buffer(struct ixgbe_ring *rx_ring, in ixgbe_get_rx_buffer() argument
2030 rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; in ixgbe_get_rx_buffer()
2049 ixgbe_dma_sync_frag(rx_ring, *skb); in ixgbe_get_rx_buffer()
2053 dma_sync_single_range_for_cpu(rx_ring->dev, in ixgbe_get_rx_buffer()
2064 static void ixgbe_put_rx_buffer(struct ixgbe_ring *rx_ring, in ixgbe_put_rx_buffer() argument
2071 ixgbe_reuse_rx_page(rx_ring, rx_buffer); in ixgbe_put_rx_buffer()
2078 dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, in ixgbe_put_rx_buffer()
2079 ixgbe_rx_pg_size(rx_ring), in ixgbe_put_rx_buffer()
2092 static struct sk_buff *ixgbe_construct_skb(struct ixgbe_ring *rx_ring, in ixgbe_construct_skb() argument
2099 unsigned int truesize = ixgbe_rx_pg_size(rx_ring) / 2; in ixgbe_construct_skb()
2126 skb = napi_alloc_skb(&rx_ring->q_vector->napi, IXGBE_RX_HDR_SIZE); in ixgbe_construct_skb()
2151 static struct sk_buff *ixgbe_build_skb(struct ixgbe_ring *rx_ring, in ixgbe_build_skb() argument
2158 unsigned int truesize = ixgbe_rx_pg_size(rx_ring) / 2; in ixgbe_build_skb()
2199 struct ixgbe_ring *rx_ring, in ixgbe_run_xdp() argument
2208 xdp_prog = READ_ONCE(rx_ring->xdp_prog); in ixgbe_run_xdp()
2238 trace_xdp_exception(rx_ring->netdev, xdp_prog, act); in ixgbe_run_xdp()
2249 static unsigned int ixgbe_rx_frame_truesize(struct ixgbe_ring *rx_ring, in ixgbe_rx_frame_truesize() argument
2255 truesize = ixgbe_rx_pg_size(rx_ring) / 2; /* Must be power-of-2 */ in ixgbe_rx_frame_truesize()
2257 truesize = ring_uses_build_skb(rx_ring) ? in ixgbe_rx_frame_truesize()
2265 static void ixgbe_rx_buffer_flip(struct ixgbe_ring *rx_ring, in ixgbe_rx_buffer_flip() argument
2269 unsigned int truesize = ixgbe_rx_frame_truesize(rx_ring, size); in ixgbe_rx_buffer_flip()
2291 struct ixgbe_ring *rx_ring, in ixgbe_clean_rx_irq() argument
2300 u16 cleaned_count = ixgbe_desc_unused(rx_ring); in ixgbe_clean_rx_irq()
2304 xdp.rxq = &rx_ring->xdp_rxq; in ixgbe_clean_rx_irq()
2308 xdp.frame_sz = ixgbe_rx_frame_truesize(rx_ring, 0); in ixgbe_clean_rx_irq()
2320 ixgbe_alloc_rx_buffers(rx_ring, cleaned_count); in ixgbe_clean_rx_irq()
2324 rx_desc = IXGBE_RX_DESC(rx_ring, rx_ring->next_to_clean); in ixgbe_clean_rx_irq()
2335 rx_buffer = ixgbe_get_rx_buffer(rx_ring, rx_desc, &skb, size, &rx_buffer_pgcnt); in ixgbe_clean_rx_irq()
2343 ixgbe_rx_offset(rx_ring); in ixgbe_clean_rx_irq()
2347 xdp.frame_sz = ixgbe_rx_frame_truesize(rx_ring, size); in ixgbe_clean_rx_irq()
2349 skb = ixgbe_run_xdp(adapter, rx_ring, &xdp); in ixgbe_clean_rx_irq()
2357 ixgbe_rx_buffer_flip(rx_ring, rx_buffer, size); in ixgbe_clean_rx_irq()
2364 ixgbe_add_rx_frag(rx_ring, rx_buffer, skb, size); in ixgbe_clean_rx_irq()
2365 } else if (ring_uses_build_skb(rx_ring)) { in ixgbe_clean_rx_irq()
2366 skb = ixgbe_build_skb(rx_ring, rx_buffer, in ixgbe_clean_rx_irq()
2369 skb = ixgbe_construct_skb(rx_ring, rx_buffer, in ixgbe_clean_rx_irq()
2375 rx_ring->rx_stats.alloc_rx_buff_failed++; in ixgbe_clean_rx_irq()
2380 ixgbe_put_rx_buffer(rx_ring, rx_buffer, skb, rx_buffer_pgcnt); in ixgbe_clean_rx_irq()
2384 if (ixgbe_is_non_eop(rx_ring, rx_desc, skb)) in ixgbe_clean_rx_irq()
2388 if (ixgbe_cleanup_headers(rx_ring, rx_desc, skb)) in ixgbe_clean_rx_irq()
2395 ixgbe_process_skb_fields(rx_ring, rx_desc, skb); in ixgbe_clean_rx_irq()
2399 if (ixgbe_rx_is_fcoe(rx_ring, rx_desc)) { in ixgbe_clean_rx_irq()
2404 mss = rx_ring->netdev->mtu - in ixgbe_clean_rx_irq()
2441 u64_stats_update_begin(&rx_ring->syncp); in ixgbe_clean_rx_irq()
2442 rx_ring->stats.packets += total_rx_packets; in ixgbe_clean_rx_irq()
2443 rx_ring->stats.bytes += total_rx_bytes; in ixgbe_clean_rx_irq()
2444 u64_stats_update_end(&rx_ring->syncp); in ixgbe_clean_rx_irq()
3691 ixgbe_enable_rx_drop(adapter, adapter->rx_ring[i]); in ixgbe_set_rx_drop_en()
3694 ixgbe_disable_rx_drop(adapter, adapter->rx_ring[i]); in ixgbe_set_rx_drop_en()
3701 struct ixgbe_ring *rx_ring) in ixgbe_configure_srrctl() argument
3705 u8 reg_idx = rx_ring->reg_idx; in ixgbe_configure_srrctl()
3721 if (rx_ring->xsk_pool) { in ixgbe_configure_srrctl()
3722 u32 xsk_buf_len = xsk_pool_get_rx_frame_size(rx_ring->xsk_pool); in ixgbe_configure_srrctl()
3736 } else if (test_bit(__IXGBE_RX_3K_BUFFER, &rx_ring->state)) { in ixgbe_configure_srrctl()
4257 struct ixgbe_ring *rx_ring; in ixgbe_set_rx_buffer_len() local
4291 rx_ring = adapter->rx_ring[i]; in ixgbe_set_rx_buffer_len()
4293 clear_ring_rsc_enabled(rx_ring); in ixgbe_set_rx_buffer_len()
4294 clear_bit(__IXGBE_RX_3K_BUFFER, &rx_ring->state); in ixgbe_set_rx_buffer_len()
4295 clear_bit(__IXGBE_RX_BUILD_SKB_ENABLED, &rx_ring->state); in ixgbe_set_rx_buffer_len()
4298 set_ring_rsc_enabled(rx_ring); in ixgbe_set_rx_buffer_len()
4300 if (test_bit(__IXGBE_RX_FCOE, &rx_ring->state)) in ixgbe_set_rx_buffer_len()
4301 set_bit(__IXGBE_RX_3K_BUFFER, &rx_ring->state); in ixgbe_set_rx_buffer_len()
4306 set_bit(__IXGBE_RX_BUILD_SKB_ENABLED, &rx_ring->state); in ixgbe_set_rx_buffer_len()
4310 set_bit(__IXGBE_RX_3K_BUFFER, &rx_ring->state); in ixgbe_set_rx_buffer_len()
4314 set_bit(__IXGBE_RX_3K_BUFFER, &rx_ring->state); in ixgbe_set_rx_buffer_len()
4401 ixgbe_configure_rx_ring(adapter, adapter->rx_ring[i]); in ixgbe_configure_rx()
4509 struct ixgbe_ring *ring = adapter->rx_ring[i]; in ixgbe_vlan_strip_disable()
4547 struct ixgbe_ring *ring = adapter->rx_ring[i]; in ixgbe_vlan_strip_enable()
5278 queue = adapter->rx_ring[ring]->reg_idx; in ixgbe_fdir_filter_restore()
5295 static void ixgbe_clean_rx_ring(struct ixgbe_ring *rx_ring) in ixgbe_clean_rx_ring() argument
5297 u16 i = rx_ring->next_to_clean; in ixgbe_clean_rx_ring()
5298 struct ixgbe_rx_buffer *rx_buffer = &rx_ring->rx_buffer_info[i]; in ixgbe_clean_rx_ring()
5300 if (rx_ring->xsk_pool) { in ixgbe_clean_rx_ring()
5301 ixgbe_xsk_clean_rx_ring(rx_ring); in ixgbe_clean_rx_ring()
5306 while (i != rx_ring->next_to_alloc) { in ixgbe_clean_rx_ring()
5310 dma_unmap_page_attrs(rx_ring->dev, in ixgbe_clean_rx_ring()
5312 ixgbe_rx_pg_size(rx_ring), in ixgbe_clean_rx_ring()
5321 dma_sync_single_range_for_cpu(rx_ring->dev, in ixgbe_clean_rx_ring()
5324 ixgbe_rx_bufsz(rx_ring), in ixgbe_clean_rx_ring()
5328 dma_unmap_page_attrs(rx_ring->dev, rx_buffer->dma, in ixgbe_clean_rx_ring()
5329 ixgbe_rx_pg_size(rx_ring), in ixgbe_clean_rx_ring()
5337 if (i == rx_ring->count) { in ixgbe_clean_rx_ring()
5339 rx_buffer = rx_ring->rx_buffer_info; in ixgbe_clean_rx_ring()
5344 rx_ring->next_to_alloc = 0; in ixgbe_clean_rx_ring()
5345 rx_ring->next_to_clean = 0; in ixgbe_clean_rx_ring()
5346 rx_ring->next_to_use = 0; in ixgbe_clean_rx_ring()
5371 adapter->rx_ring[baseq + i]->netdev = vdev; in ixgbe_fwd_ring_up()
5390 adapter->rx_ring[baseq + i]->netdev = NULL; in ixgbe_fwd_ring_up()
5770 struct ixgbe_ring *ring = adapter->rx_ring[i]; in ixgbe_disable_rx()
5813 struct ixgbe_ring *ring = adapter->rx_ring[i]; in ixgbe_disable_rx()
6067 ixgbe_clean_rx_ring(adapter->rx_ring[i]); in ixgbe_clean_all_rx_rings()
6551 struct ixgbe_ring *rx_ring) in ixgbe_setup_rx_resources() argument
6553 struct device *dev = rx_ring->dev; in ixgbe_setup_rx_resources()
6558 size = sizeof(struct ixgbe_rx_buffer) * rx_ring->count; in ixgbe_setup_rx_resources()
6560 if (rx_ring->q_vector) in ixgbe_setup_rx_resources()
6561 ring_node = rx_ring->q_vector->numa_node; in ixgbe_setup_rx_resources()
6563 rx_ring->rx_buffer_info = vmalloc_node(size, ring_node); in ixgbe_setup_rx_resources()
6564 if (!rx_ring->rx_buffer_info) in ixgbe_setup_rx_resources()
6565 rx_ring->rx_buffer_info = vmalloc(size); in ixgbe_setup_rx_resources()
6566 if (!rx_ring->rx_buffer_info) in ixgbe_setup_rx_resources()
6570 rx_ring->size = rx_ring->count * sizeof(union ixgbe_adv_rx_desc); in ixgbe_setup_rx_resources()
6571 rx_ring->size = ALIGN(rx_ring->size, 4096); in ixgbe_setup_rx_resources()
6574 rx_ring->desc = dma_alloc_coherent(dev, in ixgbe_setup_rx_resources()
6575 rx_ring->size, in ixgbe_setup_rx_resources()
6576 &rx_ring->dma, in ixgbe_setup_rx_resources()
6579 if (!rx_ring->desc) in ixgbe_setup_rx_resources()
6580 rx_ring->desc = dma_alloc_coherent(dev, rx_ring->size, in ixgbe_setup_rx_resources()
6581 &rx_ring->dma, GFP_KERNEL); in ixgbe_setup_rx_resources()
6582 if (!rx_ring->desc) in ixgbe_setup_rx_resources()
6585 rx_ring->next_to_clean = 0; in ixgbe_setup_rx_resources()
6586 rx_ring->next_to_use = 0; in ixgbe_setup_rx_resources()
6589 if (xdp_rxq_info_reg(&rx_ring->xdp_rxq, adapter->netdev, in ixgbe_setup_rx_resources()
6590 rx_ring->queue_index) < 0) in ixgbe_setup_rx_resources()
6593 rx_ring->xdp_prog = adapter->xdp_prog; in ixgbe_setup_rx_resources()
6597 vfree(rx_ring->rx_buffer_info); in ixgbe_setup_rx_resources()
6598 rx_ring->rx_buffer_info = NULL; in ixgbe_setup_rx_resources()
6618 err = ixgbe_setup_rx_resources(adapter, adapter->rx_ring[i]); in ixgbe_setup_all_rx_resources()
6634 ixgbe_free_rx_resources(adapter->rx_ring[i]); in ixgbe_setup_all_rx_resources()
6685 void ixgbe_free_rx_resources(struct ixgbe_ring *rx_ring) in ixgbe_free_rx_resources() argument
6687 ixgbe_clean_rx_ring(rx_ring); in ixgbe_free_rx_resources()
6689 rx_ring->xdp_prog = NULL; in ixgbe_free_rx_resources()
6690 xdp_rxq_info_unreg(&rx_ring->xdp_rxq); in ixgbe_free_rx_resources()
6691 vfree(rx_ring->rx_buffer_info); in ixgbe_free_rx_resources()
6692 rx_ring->rx_buffer_info = NULL; in ixgbe_free_rx_resources()
6695 if (!rx_ring->desc) in ixgbe_free_rx_resources()
6698 dma_free_coherent(rx_ring->dev, rx_ring->size, in ixgbe_free_rx_resources()
6699 rx_ring->desc, rx_ring->dma); in ixgbe_free_rx_resources()
6701 rx_ring->desc = NULL; in ixgbe_free_rx_resources()
6719 if (adapter->rx_ring[i]->desc) in ixgbe_free_all_rx_resources()
6720 ixgbe_free_rx_resources(adapter->rx_ring[i]); in ixgbe_free_all_rx_resources()
6740 struct ixgbe_ring *ring = adapter->rx_ring[i]; in ixgbe_change_mtu()
7042 rsc_count += adapter->rx_ring[i]->rx_stats.rsc_count; in ixgbe_update_stats()
7043 rsc_flush += adapter->rx_ring[i]->rx_stats.rsc_flush; in ixgbe_update_stats()
7050 struct ixgbe_ring *rx_ring = READ_ONCE(adapter->rx_ring[i]); in ixgbe_update_stats() local
7052 if (!rx_ring) in ixgbe_update_stats()
7054 non_eop_descs += rx_ring->rx_stats.non_eop_descs; in ixgbe_update_stats()
7055 alloc_rx_page += rx_ring->rx_stats.alloc_rx_page; in ixgbe_update_stats()
7056 alloc_rx_page_failed += rx_ring->rx_stats.alloc_rx_page_failed; in ixgbe_update_stats()
7057 alloc_rx_buff_failed += rx_ring->rx_stats.alloc_rx_buff_failed; in ixgbe_update_stats()
7058 hw_csum_rx_error += rx_ring->rx_stats.csum_err; in ixgbe_update_stats()
7059 bytes += rx_ring->stats.bytes; in ixgbe_update_stats()
7060 packets += rx_ring->stats.packets; in ixgbe_update_stats()
8937 struct ixgbe_ring *ring = READ_ONCE(adapter->rx_ring[i]); in ixgbe_get_stats64()
9272 data->queue = adapter->rx_ring[vadapter->rx_base_queue]->reg_idx; in get_macvlan_queue()
10051 struct ixgbe_ring *ring = adapter->rx_ring[rxbase + i]; in ixgbe_fwd_del()
10127 struct ixgbe_ring *ring = adapter->rx_ring[i]; in ixgbe_xdp_setup()
10157 (void)xchg(&adapter->rx_ring[i]->xdp_prog, in ixgbe_xdp_setup()
10326 struct ixgbe_ring *rx_ring) in ixgbe_disable_rxr_hw() argument
10330 u8 reg_idx = rx_ring->reg_idx; in ixgbe_disable_rxr_hw()
10370 static void ixgbe_reset_rxr_stats(struct ixgbe_ring *rx_ring) in ixgbe_reset_rxr_stats() argument
10372 memset(&rx_ring->stats, 0, sizeof(rx_ring->stats)); in ixgbe_reset_rxr_stats()
10373 memset(&rx_ring->rx_stats, 0, sizeof(rx_ring->rx_stats)); in ixgbe_reset_rxr_stats()
10386 struct ixgbe_ring *rx_ring, *tx_ring, *xdp_ring; in ixgbe_txrx_ring_disable() local
10388 rx_ring = adapter->rx_ring[ring]; in ixgbe_txrx_ring_disable()
10395 ixgbe_disable_rxr_hw(adapter, rx_ring); in ixgbe_txrx_ring_disable()
10401 napi_disable(&rx_ring->q_vector->napi); in ixgbe_txrx_ring_disable()
10406 ixgbe_clean_rx_ring(rx_ring); in ixgbe_txrx_ring_disable()
10411 ixgbe_reset_rxr_stats(rx_ring); in ixgbe_txrx_ring_disable()
10424 struct ixgbe_ring *rx_ring, *tx_ring, *xdp_ring; in ixgbe_txrx_ring_enable() local
10426 rx_ring = adapter->rx_ring[ring]; in ixgbe_txrx_ring_enable()
10431 napi_enable(&rx_ring->q_vector->napi); in ixgbe_txrx_ring_enable()
10436 ixgbe_configure_rx_ring(adapter, rx_ring); in ixgbe_txrx_ring_enable()
10935 u64_stats_init(&adapter->rx_ring[i]->syncp); in ixgbe_probe()