Home
last modified time | relevance | path

Searched refs:xsk_pool (Results 1 – 25 of 26) sorted by relevance

12

/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
Drx.h25 dma_info->xsk = xsk_buff_alloc(rq->xsk_pool); in mlx5e_xsk_page_alloc_pool()
41 if (!xsk_uses_need_wakeup(rq->xsk_pool)) in mlx5e_xsk_update_rx_wakeup()
45 xsk_set_rx_need_wakeup(rq->xsk_pool); in mlx5e_xsk_update_rx_wakeup()
47 xsk_clear_rx_need_wakeup(rq->xsk_pool); in mlx5e_xsk_update_rx_wakeup()
Dtx.h18 if (!xsk_uses_need_wakeup(sq->xsk_pool)) in mlx5e_xsk_update_tx_wakeup()
22 xsk_clear_tx_need_wakeup(sq->xsk_pool); in mlx5e_xsk_update_tx_wakeup()
24 xsk_set_tx_need_wakeup(sq->xsk_pool); in mlx5e_xsk_update_tx_wakeup()
Drx.c50 xsk_buff_dma_sync_for_cpu(xdp, rq->xsk_pool); in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
96 xsk_buff_dma_sync_for_cpu(xdp, rq->xsk_pool); in mlx5e_xsk_skb_from_cqe_linear()
Dtx.c69 struct xsk_buff_pool *pool = sq->xsk_pool; in mlx5e_xsk_tx()
Dsetup.c69 rq->xsk_pool = pool; in mlx5e_init_xsk_rq()
/Linux-v5.15/drivers/net/ethernet/intel/i40e/
Di40e_xsk.c203 xdp = xsk_buff_alloc(rx_ring->xsk_pool); in i40e_alloc_rx_buffers_zc()
369 xsk_buff_dma_sync_for_cpu(bi, rx_ring->xsk_pool); in i40e_clean_rx_irq_zc()
389 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in i40e_clean_rx_irq_zc()
391 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in i40e_clean_rx_irq_zc()
393 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in i40e_clean_rx_irq_zc()
406 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc->addr); in i40e_xmit_pkt()
407 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, desc->len); in i40e_xmit_pkt()
426 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc[i].addr); in i40e_xmit_pkt_batch()
427 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, desc[i].len); in i40e_xmit_pkt_batch()
476 nb_pkts = xsk_tx_peek_release_desc_batch(xdp_ring->xsk_pool, descs, budget); in i40e_xmit_zc()
[all …]
Di40e_txrx.h392 struct xsk_buff_pool *xsk_pool; member
Di40e_txrx.c793 if (ring_is_xdp(tx_ring) && tx_ring->xsk_pool) { in i40e_clean_tx_ring()
1503 if (rx_ring->xsk_pool) { in i40e_clean_rx_ring()
1537 if (rx_ring->xsk_pool) in i40e_clean_rx_ring()
2709 bool wd = ring->xsk_pool ? in i40e_napi_poll()
2737 int cleaned = ring->xsk_pool ? in i40e_napi_poll()
Di40e_main.c3164 ring->xsk_pool = i40e_xsk_pool(ring); in i40e_configure_tx_ring()
3298 ring->xsk_pool = i40e_xsk_pool(ring); in i40e_configure_rx_ring()
3299 if (ring->xsk_pool) { in i40e_configure_rx_ring()
3304 xsk_pool_get_rx_frame_size(ring->xsk_pool); in i40e_configure_rx_ring()
3389 if (ring->xsk_pool) { in i40e_configure_rx_ring()
3390 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in i40e_configure_rx_ring()
3401 ring->xsk_pool ? "AF_XDP ZC enabled " : "", in i40e_configure_rx_ring()
12976 if (vsi->xdp_rings[i]->xsk_pool) in i40e_xdp_setup()
/Linux-v5.15/drivers/net/ethernet/intel/ice/
Dice_xsk.c236 xdp_ring->xsk_pool = ice_xsk_pool(xdp_ring); in ice_qp_ena()
378 rx_buf->xdp = xsk_buff_alloc(rx_ring->xsk_pool); in ice_alloc_rx_bufs_zc()
549 xsk_buff_dma_sync_for_cpu(rx_buf->xdp, rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
602 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ice_clean_rx_irq_zc()
604 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
606 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ice_clean_rx_irq_zc()
639 if (!xsk_tx_peek_desc(xdp_ring->xsk_pool, &desc)) in ice_xmit_zc()
642 dma = xsk_buff_raw_get_dma(xdp_ring->xsk_pool, desc.addr); in ice_xmit_zc()
643 xsk_buff_raw_dma_sync_for_device(xdp_ring->xsk_pool, dma, in ice_xmit_zc()
660 xsk_tx_release(xdp_ring->xsk_pool); in ice_xmit_zc()
[all …]
Dice_base.c387 if (ring->xsk_pool) in ice_setup_rx_ctx()
456 ring->xsk_pool = ice_xsk_pool(ring); in ice_vsi_cfg_rxq()
457 if (ring->xsk_pool) { in ice_vsi_cfg_rxq()
461 xsk_pool_get_rx_frame_size(ring->xsk_pool); in ice_vsi_cfg_rxq()
467 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in ice_vsi_cfg_rxq()
493 if (ring->xsk_pool) { in ice_vsi_cfg_rxq()
496 if (!xsk_buff_can_alloc(ring->xsk_pool, num_bufs)) { in ice_vsi_cfg_rxq()
Dice_txrx.h299 struct xsk_buff_pool *xsk_pool; member
Dice_txrx.c149 if (ice_ring_is_xdp(tx_ring) && tx_ring->xsk_pool) { in ice_clean_tx_ring()
387 if (rx_ring->xsk_pool) { in ice_clean_rx_ring()
1399 bool wd = ring->xsk_pool ? in ice_napi_poll()
1429 cleaned = ring->xsk_pool ? in ice_napi_poll()
Dice_lib.c1821 vsi->xdp_rings[i]->xsk_pool = ice_xsk_pool(vsi->xdp_rings[i]); in ice_vsi_cfg_xdp_txqs()
/Linux-v5.15/drivers/net/ethernet/intel/ixgbe/
Dixgbe_xsk.c158 bi->xdp = xsk_buff_alloc(rx_ring->xsk_pool); in ixgbe_alloc_rx_buffers_zc()
295 xsk_buff_dma_sync_for_cpu(bi->xdp, rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc()
353 if (xsk_uses_need_wakeup(rx_ring->xsk_pool)) { in ixgbe_clean_rx_irq_zc()
355 xsk_set_rx_need_wakeup(rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc()
357 xsk_clear_rx_need_wakeup(rx_ring->xsk_pool); in ixgbe_clean_rx_irq_zc()
382 struct xsk_buff_pool *pool = xdp_ring->xsk_pool; in ixgbe_xmit_zc()
448 struct xsk_buff_pool *pool = tx_ring->xsk_pool; in ixgbe_clean_xdp_tx_irq()
520 if (!ring->xsk_pool) in ixgbe_xsk_wakeup()
535 struct xsk_buff_pool *pool = tx_ring->xsk_pool; in ixgbe_xsk_clean_tx_ring()
Dixgbe_main.c3157 bool wd = ring->xsk_pool ? in ixgbe_poll()
3177 int cleaned = ring->xsk_pool ? in ixgbe_poll()
3472 ring->xsk_pool = NULL; in ixgbe_configure_tx_ring()
3474 ring->xsk_pool = ixgbe_xsk_pool(adapter, ring); in ixgbe_configure_tx_ring()
3714 if (rx_ring->xsk_pool) { in ixgbe_configure_srrctl()
3715 u32 xsk_buf_len = xsk_pool_get_rx_frame_size(rx_ring->xsk_pool); in ixgbe_configure_srrctl()
4060 ring->xsk_pool = ixgbe_xsk_pool(adapter, ring); in ixgbe_configure_rx_ring()
4061 if (ring->xsk_pool) { in ixgbe_configure_rx_ring()
4065 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in ixgbe_configure_rx_ring()
4122 if (ring->xsk_pool && hw->mac.type != ixgbe_mac_82599EB) { in ixgbe_configure_rx_ring()
[all …]
Dixgbe.h354 struct xsk_buff_pool *xsk_pool; member
/Linux-v5.15/drivers/net/ethernet/stmicro/stmmac/
Dstmmac.h73 struct xsk_buff_pool *xsk_pool; member
99 struct xsk_buff_pool *xsk_pool; member
Dstmmac_main.c226 if (rx_q->xsk_pool) { in stmmac_disable_all_queues()
1600 buf->xdp = xsk_buff_alloc(rx_q->xsk_pool); in stmmac_alloc_rx_buffers_zc()
1642 rx_q->xsk_pool = stmmac_get_xsk_pool(priv, queue); in __init_dma_rx_desc_rings()
1644 if (rx_q->xsk_pool) { in __init_dma_rx_desc_rings()
1651 xsk_pool_set_rxq_info(rx_q->xsk_pool, &rx_q->xdp_rxq); in __init_dma_rx_desc_rings()
1661 if (rx_q->xsk_pool) { in __init_dma_rx_desc_rings()
1713 if (rx_q->xsk_pool) in init_dma_rx_desc_rings()
1719 rx_q->xsk_pool = NULL; in init_dma_rx_desc_rings()
1759 tx_q->xsk_pool = stmmac_get_xsk_pool(priv, queue); in __init_dma_tx_desc_rings()
1845 if (tx_q->xsk_pool && tx_q->xsk_frames_done) { in dma_free_tx_skbufs()
[all …]
/Linux-v5.15/drivers/net/ethernet/intel/igc/
Digc_main.c251 if (tx_ring->xsk_pool && xsk_frames) in igc_clean_tx_ring()
252 xsk_tx_completed(tx_ring->xsk_pool, xsk_frames); in igc_clean_tx_ring()
430 if (ring->xsk_pool) in igc_clean_rx_ring()
599 ring->xsk_pool = igc_get_xsk_pool(adapter, ring); in igc_configure_rx_ring()
600 if (ring->xsk_pool) { in igc_configure_rx_ring()
604 xsk_pool_set_rxq_info(ring->xsk_pool, &ring->xdp_rxq); in igc_configure_rx_ring()
633 if (ring->xsk_pool) in igc_configure_rx_ring()
634 buf_size = xsk_pool_get_rx_frame_size(ring->xsk_pool); in igc_configure_rx_ring()
696 ring->xsk_pool = igc_get_xsk_pool(adapter, ring); in igc_configure_tx_ring()
2077 bi->xdp = xsk_buff_alloc(ring->xsk_pool); in igc_alloc_rx_buffers_zc()
[all …]
Digc.h130 struct xsk_buff_pool *xsk_pool; member
/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/
Den_rx.c288 if (rq->xsk_pool) in mlx5e_page_alloc()
320 if (rq->xsk_pool) in mlx5e_page_release()
407 if (rq->xsk_pool) { in mlx5e_alloc_rx_wqes()
414 if (unlikely(!xsk_buff_can_alloc(rq->xsk_pool, pages_desired))) in mlx5e_alloc_rx_wqes()
511 if (rq->xsk_pool && in mlx5e_alloc_rx_mpwqe()
512 unlikely(!xsk_buff_can_alloc(rq->xsk_pool, MLX5_MPWRQ_PAGES_PER_WQE))) { in mlx5e_alloc_rx_mpwqe()
768 if (unlikely(alloc_err == -ENOMEM && rq->xsk_pool)) in mlx5e_post_rx_mpwqes()
Den_main.c496 xsk_pool_set_rxq_info(rq->xsk_pool, &rq->xdp_rxq); in mlx5e_alloc_rq()
967 struct xsk_buff_pool *xsk_pool, in mlx5e_alloc_xdpsq() argument
983 sq->xsk_pool = xsk_pool; in mlx5e_alloc_xdpsq()
985 sq->stats = sq->xsk_pool ? in mlx5e_alloc_xdpsq()
1459 struct mlx5e_sq_param *param, struct xsk_buff_pool *xsk_pool, in mlx5e_open_xdpsq() argument
1465 err = mlx5e_alloc_xdpsq(c, params, xsk_pool, param, sq, is_redirect); in mlx5e_open_xdpsq()
1969 struct xsk_buff_pool *xsk_pool, in mlx5e_open_channel() argument
2007 if (xsk_pool) { in mlx5e_open_channel()
2008 mlx5e_build_xsk_param(xsk_pool, &xsk); in mlx5e_open_channel()
2009 err = mlx5e_open_xsk(priv, params, &xsk, xsk_pool, c); in mlx5e_open_channel()
[all …]
Den.h488 struct xsk_buff_pool *xsk_pool; member
658 struct xsk_buff_pool *xsk_pool; member
951 struct mlx5e_sq_param *param, struct xsk_buff_pool *xsk_pool,
/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/en/
Dxdp.c457 xsk_tx_completed(sq->xsk_pool, xsk_frames); in mlx5e_poll_xdpsq_cq()
495 xsk_tx_completed(sq->xsk_pool, xsk_frames); in mlx5e_free_xdpsq_descs()

12