/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | xdp.c | 42 xdpi.xdpf = convert_to_xdp_frame(xdp); in mlx5e_xmit_xdp_buff() 43 if (unlikely(!xdpi.xdpf)) in mlx5e_xmit_xdp_buff() 45 xdpi.dma_addr = di->addr + (xdpi.xdpf->data - (void *)xdpi.xdpf); in mlx5e_xmit_xdp_buff() 47 xdpi.xdpf->len, PCI_DMA_TODEVICE); in mlx5e_xmit_xdp_buff() 115 struct xdp_frame *xdpf = xdpi->xdpf; in mlx5e_xmit_xdp_frame() local 117 unsigned int dma_len = xdpf->len; in mlx5e_xmit_xdp_frame() 142 memcpy(eseg->inline_hdr.start, xdpf->data, MLX5E_XDP_MIN_INLINE); in mlx5e_xmit_xdp_frame() 210 xdp_return_frame(xdpi->xdpf); in mlx5e_poll_xdpsq_cq() 212 xdpi->xdpf->len, DMA_TO_DEVICE); in mlx5e_poll_xdpsq_cq() 246 xdp_return_frame(xdpi->xdpf); in mlx5e_free_xdpsq_descs() [all …]
|
/Linux-v4.19/kernel/bpf/ |
D | cpumap.c | 163 struct xdp_frame *xdpf) in cpu_map_build_skb() argument 186 frame_size = SKB_DATA_ALIGN(xdpf->len) + xdpf->headroom + in cpu_map_build_skb() 189 pkt_data_start = xdpf->data - xdpf->headroom; in cpu_map_build_skb() 194 skb_reserve(skb, xdpf->headroom); in cpu_map_build_skb() 195 __skb_put(skb, xdpf->len); in cpu_map_build_skb() 196 if (xdpf->metasize) in cpu_map_build_skb() 197 skb_metadata_set(skb, xdpf->metasize); in cpu_map_build_skb() 200 skb->protocol = eth_type_trans(skb, xdpf->dev_rx); in cpu_map_build_skb() 218 struct xdp_frame *xdpf; in __cpu_map_ring_cleanup() local 220 while ((xdpf = ptr_ring_consume(ring))) in __cpu_map_ring_cleanup() [all …]
|
D | devmap.c | 232 struct xdp_frame *xdpf = bq->q[i]; in bq_xmit_all() local 234 prefetch(xdpf); in bq_xmit_all() 256 struct xdp_frame *xdpf = bq->q[i]; in bq_xmit_all() local 260 xdp_return_frame_rx_napi(xdpf); in bq_xmit_all() 262 xdp_return_frame(xdpf); in bq_xmit_all() 317 static int bq_enqueue(struct bpf_dtab_netdev *obj, struct xdp_frame *xdpf, in bq_enqueue() argument 333 bq->q[bq->count++] = xdpf; in bq_enqueue() 341 struct xdp_frame *xdpf; in dev_map_enqueue() local 351 xdpf = convert_to_xdp_frame(xdp); in dev_map_enqueue() 352 if (unlikely(!xdpf)) in dev_map_enqueue() [all …]
|
/Linux-v4.19/net/core/ |
D | xdp.c | 353 void xdp_return_frame(struct xdp_frame *xdpf) in xdp_return_frame() argument 355 __xdp_return(xdpf->data, &xdpf->mem, false, 0); in xdp_return_frame() 359 void xdp_return_frame_rx_napi(struct xdp_frame *xdpf) in xdp_return_frame_rx_napi() argument 361 __xdp_return(xdpf->data, &xdpf->mem, true, 0); in xdp_return_frame_rx_napi()
|
D | filter.c | 3158 struct xdp_frame *xdpf; in __bpf_tx_xdp() local 3169 xdpf = convert_to_xdp_frame(xdp); in __bpf_tx_xdp() 3170 if (unlikely(!xdpf)) in __bpf_tx_xdp() 3173 sent = dev->netdev_ops->ndo_xdp_xmit(dev, 1, &xdpf, XDP_XMIT_FLUSH); in __bpf_tx_xdp()
|
/Linux-v4.19/include/net/ |
D | xdp.h | 127 void xdp_return_frame(struct xdp_frame *xdpf); 128 void xdp_return_frame_rx_napi(struct xdp_frame *xdpf);
|
/Linux-v4.19/drivers/net/ |
D | virtio_net.c | 440 struct xdp_frame *xdpf) in __virtnet_xdp_xmit_one() argument 446 if (unlikely(xdpf->metasize > 0)) in __virtnet_xdp_xmit_one() 449 if (unlikely(xdpf->headroom < vi->hdr_len)) in __virtnet_xdp_xmit_one() 453 xdpf->data -= vi->hdr_len; in __virtnet_xdp_xmit_one() 455 hdr = xdpf->data; in __virtnet_xdp_xmit_one() 457 xdpf->len += vi->hdr_len; in __virtnet_xdp_xmit_one() 459 sg_init_one(sq->sg, xdpf->data, xdpf->len); in __virtnet_xdp_xmit_one() 461 err = virtqueue_add_outbuf(sq->vq, sq->sg, 1, xdpf, GFP_ATOMIC); in __virtnet_xdp_xmit_one() 513 struct xdp_frame *xdpf = frames[i]; in virtnet_xdp_xmit() local 515 err = __virtnet_xdp_xmit_one(vi, sq, xdpf); in virtnet_xdp_xmit() [all …]
|
D | tun.c | 666 struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); in tun_ptr_free() local 668 xdp_return_frame(xdpf); in tun_ptr_free() 2167 struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); in tun_do_read() local 2169 ret = tun_put_user_xdp(tun, tfile, xdpf, to); in tun_do_read() 2170 xdp_return_frame(xdpf); in tun_do_read() 2411 struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); in tun_ptr_peek_len() local 2413 return xdpf->len; in tun_ptr_peek_len()
|
/Linux-v4.19/drivers/net/ethernet/intel/i40e/ |
D | i40e_txrx.c | 617 xdp_return_frame(tx_buffer->xdpf); in i40e_unmap_and_free_tx_resource() 820 xdp_return_frame(tx_buf->xdpf); in i40e_clean_tx_irq() 2207 static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf, 2213 struct xdp_frame *xdpf = convert_to_xdp_frame(xdp); in i40e_xmit_xdp_tx_ring() local 2215 if (unlikely(!xdpf)) in i40e_xmit_xdp_tx_ring() 2218 return i40e_xmit_xdp_ring(xdpf, xdp_ring); in i40e_xmit_xdp_tx_ring() 3501 static int i40e_xmit_xdp_ring(struct xdp_frame *xdpf, in i40e_xmit_xdp_ring() argument 3507 u32 size = xdpf->len; in i40e_xmit_xdp_ring() 3515 dma = dma_map_single(xdp_ring->dev, xdpf->data, size, DMA_TO_DEVICE); in i40e_xmit_xdp_ring() 3522 tx_bi->xdpf = xdpf; in i40e_xmit_xdp_ring() [all …]
|
D | i40e_txrx.h | 285 struct xdp_frame *xdpf; member
|
/Linux-v4.19/drivers/net/ethernet/intel/ixgbe/ |
D | ixgbe_main.c | 1149 xdp_return_frame(tx_buffer->xdpf); in ixgbe_clean_tx_irq() 2195 struct xdp_frame *xdpf); 2203 struct xdp_frame *xdpf; in ixgbe_run_xdp() local 2219 xdpf = convert_to_xdp_frame(xdp); in ixgbe_run_xdp() 2220 if (unlikely(!xdpf)) { in ixgbe_run_xdp() 2224 result = ixgbe_xmit_xdp_ring(adapter, xdpf); in ixgbe_run_xdp() 5893 xdp_return_frame(tx_buffer->xdpf); in ixgbe_clean_tx_ring() 8427 struct xdp_frame *xdpf) in ixgbe_xmit_xdp_ring() argument 8436 len = xdpf->len; in ixgbe_xmit_xdp_ring() 8441 dma = dma_map_single(ring->dev, xdpf->data, len, DMA_TO_DEVICE); in ixgbe_xmit_xdp_ring() [all …]
|
D | ixgbe.h | 218 struct xdp_frame *xdpf; member
|
/Linux-v4.19/drivers/vhost/ |
D | net.c | 194 struct xdp_frame *xdpf = tun_ptr_to_xdp(ptr); in vhost_net_buf_peek_len() local 196 return xdpf->len; in vhost_net_buf_peek_len()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en.h | 405 struct xdp_frame *xdpf; member
|