Home
last modified time | relevance | path

Searched refs:skb_shinfo (Results 1 – 25 of 302) sorted by relevance

12345678910>>...13

/Linux-v5.4/net/core/
Dskbuff.c237 shinfo = skb_shinfo(skb); in __alloc_skb()
280 shinfo = skb_shinfo(skb); in __build_skb_around()
556 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag()
573 skb_drop_list(&skb_shinfo(skb)->frag_list); in skb_drop_fraglist()
596 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_release_data()
721 struct skb_shared_info *sh = skb_shinfo(skb); in skb_dump()
782 for (i = 0; len && i < skb_shinfo(skb)->nr_frags; i++) { in skb_dump()
783 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump()
1006 atomic_inc(&(skb_shinfo(skb)->dataref)); in __skb_clone()
1029 skb_shinfo(n)->frag_list = first; in alloc_skb_for_msg()
[all …]
Dtso.c12 return skb_shinfo(skb)->gso_segs * 2 + skb_shinfo(skb)->nr_frags; in tso_count_descs()
54 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_build_data()
55 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data()
78 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_start()
79 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
/Linux-v5.4/net/ipv4/
Dtcp_offload.c19 skb_shinfo(skb)->tx_flags |= SKBTX_SW_TSTAMP; in tcp_gso_tstamp()
20 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp()
32 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4)) in tcp4_gso_segment()
80 mss = skb_shinfo(skb)->gso_size; in tcp_gso_segment()
87 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in tcp_gso_segment()
110 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment()
118 if (unlikely(skb_shinfo(gso_skb)->tx_flags & SKBTX_SW_TSTAMP)) in tcp_gso_segment()
119 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment()
258 mss = skb_shinfo(p)->gso_size; in tcp_gro_receive()
296 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in tcp_gro_complete()
[all …]
Dudp_offload.c40 if (skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) in __skb_udp_tunnel_segment()
55 need_csum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM); in __skb_udp_tunnel_segment()
58 remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM); in __skb_udp_tunnel_segment()
89 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment()
121 uh->len = htons(skb_shinfo(skb)->gso_size + in __skb_udp_tunnel_segment()
199 mss = skb_shinfo(gso_skb)->gso_size; in __udp_gso_segment()
222 mss *= skb_shinfo(segs)->gso_segs; in __udp_gso_segment()
228 skb_shinfo(seg)->tskey = skb_shinfo(gso_skb)->tskey; in __udp_gso_segment()
229 skb_shinfo(seg)->tx_flags |= in __udp_gso_segment()
230 (skb_shinfo(gso_skb)->tx_flags & SKBTX_ANY_TSTAMP); in __udp_gso_segment()
[all …]
Dgre_offload.c43 need_csum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_GRE_CSUM); in gre_gso_segment()
56 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment()
94 skb_shinfo(skb)->gso_size; in gre_gso_segment()
236 skb_shinfo(skb)->gso_type = SKB_GSO_GRE; in gre_gro_complete()
Dinet_fragment.c457 skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; in inet_frag_reasm_prepare()
459 for (i = 0; i < skb_shinfo(head)->nr_frags; i++) in inet_frag_reasm_prepare()
460 plen += skb_frag_size(&skb_shinfo(head)->frags[i]); in inet_frag_reasm_prepare()
467 skb_shinfo(head)->frag_list = clone; in inet_frag_reasm_prepare()
470 nextp = &skb_shinfo(head)->frag_list; in inet_frag_reasm_prepare()
/Linux-v5.4/drivers/net/xen-netback/
Dnetback.c355 skb_shinfo(skb)->destructor_arg = NULL; in xenvif_alloc_skb()
367 struct skb_shared_info *shinfo = skb_shinfo(skb); in xenvif_get_requests()
389 shinfo = skb_shinfo(nskb); in xenvif_get_requests()
402 skb_shinfo(skb)->frag_list = nskb; in xenvif_get_requests()
445 struct skb_shared_info *shinfo = skb_shinfo(skb); in xenvif_tx_check_gop()
549 first_shinfo = skb_shinfo(skb); in xenvif_tx_check_gop()
550 shinfo = skb_shinfo(skb_shinfo(skb)->frag_list); in xenvif_tx_check_gop()
562 struct skb_shared_info *shinfo = skb_shinfo(skb); in xenvif_fill_frags()
577 skb_shinfo(skb)->destructor_arg = in xenvif_fill_frags()
644 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in xenvif_set_skb_gso()
[all …]
/Linux-v5.4/include/linux/
Dvirtio_net.h84 skb_shinfo(skb)->gso_size = gso_size; in virtio_net_hdr_to_skb()
85 skb_shinfo(skb)->gso_type = gso_type; in virtio_net_hdr_to_skb()
88 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; in virtio_net_hdr_to_skb()
89 skb_shinfo(skb)->gso_segs = 0; in virtio_net_hdr_to_skb()
104 struct skb_shared_info *sinfo = skb_shinfo(skb); in virtio_net_hdr_from_skb()
Dudp.h126 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) { in udp_cmsg_recv()
127 gso_size = skb_shinfo(skb)->gso_size; in udp_cmsg_recv()
135 skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4; in udp_unexpected_gso()
Dskbuff.h483 #define skb_uarg(SKB) ((struct ubuf_info *)(skb_shinfo(SKB)->destructor_arg))
1403 #define skb_shinfo(SKB) ((struct skb_shared_info *)(skb_end_pointer(SKB))) macro
1407 return &skb_shinfo(skb)->hwtstamps; in skb_hwtstamps()
1412 bool is_zcopy = skb && skb_shinfo(skb)->tx_flags & SKBTX_DEV_ZEROCOPY; in skb_zcopy()
1425 skb_shinfo(skb)->destructor_arg = uarg; in skb_zcopy_set()
1426 skb_shinfo(skb)->tx_flags |= SKBTX_ZEROCOPY_FRAG; in skb_zcopy_set()
1432 skb_shinfo(skb)->destructor_arg = (void *)((uintptr_t) val | 0x1UL); in skb_zcopy_set_nouarg()
1433 skb_shinfo(skb)->tx_flags |= SKBTX_ZEROCOPY_FRAG; in skb_zcopy_set_nouarg()
1438 return (uintptr_t) skb_shinfo(skb)->destructor_arg & 0x1UL; in skb_zcopy_is_nouarg()
1443 return (void *)((uintptr_t) skb_shinfo(skb)->destructor_arg & ~0x1UL); in skb_zcopy_get_nouarg()
[all …]
/Linux-v5.4/net/ipv6/
Dip6_offload.c103 skb_shinfo(skb)->gso_type & (SKB_GSO_IPXIP4 | SKB_GSO_IPXIP6)) in ipv6_gso_segment()
105 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP); in ipv6_gso_segment()
108 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP); in ipv6_gso_segment()
119 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment()
124 payload_len = skb_shinfo(skb)->gso_size + in ipv6_gso_segment()
355 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP4; in sit_gro_complete()
362 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP6; in ip6ip6_gro_complete()
369 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP6; in ip4ip6_gro_complete()
385 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_IPXIP4)) in sit_gso_segment()
394 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_IPXIP6)) in ip4ip6_gso_segment()
[all …]
Dudp_offload.c31 mss = skb_shinfo(skb)->gso_size; in udp6_ufo_fragment()
35 if (skb->encapsulation && skb_shinfo(skb)->gso_type & in udp6_ufo_fragment()
42 if (!(skb_shinfo(skb)->gso_type & (SKB_GSO_UDP | SKB_GSO_UDP_L4))) in udp6_ufo_fragment()
48 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in udp6_ufo_fragment()
Dtcpv6_offload.c36 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV6; in tcp6_gro_complete()
46 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6)) in tcp6_gso_segment()
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/
Den_tx.c177 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in mlx5e_tx_get_gso_ihs()
212 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlx5e_txwqe_build_dsegs()
213 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mlx5e_txwqe_build_dsegs()
254 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in mlx5e_txwqe_complete()
255 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in mlx5e_txwqe_complete()
290 mss = cpu_to_be16(skb_shinfo(skb)->gso_size); in mlx5e_sq_xmit()
292 num_bytes = skb->len + (skb_shinfo(skb)->gso_segs - 1) * ihs; in mlx5e_sq_xmit()
293 stats->packets += skb_shinfo(skb)->gso_segs; in mlx5e_sq_xmit()
309 ds_cnt += skb_shinfo(skb)->nr_frags; in mlx5e_sq_xmit()
490 if (unlikely(skb_shinfo(skb)->tx_flags & in mlx5e_poll_tx_cq()
[all …]
/Linux-v5.4/drivers/net/ethernet/sfc/
Dtx_tso.c291 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size; in tso_start_new_packet()
295 st->packet_space = skb_shinfo(skb)->gso_size; in tso_start_new_packet()
341 st->seqnum += skb_shinfo(skb)->gso_size; in tso_start_new_packet()
386 EFX_WARN_ON_ONCE_PARANOID(skb_shinfo(skb)->nr_frags < 1); in efx_enqueue_skb_tso()
389 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()
408 if (++frag_i >= skb_shinfo(skb)->nr_frags) in efx_enqueue_skb_tso()
412 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()
/Linux-v5.4/include/trace/events/
Dnet.h55 __entry->tx_flags = skb_shinfo(skb)->tx_flags;
56 __entry->gso_size = skb_shinfo(skb)->gso_size;
57 __entry->gso_segs = skb_shinfo(skb)->gso_segs;
58 __entry->gso_type = skb_shinfo(skb)->gso_type;
213 __entry->nr_frags = skb_shinfo(skb)->nr_frags;
214 __entry->gso_size = skb_shinfo(skb)->gso_size;
215 __entry->gso_type = skb_shinfo(skb)->gso_type;
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
Dtls_rxtx.c168 skb_shinfo(nskb)->gso_size = 0; in mlx5e_tls_complete_sync_skb()
170 skb_shinfo(nskb)->gso_size = mss; in mlx5e_tls_complete_sync_skb()
171 skb_shinfo(nskb)->gso_segs = DIV_ROUND_UP(data_len, mss); in mlx5e_tls_complete_sync_skb()
173 skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; in mlx5e_tls_complete_sync_skb()
242 skb_shinfo(nskb)->frags[i] = info.frags[i]; in mlx5e_tls_handle_ooo()
244 skb_shinfo(nskb)->nr_frags = info.nr_frags; in mlx5e_tls_handle_ooo()
Den_accel.h98 int payload_len = skb_shinfo(skb)->gso_size + sizeof(struct udphdr); in mlx5e_udp_gso_handle_tx_skb()
126 if (skb_is_gso(skb) && skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in mlx5e_accel_handle_tx()
/Linux-v5.4/drivers/net/ethernet/chelsio/cxgb4/
Dcxgb4_ptp.h57 return skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP; in cxgb4_xmit_with_hwtstamp()
62 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in cxgb4_xmit_hwtstamp_pending()
/Linux-v5.4/drivers/staging/octeon/
Dethernet-tx.c191 if (unlikely(skb_shinfo(skb)->nr_frags > 5)) { in cvm_oct_xmit()
263 if (skb_shinfo(skb)->nr_frags == 0) { in cvm_oct_xmit()
272 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in cvm_oct_xmit()
273 skb_frag_t *fs = skb_shinfo(skb)->frags + i; in cvm_oct_xmit()
282 hw_buffer.s.size = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit()
283 pko_command.s.segs = skb_shinfo(skb)->nr_frags + 1; in cvm_oct_xmit()
325 if (unlikely(skb_shinfo(skb)->nr_frags)) { in cvm_oct_xmit()
/Linux-v5.4/drivers/net/ethernet/qlogic/qede/
Dqede_fp.c136 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++, bds_consumed++) { in qede_free_tx_pkt()
216 unsigned short gso_type = skb_shinfo(skb)->gso_type; in qede_xmit_type()
310 return (skb_shinfo(skb)->nr_frags > allowed_frags); in qede_pkt_req_lin()
633 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in qede_set_gro_params()
635 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in qede_set_gro_params()
637 skb_shinfo(skb)->gso_size = __le16_to_cpu(cqe->len_on_first_bd) - in qede_set_gro_params()
902 skb_shinfo(skb)->gso_type = 0; in qede_gro_receive()
903 skb_shinfo(skb)->gso_size = 0; in qede_gro_receive()
908 if (skb_shinfo(skb)->gso_size) { in qede_gro_receive()
1163 skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags++, in qede_rx_build_jumbo()
[all …]
/Linux-v5.4/net/sctp/
Dinqueue.c125 chunk->skb = skb_shinfo(chunk->skb)->frag_list; in sctp_inq_pop()
162 if (skb_shinfo(chunk->skb)->frag_list) in sctp_inq_pop()
167 chunk->skb = skb_shinfo(chunk->skb)->frag_list; in sctp_inq_pop()
/Linux-v5.4/drivers/net/
Dxen-netfront.c316 page = skb_frag_page(&skb_shinfo(skb)->frags[0]); in xennet_alloc_rx_buffers()
525 int i, frags = skb_shinfo(skb)->nr_frags; in xennet_count_skb_slots()
532 skb_frag_t *frag = skb_shinfo(skb)->frags + i; in xennet_count_skb_slots()
652 if (skb_shinfo(skb)->gso_size) { in xennet_start_xmit()
660 gso->u.gso.size = skb_shinfo(skb)->gso_size; in xennet_start_xmit()
661 gso->u.gso.type = (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) ? in xennet_start_xmit()
675 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xennet_start_xmit()
676 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in xennet_start_xmit()
877 skb_shinfo(skb)->gso_size = gso->u.gso.size; in xennet_set_skb_gso()
878 skb_shinfo(skb)->gso_type = in xennet_set_skb_gso()
[all …]
/Linux-v5.4/drivers/net/ethernet/sun/
Dsunvnet_common.c1083 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_map()
1084 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_map()
1123 docopy = skb_shinfo(skb)->nr_frags >= ncookies; in vnet_skb_shape()
1124 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_shape()
1125 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_shape()
1209 skb_shinfo(nskb)->gso_size = skb_shinfo(skb)->gso_size; in vnet_skb_shape()
1210 skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; in vnet_skb_shape()
1249 gso_size = skb_shinfo(skb)->gso_size; in vnet_handle_offloads()
1250 gso_type = skb_shinfo(skb)->gso_type; in vnet_handle_offloads()
1251 gso_segs = skb_shinfo(skb)->gso_segs; in vnet_handle_offloads()
[all …]
/Linux-v5.4/net/xfrm/
Dxfrm_ipcomp.c72 if (WARN_ON(skb_shinfo(skb)->nr_frags >= MAX_SKB_FRAGS)) in ipcomp_decompress()
75 frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags; in ipcomp_decompress()
96 skb_shinfo(skb)->nr_frags++; in ipcomp_decompress()

12345678910>>...13