Home
last modified time | relevance | path

Searched refs:segs (Results 1 – 25 of 97) sorted by relevance

1234

/Linux-v5.15/drivers/net/ethernet/intel/ice/
Dice_flow.c629 ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) in ice_flow_val_hdrs() argument
635 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK && in ice_flow_val_hdrs()
636 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK)) in ice_flow_val_hdrs()
640 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK && in ice_flow_val_hdrs()
641 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)) in ice_flow_val_hdrs()
669 sz = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_VLAN) ? in ice_flow_calc_seg_sz()
673 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) in ice_flow_calc_seg_sz()
675 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6) in ice_flow_calc_seg_sz()
677 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_ARP) in ice_flow_calc_seg_sz()
679 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK) in ice_flow_calc_seg_sz()
[all …]
/Linux-v5.15/net/sctp/
Doffload.c41 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local
67 segs = NULL; in sctp_gso_segment()
71 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment()
72 if (IS_ERR(segs)) in sctp_gso_segment()
77 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment()
86 return segs; in sctp_gso_segment()
/Linux-v5.15/net/ipv4/
Dudp_offload.c22 struct sk_buff *segs = ERR_PTR(-EINVAL); in __skb_udp_tunnel_segment() local
85 segs = gso_inner_segment(skb, features); in __skb_udp_tunnel_segment()
86 if (IS_ERR_OR_NULL(segs)) { in __skb_udp_tunnel_segment()
92 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment()
96 skb = segs; in __skb_udp_tunnel_segment()
148 return segs; in __skb_udp_tunnel_segment()
158 struct sk_buff *segs = ERR_PTR(-EINVAL); in skb_udp_tunnel_segment() local
180 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment, in skb_udp_tunnel_segment()
186 return segs; in skb_udp_tunnel_segment()
217 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) in __udpv4_gso_segment_list_csum() argument
[all …]
Dtcp_offload.c57 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local
89 segs = NULL; in tcp_gso_segment()
98 segs = skb_segment(skb, features); in tcp_gso_segment()
99 if (IS_ERR(segs)) in tcp_gso_segment()
103 segs->ooo_okay = ooo_okay; in tcp_gso_segment()
109 if (skb_is_gso(segs)) in tcp_gso_segment()
110 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment()
114 skb = segs; in tcp_gso_segment()
119 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment()
177 return segs; in tcp_gso_segment()
Dgre_offload.c19 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local
56 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment()
57 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment()
63 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment()
67 skb = segs; in gre_gso_segment()
117 return segs; in gre_gso_segment()
Desp4_offload.c121 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment() local
127 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment()
129 return segs; in xfrm4_transport_gso_segment()
137 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment() local
165 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment()
167 return segs; in xfrm4_beet_gso_segment()
/Linux-v5.15/net/mpls/
Dmpls_gso.c22 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local
47 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment()
48 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment()
53 skb = segs; in mpls_gso_segment()
69 return segs; in mpls_gso_segment()
/Linux-v5.15/net/nsh/
Dnsh.c79 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local
108 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment()
109 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment()
116 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment()
125 return segs; in nsh_gso_segment()
/Linux-v5.15/net/ipv6/
Dip6_offload.c77 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local
101 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment()
116 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment()
119 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment()
122 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment()
124 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment()
139 kfree_skb_list(segs); in ipv6_gso_segment()
154 return segs; in ipv6_gso_segment()
Dudp_offload.c20 struct sk_buff *segs = ERR_PTR(-EINVAL); in udp6_ufo_fragment() local
33 segs = skb_udp_tunnel_segment(skb, features, true); in udp6_ufo_fragment()
107 segs = skb_segment(skb, features); in udp6_ufo_fragment()
111 return segs; in udp6_ufo_fragment()
Desp6_offload.c156 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_transport_gso_segment() local
162 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_transport_gso_segment()
164 return segs; in xfrm6_transport_gso_segment()
172 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_beet_gso_segment() local
204 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_beet_gso_segment()
206 return segs; in xfrm6_beet_gso_segment()
/Linux-v5.15/net/sched/
Dsch_tbf.c194 struct sk_buff *segs, *nskb; in tbf_segment() local
199 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment()
201 if (IS_ERR_OR_NULL(segs)) in tbf_segment()
205 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment()
206 skb_mark_not_on_list(segs); in tbf_segment()
207 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment()
208 len += segs->len; in tbf_segment()
209 ret = qdisc_enqueue(segs, q->qdisc, to_free); in tbf_segment()
Dsch_netem.c415 struct sk_buff *segs; in netem_segment() local
418 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in netem_segment()
420 if (IS_ERR_OR_NULL(segs)) { in netem_segment()
425 return segs; in netem_segment()
441 struct sk_buff *segs = NULL; in netem_enqueue() local
499 segs = skb->next; in netem_enqueue()
522 skb->next = segs; in netem_enqueue()
595 if (segs) { in netem_enqueue()
602 while (segs) { in netem_enqueue()
603 skb2 = segs->next; in netem_enqueue()
[all …]
Dsch_cake.c1355 u16 segs = 1; in cake_overhead() local
1384 segs = DIV_ROUND_UP(skb->len - hdr_len, in cake_overhead()
1387 segs = shinfo->gso_segs; in cake_overhead()
1390 last_len = skb->len - shinfo->gso_size * (segs - 1); in cake_overhead()
1392 return (cake_calc_overhead(q, len, off) * (segs - 1) + in cake_overhead()
1742 struct sk_buff *segs, *nskb; in cake_enqueue() local
1746 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in cake_enqueue()
1747 if (IS_ERR_OR_NULL(segs)) in cake_enqueue()
1750 skb_list_walk_safe(segs, segs, nskb) { in cake_enqueue()
1751 skb_mark_not_on_list(segs); in cake_enqueue()
[all …]
/Linux-v5.15/drivers/infiniband/sw/rdmavt/
Dmr.c379 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr()
380 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr()
569 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page()
570 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page()
599 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; in rvt_map_mr_sg()
787 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok()
788 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok()
797 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok()
798 isge->length = mr->map[m]->segs[n].length - off; in rvt_lkey_ok()
894 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok()
[all …]
/Linux-v5.15/include/rdma/
Drdmavt_mr.h28 struct rvt_seg segs[RVT_SEGSZ]; member
133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge()
134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
/Linux-v5.15/fs/
Dbinfmt_elf_fdpic.c773 seg = loadmap->segs; in elf_fdpic_map_file()
794 seg = loadmap->segs; in elf_fdpic_map_file()
819 seg = loadmap->segs; in elf_fdpic_map_file()
840 seg = loadmap->segs; in elf_fdpic_map_file()
876 mseg = loadmap->segs; in elf_fdpic_map_file()
904 seg = loadmap->segs; in elf_fdpic_map_file()
935 seg = params->loadmap->segs; in elf_fdpic_map_file_constdisp_on_uclinux()
1022 seg = params->loadmap->segs; in elf_fdpic_map_file_by_direct_mmap()
1237 static inline void fill_elf_fdpic_header(struct elfhdr *elf, int segs) in fill_elf_fdpic_header() argument
1255 elf->e_phnum = segs; in fill_elf_fdpic_header()
[all …]
/Linux-v5.15/net/xfrm/
Dxfrm_output.c619 struct sk_buff *segs, *nskb; in xfrm_output_gso() local
623 segs = skb_gso_segment(skb, 0); in xfrm_output_gso()
625 if (IS_ERR(segs)) in xfrm_output_gso()
626 return PTR_ERR(segs); in xfrm_output_gso()
627 if (segs == NULL) in xfrm_output_gso()
630 skb_list_walk_safe(segs, segs, nskb) { in xfrm_output_gso()
633 skb_mark_not_on_list(segs); in xfrm_output_gso()
634 err = xfrm_output2(net, sk, segs); in xfrm_output_gso()
Dxfrm_device.c138 struct sk_buff *segs; in validate_xmit_xfrm() local
143 segs = skb_gso_segment(skb, esp_features); in validate_xmit_xfrm()
144 if (IS_ERR(segs)) { in validate_xmit_xfrm()
150 skb = segs; in validate_xmit_xfrm()
/Linux-v5.15/drivers/net/wireguard/
Ddevice.c158 struct sk_buff *segs = skb_gso_segment(skb, 0); in wg_xmit() local
160 if (IS_ERR(segs)) { in wg_xmit()
161 ret = PTR_ERR(segs); in wg_xmit()
165 skb = segs; in wg_xmit()
/Linux-v5.15/include/uapi/linux/
Delf-fdpic.h30 struct elf32_fdpic_loadseg segs[]; member
/Linux-v5.15/drivers/net/ethernet/marvell/octeontx2/nic/
Dotx2_txrx.c199 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg()
265 if (cqe->sg.segs) in otx2_check_rcv_errors()
298 for (seg = 0; seg < sg->segs; seg++, seg_addr++) { in otx2_rcv_pkt_handler()
482 sg->segs = 0; in otx2_sqe_add_sg()
498 sg->segs++; in otx2_sqe_add_sg()
709 sg->segs = 0; in otx2_sqe_tso_add_sg()
722 sg->segs++; in otx2_sqe_tso_add_sg()
943 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes()
Dotx2_struct.h153 u64 segs : 2; member
228 u64 segs : 2; member
/Linux-v5.15/include/net/
Dudp.h485 struct sk_buff *segs; in udp_rcv_segment() local
507 segs = __skb_gso_segment(skb, features, false); in udp_rcv_segment()
508 if (IS_ERR_OR_NULL(segs)) { in udp_rcv_segment()
518 return segs; in udp_rcv_segment()
/Linux-v5.15/arch/powerpc/platforms/powernv/
Dpci-ioda.c1520 unsigned int tce32_segsz, base, segs, avail, i; in pnv_pci_ioda1_setup_dma_pe() local
1533 segs = (weight * phb->ioda.dma32_count) / total_weight; in pnv_pci_ioda1_setup_dma_pe()
1534 if (!segs) in pnv_pci_ioda1_setup_dma_pe()
1535 segs = 1; in pnv_pci_ioda1_setup_dma_pe()
1544 for (base = 0; base <= phb->ioda.dma32_count - segs; base++) { in pnv_pci_ioda1_setup_dma_pe()
1545 for (avail = 0, i = base; i < base + segs; i++) { in pnv_pci_ioda1_setup_dma_pe()
1551 if (avail == segs) in pnv_pci_ioda1_setup_dma_pe()
1554 } while (--segs); in pnv_pci_ioda1_setup_dma_pe()
1556 if (!segs) { in pnv_pci_ioda1_setup_dma_pe()
1572 weight, total_weight, base, segs); in pnv_pci_ioda1_setup_dma_pe()
[all …]

1234