/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
D | ice_flow.c | 225 ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) in ice_flow_val_hdrs() argument 231 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK && in ice_flow_val_hdrs() 232 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK)) in ice_flow_val_hdrs() 236 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK && in ice_flow_val_hdrs() 237 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)) in ice_flow_val_hdrs() 262 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) in ice_flow_calc_seg_sz() 264 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6) in ice_flow_calc_seg_sz() 268 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_TCP) in ice_flow_calc_seg_sz() 270 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_UDP) in ice_flow_calc_seg_sz() 272 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_SCTP) in ice_flow_calc_seg_sz() [all …]
|
/kernel/linux/linux-5.10/net/ipv4/ |
D | udp_offload.c | 22 struct sk_buff *segs = ERR_PTR(-EINVAL); in __skb_udp_tunnel_segment() local 82 segs = gso_inner_segment(skb, features); in __skb_udp_tunnel_segment() 83 if (IS_ERR_OR_NULL(segs)) { in __skb_udp_tunnel_segment() 89 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment() 93 skb = segs; in __skb_udp_tunnel_segment() 145 return segs; in __skb_udp_tunnel_segment() 155 struct sk_buff *segs = ERR_PTR(-EINVAL); in skb_udp_tunnel_segment() local 177 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment, in skb_udp_tunnel_segment() 183 return segs; in skb_udp_tunnel_segment() 214 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) in __udpv4_gso_segment_list_csum() argument [all …]
|
D | tcp_offload.c | 57 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local 89 segs = NULL; in tcp_gso_segment() 98 segs = skb_segment(skb, features); in tcp_gso_segment() 99 if (IS_ERR(segs)) in tcp_gso_segment() 103 segs->ooo_okay = ooo_okay; in tcp_gso_segment() 109 if (skb_is_gso(segs)) in tcp_gso_segment() 110 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment() 114 skb = segs; in tcp_gso_segment() 119 tcp_gso_tstamp(segs, skb_shinfo(gso_skb)->tskey, seq, mss); in tcp_gso_segment() 177 return segs; in tcp_gso_segment()
|
D | gre_offload.c | 19 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local 50 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment() 51 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment() 57 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 61 skb = segs; in gre_gso_segment() 113 return segs; in gre_gso_segment()
|
D | esp4_offload.c | 122 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment() local 128 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment() 130 return segs; in xfrm4_transport_gso_segment() 138 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment() local 169 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment() 171 return segs; in xfrm4_beet_gso_segment()
|
/kernel/linux/linux-5.10/net/sctp/ |
D | offload.c | 37 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local 63 segs = NULL; in sctp_gso_segment() 67 segs = skb_segment(skb, features | NETIF_F_HW_CSUM | NETIF_F_SG); in sctp_gso_segment() 68 if (IS_ERR(segs)) in sctp_gso_segment() 73 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment() 82 return segs; in sctp_gso_segment()
|
/kernel/linux/linux-5.10/net/mpls/ |
D | mpls_gso.c | 22 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local 47 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment() 48 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment() 53 skb = segs; in mpls_gso_segment() 69 return segs; in mpls_gso_segment()
|
/kernel/linux/linux-5.10/net/nsh/ |
D | nsh.c | 79 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local 107 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment() 108 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment() 114 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment() 123 return segs; in nsh_gso_segment()
|
/kernel/linux/linux-5.10/net/ipv6/ |
D | ip6_offload.c | 76 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local 100 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment() 115 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment() 116 if (!segs) in ipv6_gso_segment() 120 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment() 123 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 125 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment() 140 kfree_skb_list(segs); in ipv6_gso_segment() 155 return segs; in ipv6_gso_segment()
|
D | udp_offload.c | 20 struct sk_buff *segs = ERR_PTR(-EINVAL); in udp6_ufo_fragment() local 37 segs = skb_udp_tunnel_segment(skb, features, true); in udp6_ufo_fragment() 107 segs = skb_segment(skb, features); in udp6_ufo_fragment() 111 return segs; in udp6_ufo_fragment()
|
D | esp6_offload.c | 156 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_transport_gso_segment() local 162 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_transport_gso_segment() 164 return segs; in xfrm6_transport_gso_segment() 172 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_beet_gso_segment() local 207 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_beet_gso_segment() 209 return segs; in xfrm6_beet_gso_segment()
|
/kernel/linux/linux-5.10/net/sched/ |
D | sch_tbf.c | 194 struct sk_buff *segs, *nskb; in tbf_segment() local 199 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment() 201 if (IS_ERR_OR_NULL(segs)) in tbf_segment() 205 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment() 206 skb_mark_not_on_list(segs); in tbf_segment() 207 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment() 208 len += segs->len; in tbf_segment() 209 ret = qdisc_enqueue(segs, q->qdisc, to_free); in tbf_segment()
|
D | sch_netem.c | 415 struct sk_buff *segs; in netem_segment() local 418 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in netem_segment() 420 if (IS_ERR_OR_NULL(segs)) { in netem_segment() 425 return segs; in netem_segment() 441 struct sk_buff *segs = NULL; in netem_enqueue() local 499 segs = skb->next; in netem_enqueue() 522 skb->next = segs; in netem_enqueue() 595 if (segs) { in netem_enqueue() 602 while (segs) { in netem_enqueue() 603 skb2 = segs->next; in netem_enqueue() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
D | mr.c | 420 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr() 421 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr() 610 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page() 611 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page() 640 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; in rvt_map_mr_sg() 828 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok() 829 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok() 838 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok() 839 isge->length = mr->map[m]->segs[n].length - off; in rvt_lkey_ok() 935 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok() [all …]
|
/kernel/linux/linux-5.10/include/rdma/ |
D | rdmavt_mr.h | 28 struct rvt_seg segs[RVT_SEGSZ]; member 133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge() 134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
|
/kernel/linux/linux-5.10/fs/ |
D | binfmt_elf_fdpic.c | 771 seg = loadmap->segs; in elf_fdpic_map_file() 792 seg = loadmap->segs; in elf_fdpic_map_file() 817 seg = loadmap->segs; in elf_fdpic_map_file() 838 seg = loadmap->segs; in elf_fdpic_map_file() 874 mseg = loadmap->segs; in elf_fdpic_map_file() 902 seg = loadmap->segs; in elf_fdpic_map_file() 933 seg = params->loadmap->segs; in elf_fdpic_map_file_constdisp_on_uclinux() 1024 seg = params->loadmap->segs; in elf_fdpic_map_file_by_direct_mmap() 1253 static inline void fill_elf_fdpic_header(struct elfhdr *elf, int segs) in fill_elf_fdpic_header() argument 1271 elf->e_phnum = segs; in fill_elf_fdpic_header() [all …]
|
/kernel/linux/linux-5.10/net/xfrm/ |
D | xfrm_output.c | 542 struct sk_buff *segs, *nskb; in xfrm_output_gso() local 546 segs = skb_gso_segment(skb, 0); in xfrm_output_gso() 548 if (IS_ERR(segs)) in xfrm_output_gso() 549 return PTR_ERR(segs); in xfrm_output_gso() 550 if (segs == NULL) in xfrm_output_gso() 553 skb_list_walk_safe(segs, segs, nskb) { in xfrm_output_gso() 556 skb_mark_not_on_list(segs); in xfrm_output_gso() 557 err = xfrm_output2(net, sk, segs); in xfrm_output_gso()
|
D | xfrm_device.c | 151 struct sk_buff *segs; in validate_xmit_xfrm() local 156 segs = skb_gso_segment(skb, esp_features); in validate_xmit_xfrm() 157 if (IS_ERR(segs)) { in validate_xmit_xfrm() 163 skb = segs; in validate_xmit_xfrm()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | otx2_txrx.c | 201 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg() 258 if (cqe->sg.segs == 1) in otx2_check_rcv_errors() 263 if (cqe->sg.segs == 1 && (pfvf->netdev->features & NETIF_F_RXALL)) in otx2_check_rcv_errors() 267 if (cqe->sg.segs) in otx2_check_rcv_errors() 280 if (unlikely(parse->errlev || parse->errcode || cqe->sg.segs > 1)) { in otx2_rcv_pkt_handler() 475 sg->segs = 0; in otx2_sqe_add_sg() 491 sg->segs++; in otx2_sqe_add_sg() 687 sg->segs = 0; in otx2_sqe_tso_add_sg() 700 sg->segs++; in otx2_sqe_tso_add_sg() 916 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes()
|
/kernel/linux/common_modules/newip/third_party/linux-5.10/include/net/ |
D | tcp_nip.h | 107 int __tcp_nip_retransmit_skb(struct sock *sk, struct sk_buff *skb, int segs); 108 int tcp_nip_retransmit_skb(struct sock *sk, struct sk_buff *skb, int segs);
|
/kernel/linux/patches/linux-5.10/prebuilts/usr/include/linux/ |
D | elf-fdpic.h | 31 struct elf32_fdpic_loadseg segs[]; member
|
/kernel/linux/patches/linux-4.19/prebuilts/usr/include/linux/ |
D | elf-fdpic.h | 18 struct elf32_fdpic_loadseg segs[]; member
|
/kernel/linux/linux-5.10/include/uapi/linux/ |
D | elf-fdpic.h | 30 struct elf32_fdpic_loadseg segs[]; member
|
/kernel/linux/linux-5.10/drivers/net/wireguard/ |
D | device.c | 162 struct sk_buff *segs = skb_gso_segment(skb, 0); in wg_xmit() local 164 if (unlikely(IS_ERR(segs))) { in wg_xmit() 165 ret = PTR_ERR(segs); in wg_xmit() 169 skb = segs; in wg_xmit()
|
/kernel/linux/linux-5.10/include/net/ |
D | udp.h | 481 struct sk_buff *segs; in udp_rcv_segment() local 503 segs = __skb_gso_segment(skb, features, false); in udp_rcv_segment() 504 if (IS_ERR_OR_NULL(segs)) { in udp_rcv_segment() 514 return segs; in udp_rcv_segment()
|