Home
last modified time | relevance | path

Searched refs:segs (Results 1 – 25 of 115) sorted by relevance

12345

/linux/net/core/
A Dnet_test.c52 const unsigned int *segs; member
61 .segs = (const unsigned int[]) { GSO_TEST_SIZE },
94 .segs = (const unsigned int[]) { 2 * GSO_TEST_SIZE, 3 },
136 .segs = (const unsigned int[]) { 100, 200, 300, 400 },
150 struct sk_buff *skb, *segs, *cur, *next, *last; in gso_test_func() local
229 segs = skb_segment(skb, features); in gso_test_func()
230 if (IS_ERR(segs)) { in gso_test_func()
231 KUNIT_FAIL(test, "segs error %pe", segs); in gso_test_func()
233 } else if (!segs) { in gso_test_func()
238 last = segs->prev; in gso_test_func()
[all …]
A Dgso.c16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_eth_gso_segment() local
22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment()
28 return segs; in skb_eth_gso_segment()
40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_mac_gso_segment() local
53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment()
61 return segs; in skb_mac_gso_segment()
91 struct sk_buff *segs; in __skb_gso_segment() local
124 segs = skb_mac_gso_segment(skb, features); in __skb_gso_segment()
126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs))) in __skb_gso_segment()
129 return segs; in __skb_gso_segment()
A Ddev.h112 unsigned int segs) in netif_set_gso_max_segs() argument
115 WRITE_ONCE(dev->gso_max_segs, segs); in netif_set_gso_max_segs()
/linux/net/sctp/
A Doffload.c42 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local
68 segs = NULL; in sctp_gso_segment()
72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment()
73 if (IS_ERR(segs)) in sctp_gso_segment()
78 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment()
87 return segs; in sctp_gso_segment()
/linux/drivers/net/ethernet/intel/ice/
A Dice_flow.c1302 if (segs[i].hdrs != p->segs[i].hdrs || in ice_flow_find_prof_conds()
1304 segs[i].match != p->segs[i].match)) in ice_flow_find_prof_conds()
1415 memcpy(&params->prof->segs[i], &segs[i], sizeof(*segs)); in ice_flow_add_prof_sync()
1656 if (!segs) in ice_flow_add_prof()
1998 seg = &segs[seg_cnt - 1]; in ice_flow_set_rss_seg_info()
2348 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_add_rss_cfg_sync()
2349 if (!segs) in ice_add_rss_cfg_sync()
2426 kfree(segs); in ice_add_rss_cfg_sync()
2495 segs = kcalloc(segs_cnt, sizeof(*segs), GFP_KERNEL); in ice_rem_rss_cfg_sync()
2496 if (!segs) in ice_rem_rss_cfg_sync()
[all …]
/linux/net/ipv4/
A Dudp_offload.c98 skb = segs; in __skb_udp_tunnel_segment()
150 return segs; in __skb_udp_tunnel_segment()
188 return segs; in skb_udp_tunnel_segment()
225 seg = segs; in __udpv4_gso_segment_list_csum()
233 return segs; in __udpv4_gso_segment_list_csum()
247 return segs; in __udpv4_gso_segment_list_csum()
331 return segs; in __udp_gso_segment()
338 if (skb_is_gso(segs)) in __udp_gso_segment()
341 seg = segs; in __udp_gso_segment()
409 return segs; in __udp_gso_segment()
[all …]
A Dtcp_offload.c60 seg = segs; in __tcpv4_gso_segment_list_csum()
68 return segs; in __tcpv4_gso_segment_list_csum()
82 return segs; in __tcpv4_gso_segment_list_csum()
167 segs = NULL; in tcp_gso_segment()
176 segs = skb_segment(skb, features); in tcp_gso_segment()
177 if (IS_ERR(segs)) in tcp_gso_segment()
181 segs->ooo_okay = ooo_okay; in tcp_gso_segment()
187 if (skb_is_gso(segs)) in tcp_gso_segment()
188 mss *= skb_shinfo(segs)->gso_segs; in tcp_gso_segment()
192 skb = segs; in tcp_gso_segment()
[all …]
A Dgre_offload.c21 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local
58 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment()
59 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment()
65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment()
69 skb = segs; in gre_gso_segment()
119 return segs; in gre_gso_segment()
A Desp4_offload.c136 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment() local
142 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment()
144 return segs; in xfrm4_transport_gso_segment()
152 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment() local
183 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment()
185 return segs; in xfrm4_beet_gso_segment()
/linux/net/mpls/
A Dmpls_gso.c23 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local
51 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment()
52 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment()
57 skb = segs; in mpls_gso_segment()
73 return segs; in mpls_gso_segment()
/linux/net/nsh/
A Dnsh.c81 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local
110 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment()
111 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment()
117 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment()
126 return segs; in nsh_gso_segment()
/linux/tools/lib/bpf/
A Dusdt.c427 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); in parse_elf_segs()
431 *segs = tmp; in parse_elf_segs()
432 seg = *segs + *seg_cnt; in parse_elf_segs()
446 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs); in parse_elf_segs()
510 tmp = libbpf_reallocarray(*segs, *seg_cnt + 1, sizeof(**segs)); in parse_vma_segs()
516 *segs = tmp; in parse_vma_segs()
517 seg = *segs + *seg_cnt; in parse_vma_segs()
533 qsort(*segs, *seg_cnt, sizeof(**segs), cmp_elf_segs); in parse_vma_segs()
583 struct elf_seg *segs = NULL, *vma_segs = NULL; in collect_usdt_targets() local
670 seg = find_elf_seg(segs, seg_cnt, usdt_abs_ip); in collect_usdt_targets()
[all …]
/linux/net/ipv6/
A Dip6_offload.c110 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local
137 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment()
152 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment()
153 if (!segs) in ipv6_gso_segment()
157 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment()
160 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment()
162 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment()
177 kfree_skb_list(segs); in ipv6_gso_segment()
192 return segs; in ipv6_gso_segment()
A Dtcpv6_offload.c109 static struct sk_buff *__tcpv6_gso_segment_list_csum(struct sk_buff *segs) in __tcpv6_gso_segment_list_csum() argument
117 seg = segs; in __tcpv6_gso_segment_list_csum()
126 return segs; in __tcpv6_gso_segment_list_csum()
138 return segs; in __tcpv6_gso_segment_list_csum()
A Dudp_offload.c22 struct sk_buff *segs = ERR_PTR(-EINVAL); in udp6_ufo_fragment() local
35 segs = skb_udp_tunnel_segment(skb, features, true); in udp6_ufo_fragment()
109 segs = skb_segment(skb, features); in udp6_ufo_fragment()
113 return segs; in udp6_ufo_fragment()
A Desp6_offload.c172 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_transport_gso_segment() local
178 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_transport_gso_segment()
180 return segs; in xfrm6_transport_gso_segment()
188 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_beet_gso_segment() local
223 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_beet_gso_segment()
225 return segs; in xfrm6_beet_gso_segment()
/linux/net/sched/
A Dsch_tbf.c209 struct sk_buff *segs, *nskb; in tbf_segment() local
214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment()
216 if (IS_ERR_OR_NULL(segs)) in tbf_segment()
220 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment()
221 skb_mark_not_on_list(segs); in tbf_segment()
222 qdisc_skb_cb(segs)->pkt_len = segs->len; in tbf_segment()
223 len += segs->len; in tbf_segment()
224 ret = qdisc_enqueue(segs, q->qdisc, to_free); in tbf_segment()
A Dsch_netem.c424 struct sk_buff *segs; in netem_segment() local
429 if (IS_ERR_OR_NULL(segs)) { in netem_segment()
434 return segs; in netem_segment()
499 segs = skb->next; in netem_enqueue()
522 skb->next = segs; in netem_enqueue()
615 if (segs) { in netem_enqueue()
622 while (segs) { in netem_enqueue()
623 skb2 = segs->next; in netem_enqueue()
625 qdisc_skb_cb(segs)->pkt_len = segs->len; in netem_enqueue()
626 last_len = segs->len; in netem_enqueue()
[all …]
/linux/drivers/infiniband/sw/rdmavt/
A Dmr.c379 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr()
380 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr()
569 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page()
570 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page()
787 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok()
788 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok()
797 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok()
894 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok()
895 off -= mr->map[m]->segs[n].length; in rvt_rkey_ok()
904 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_rkey_ok()
[all …]
/linux/include/uapi/linux/
A Delf-fdpic.h30 struct elf32_fdpic_loadseg segs[]; member
45 struct elf64_fdpic_loadseg segs[]; member
/linux/include/rdma/
A Drdmavt_mr.h28 struct rvt_seg segs[RVT_SEGSZ]; member
133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge()
134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
/linux/fs/
A Dbinfmt_elf_fdpic.c783 seg = loadmap->segs; in elf_fdpic_map_file()
808 seg = loadmap->segs; in elf_fdpic_map_file()
829 seg = loadmap->segs; in elf_fdpic_map_file()
865 mseg = loadmap->segs; in elf_fdpic_map_file()
893 seg = loadmap->segs; in elf_fdpic_map_file()
926 seg = params->loadmap->segs; in elf_fdpic_map_file_constdisp_on_uclinux()
1013 seg = params->loadmap->segs; in elf_fdpic_map_file_by_direct_mmap()
1248 elf->e_phnum = segs; in fill_elf_fdpic_header()
1461 int segs; in elf_fdpic_core_dump() local
1507 segs++; in elf_fdpic_core_dump()
[all …]
/linux/net/xfrm/
A Dxfrm_output.c620 struct sk_buff *segs, *nskb; in xfrm_output_gso() local
624 segs = skb_gso_segment(skb, 0); in xfrm_output_gso()
626 if (IS_ERR(segs)) in xfrm_output_gso()
627 return PTR_ERR(segs); in xfrm_output_gso()
628 if (segs == NULL) in xfrm_output_gso()
631 skb_list_walk_safe(segs, segs, nskb) { in xfrm_output_gso()
634 skb_mark_not_on_list(segs); in xfrm_output_gso()
635 err = xfrm_output2(net, sk, segs); in xfrm_output_gso()
/linux/drivers/net/wireguard/
A Ddevice.c179 struct sk_buff *segs = skb_gso_segment(skb, 0); in wg_xmit() local
181 if (IS_ERR(segs)) { in wg_xmit()
182 ret = PTR_ERR(segs); in wg_xmit()
186 skb = segs; in wg_xmit()
/linux/include/net/
A Dudp.h463 struct sk_buff *segs; in udp_rcv_segment() local
485 segs = __skb_gso_segment(skb, features, false); in udp_rcv_segment()
486 if (IS_ERR_OR_NULL(segs)) { in udp_rcv_segment()
496 return segs; in udp_rcv_segment()

Completed in 63 milliseconds

12345