Home
last modified time | relevance | path

Searched refs:segs (Results 1 – 25 of 95) sorted by relevance

1234

/linux/net/core/
H A Dnet_test.c52 const unsigned int *segs; member
61 .segs = (const unsigned int[]) { GSO_TEST_SIZE },
68 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 },
77 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 },
85 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 },
94 .segs = (const unsigned int[]) { 2 * GSO_TEST_SIZE, 3 },
104 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE },
112 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE },
122 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE, 3 },
136 .segs = (const unsigned int[]) { 100, 200, 300, 400 },
[all …]
H A Dgso.c16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_eth_gso_segment() local
22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment()
28 return segs; in skb_eth_gso_segment()
40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_mac_gso_segment() local
53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment()
61 return segs; in skb_mac_gso_segment()
91 struct sk_buff *segs; in __skb_gso_segment() local
124 segs = skb_mac_gso_segment(skb, features); in __skb_gso_segment()
126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs))) in __skb_gso_segment()
129 return segs; in __skb_gso_segment()
/linux/drivers/net/ethernet/intel/ice/
H A Dice_flow.c733 static int ice_flow_val_hdrs(struct ice_flow_seg_info *segs, u8 segs_cnt) in ice_flow_val_hdrs() argument
739 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK && in ice_flow_val_hdrs()
740 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L3_MASK)) in ice_flow_val_hdrs()
744 if (segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK && in ice_flow_val_hdrs()
745 !is_power_of_2(segs[i].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)) in ice_flow_val_hdrs()
773 sz = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_VLAN) ? in ice_flow_calc_seg_sz()
777 if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) in ice_flow_calc_seg_sz()
779 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6) in ice_flow_calc_seg_sz()
781 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_ARP) in ice_flow_calc_seg_sz()
783 else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK) in ice_flow_calc_seg_sz()
[all …]
/linux/net/sctp/
H A Doffload.c42 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local
68 segs = NULL; in sctp_gso_segment()
72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment()
73 if (IS_ERR(segs)) in sctp_gso_segment()
78 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment()
87 return segs; in sctp_gso_segment()
/linux/net/ipv4/
H A Dudp_offload.c180 struct sk_buff *segs = ERR_PTR(-EINVAL); in __skb_udp_tunnel_segment() local
243 segs = gso_inner_segment(skb, features); in __skb_udp_tunnel_segment()
244 if (IS_ERR_OR_NULL(segs)) { in __skb_udp_tunnel_segment()
250 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment()
254 skb = segs; in __skb_udp_tunnel_segment()
306 return segs; in __skb_udp_tunnel_segment()
316 struct sk_buff *segs = ERR_PTR(-EINVAL); in skb_udp_tunnel_segment() local
338 segs = __skb_udp_tunnel_segment(skb, features, gso_inner_segment, in skb_udp_tunnel_segment()
344 return segs; in skb_udp_tunnel_segment()
375 static struct sk_buff *__udpv4_gso_segment_list_csum(struct sk_buff *segs) in __udpv4_gso_segment_list_csum() argument
430 __udpv6_gso_segment_list_csum(struct sk_buff * segs) __udpv6_gso_segment_list_csum() argument
485 struct sk_buff *segs, *seg; __udp_gso_segment() local
641 struct sk_buff *segs = ERR_PTR(-EINVAL); udp4_ufo_fragment() local
[all...]
H A Dtcp_offload.c55 static struct sk_buff *__tcpv4_gso_segment_list_csum(struct sk_buff *segs) in __tcpv4_gso_segment_list_csum() argument
63 seg = segs; in __tcpv4_gso_segment_list_csum()
71 return segs; in __tcpv4_gso_segment_list_csum()
85 return segs; in __tcpv4_gso_segment_list_csum()
136 struct sk_buff *segs = ERR_PTR(-EINVAL); in tcp_gso_segment() local
172 segs = NULL; in tcp_gso_segment()
181 segs = skb_segment(skb, features); in tcp_gso_segment()
182 if (IS_ERR(segs)) in tcp_gso_segment()
186 segs->ooo_okay = ooo_okay; in tcp_gso_segment()
192 if (skb_is_gso(segs)) in tcp_gso_segment()
[all...]
H A Dgre_offload.c21 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local
58 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment()
59 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment()
65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment()
69 skb = segs; in gre_gso_segment()
119 return segs; in gre_gso_segment()
H A Desp4_offload.c138 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment() local
144 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment()
146 return segs; in xfrm4_transport_gso_segment()
154 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment() local
185 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment()
187 return segs; in xfrm4_beet_gso_segment()
/linux/net/mpls/
H A Dmpls_gso.c23 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local
51 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment()
52 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment()
57 skb = segs; in mpls_gso_segment()
73 return segs; in mpls_gso_segment()
/linux/net/nsh/
H A Dnsh.c81 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local
110 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment()
111 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment()
117 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment()
126 return segs; in nsh_gso_segment()
/linux/net/ipv6/
H A Dip6_offload.c94 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_pull_exthdrs()
118 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment()
135 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment()
136 if (!segs) in ipv6_gso_segment()
140 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment()
143 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment()
145 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment()
160 kfree_skb_list(segs); in ipv6_gso_segment()
175 return segs; in ipv6_gso_segment()
110 struct sk_buff *segs = ERR_PTR(-EINVAL); ipv6_gso_segment() local
H A Dtcpv6_offload.c115 static struct sk_buff *__tcpv6_gso_segment_list_csum(struct sk_buff *segs) in __tcpv6_gso_segment_csum()
123 seg = segs; in __tcpv6_gso_segment_list_csum()
132 return segs; in __tcpv6_gso_segment_list_csum()
144 return segs; in __tcpv6_gso_segment_list_csum()
117 __tcpv6_gso_segment_list_csum(struct sk_buff * segs) __tcpv6_gso_segment_list_csum() argument
H A Desp6_offload.c174 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_transport_gso_segment() local
180 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_transport_gso_segment()
182 return segs; in xfrm6_transport_gso_segment()
190 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_beet_gso_segment() local
225 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_beet_gso_segment()
227 return segs; in xfrm6_beet_gso_segment()
/linux/drivers/infiniband/sw/rdmavt/
H A Dmr.c384 mr->mr.map[m]->segs[n].vaddr = vaddr; in rvt_reg_user_mr()
385 mr->mr.map[m]->segs[n].length = PAGE_SIZE; in rvt_reg_user_mr()
574 mr->mr.map[m]->segs[n].vaddr = (void *)addr; in rvt_set_page()
575 mr->mr.map[m]->segs[n].length = ps; in rvt_set_page()
604 mr->mr.offset = ibmr->iova - (u64)mr->mr.map[0]->segs[0].vaddr; in rvt_map_mr_sg()
792 while (off >= mr->map[m]->segs[n].length) { in rvt_lkey_ok()
793 off -= mr->map[m]->segs[n].length; in rvt_lkey_ok()
802 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in rvt_lkey_ok()
803 isge->length = mr->map[m]->segs[n].length - off; in rvt_lkey_ok()
899 while (off >= mr->map[m]->segs[n].length) { in rvt_rkey_ok()
[all …]
/linux/net/sched/
H A Dsch_tbf.c209 struct sk_buff *segs, *nskb; in tbf_segment() local
214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment()
216 if (IS_ERR_OR_NULL(segs)) in tbf_segment()
220 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment()
221 skb_mark_not_on_list(segs); in tbf_segment()
222 seg_len = segs->len; in tbf_segment()
223 qdisc_skb_cb(segs)->pkt_len = seg_len; in tbf_segment()
224 qdisc_skb_cb(segs)->pkt_segs = 1; in tbf_segment()
225 ret = qdisc_enqueue(segs, in tbf_segment()
[all...]
H A Dsch_netem.c429 struct sk_buff *segs; in netem_segment() local
433 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in netem_segment()
435 if (IS_ERR_OR_NULL(segs)) { in netem_segment()
440 return segs; in netem_segment()
456 struct sk_buff *segs = NULL; in netem_enqueue() local
505 segs = skb->next; in netem_enqueue()
528 skb->next = segs; in netem_enqueue()
621 if (segs) { in netem_enqueue()
628 while (segs) { in netem_enqueue()
629 skb2 = segs->next; in netem_enqueue()
[all …]
H A Dsch_cake.c1416 u16 segs = qdisc_pkt_segs(skb); in cake_overhead() local
1421 if (segs == 1) in cake_overhead()
1449 last_len = skb->len - shinfo->gso_size * (segs - 1); in cake_overhead()
1451 return (cake_calc_overhead(q, len, off) * (segs - 1) + in cake_overhead()
1801 struct sk_buff *segs, *nskb; in cake_enqueue() local
1805 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in cake_enqueue()
1806 if (IS_ERR_OR_NULL(segs)) in cake_enqueue()
1809 skb_list_walk_safe(segs, segs, nskb) { in cake_enqueue()
1810 skb_mark_not_on_list(segs); in cake_enqueue()
1811 qdisc_skb_cb(segs)->pkt_len = segs->len; in cake_enqueue()
[all …]
/linux/fs/
H A Dbinfmt_elf_fdpic.c764 loadmap = kzalloc_flex(*loadmap, segs, nloads); in elf_fdpic_map_file()
792 seg = loadmap->segs; in elf_fdpic_map_file()
817 seg = loadmap->segs; in elf_fdpic_map_file()
838 seg = loadmap->segs; in elf_fdpic_map_file()
874 mseg = loadmap->segs; in elf_fdpic_map_file()
902 seg = loadmap->segs; in elf_fdpic_map_file()
935 seg = params->loadmap->segs; in elf_fdpic_map_file_constdisp_on_uclinux()
1022 seg = params->loadmap->segs; in elf_fdpic_map_file_by_direct_mmap()
1240 static inline void fill_elf_fdpic_header(struct elfhdr *elf, int segs) in fill_elf_fdpic_header() argument
1258 elf->e_phnum = segs; in fill_elf_fdpic_header()
[all …]
H A Dbinfmt_elf.c1437 static void fill_elf_header(struct elfhdr *elf, int segs, in fill_elf_header() argument
1455 elf->e_phnum = segs; in fill_elf_header()
1971 elf_addr_t e_shoff, int segs) in fill_extnum_info() argument
1983 shdr4extnum->sh_info = segs; in fill_extnum_info()
1996 int segs, i; in elf_core_dump() local
2009 segs = cprm->vma_count + elf_core_extra_phdrs(cprm); in elf_core_dump()
2012 segs++; in elf_core_dump()
2017 e_phnum = segs > PN_XNUM ? PN_XNUM : segs; in elf_core_dump()
2029 offset += segs * sizeof(struct elf_phdr); /* Program headers */ in elf_core_dump()
2056 fill_extnum_info(&elf, shdr4extnum, e_shoff, segs); in elf_core_dump()
/linux/include/uapi/linux/
H A Delf-fdpic.h30 struct elf32_fdpic_loadseg segs[]; member
45 struct elf64_fdpic_loadseg segs[]; member
/linux/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt_coredump.c98 info->segs = le16_to_cpu(*((__le16 *)(resp + in bnxt_hwrm_dbg_dma_data()
100 if (!info->segs) { in bnxt_hwrm_dbg_dma_data()
105 info->dest_buf_size = info->segs * in bnxt_hwrm_dbg_dma_data()
169 coredump->total_segs = info.segs; in bnxt_hwrm_dbg_coredump_list()
349 u32 *segs) in bnxt_get_ctx_coredump() argument
359 *segs = 0; in bnxt_get_ctx_coredump()
398 *segs += 1; in bnxt_get_ctx_coredump()
436 u32 drv_len, segs = 0; in __bnxt_get_coredump() local
438 drv_len = bnxt_get_ctx_coredump(bp, buf, offset, &segs); in __bnxt_get_coredump()
442 coredump.total_segs += segs; in __bnxt_get_coredump()
/linux/include/rdma/
H A Drdmavt_mr.h28 struct rvt_seg segs[RVT_SEGSZ]; member
133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge()
134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
/linux/tools/testing/selftests/net/
H A Dlwt_dst_cache_ref_loop.sh63 encap rpl segs 2001:db8:3::1 dev veth0 &>/dev/null
66 encap seg6 mode inline segs 2001:db8:4::1 dev veth0 &>/dev/null
186 encap rpl segs 2001:db8:2::2 \
208 encap seg6 mode inline segs 2001:db8:2::2 \
/linux/net/xfrm/
H A Dxfrm_output.c656 struct sk_buff *segs, *nskb; in xfrm_output_gso() local
660 segs = skb_gso_segment(skb, 0); in xfrm_output_gso()
662 if (IS_ERR(segs)) in xfrm_output_gso()
663 return PTR_ERR(segs); in xfrm_output_gso()
664 if (segs == NULL) in xfrm_output_gso()
667 skb_list_walk_safe(segs, segs, nskb) { in xfrm_output_gso()
670 skb_mark_not_on_list(segs); in xfrm_output_gso()
671 err = xfrm_output2(net, sk, segs); in xfrm_output_gso()
H A Dxfrm_device.c160 struct sk_buff *segs; in validate_xmit_xfrm() local
165 segs = skb_gso_segment(skb, esp_features); in validate_xmit_xfrm()
166 if (IS_ERR(segs)) { in validate_xmit_xfrm()
172 skb = segs; in validate_xmit_xfrm()

1234