| /linux/net/core/ |
| H A D | net_test.c | 52 const unsigned int *segs; member 61 .segs = (const unsigned int[]) { GSO_TEST_SIZE }, 68 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 }, 77 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 1 }, 85 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, 2 }, 94 .segs = (const unsigned int[]) { 2 * GSO_TEST_SIZE, 3 }, 104 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE }, 112 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE }, 122 .segs = (const unsigned int[]) { GSO_TEST_SIZE, GSO_TEST_SIZE, GSO_TEST_SIZE, 3 }, 136 .segs = (const unsigned int[]) { 100, 200, 300, 400 }, [all …]
|
| H A D | gso.c | 16 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_eth_gso_segment() local 22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment() 28 return segs; in skb_eth_gso_segment() 40 struct sk_buff *segs = ERR_PTR(-EPROTONOSUPPORT); in skb_mac_gso_segment() local 53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment() 61 return segs; in skb_mac_gso_segment() 91 struct sk_buff *segs; in __skb_gso_segment() local 124 segs = skb_mac_gso_segment(skb, features); in __skb_gso_segment() 126 if (segs != skb && unlikely(skb_needs_check(skb, tx_path) && !IS_ERR(segs))) in __skb_gso_segment() 129 return segs; in __skb_gso_segment()
|
| /linux/net/sctp/ |
| H A D | offload.c | 42 struct sk_buff *segs = ERR_PTR(-EINVAL); in sctp_gso_segment() local 68 segs = NULL; in sctp_gso_segment() 72 segs = skb_segment(skb, (features | NETIF_F_HW_CSUM) & ~NETIF_F_SG); in sctp_gso_segment() 73 if (IS_ERR(segs)) in sctp_gso_segment() 78 for (skb = segs; skb; skb = skb->next) { in sctp_gso_segment() 87 return segs; in sctp_gso_segment()
|
| /linux/net/mpls/ |
| H A D | mpls_gso.c | 23 struct sk_buff *segs = ERR_PTR(-EINVAL); in mpls_gso_segment() local 51 segs = skb_mac_gso_segment(skb, mpls_features); in mpls_gso_segment() 52 if (IS_ERR_OR_NULL(segs)) { in mpls_gso_segment() 57 skb = segs; in mpls_gso_segment() 73 return segs; in mpls_gso_segment()
|
| /linux/net/nsh/ |
| H A D | nsh.c | 81 struct sk_buff *segs = ERR_PTR(-EINVAL); in nsh_gso_segment() local 110 segs = skb_mac_gso_segment(skb, features); in nsh_gso_segment() 111 if (IS_ERR_OR_NULL(segs)) { in nsh_gso_segment() 117 for (skb = segs; skb; skb = skb->next) { in nsh_gso_segment() 126 return segs; in nsh_gso_segment()
|
| /linux/net/ipv4/ |
| H A D | gre_offload.c | 21 struct sk_buff *segs = ERR_PTR(-EINVAL); in gre_gso_segment() local 58 segs = skb_mac_gso_segment(skb, features); in gre_gso_segment() 59 if (IS_ERR_OR_NULL(segs)) { in gre_gso_segment() 65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 69 skb = segs; in gre_gso_segment() 119 return segs; in gre_gso_segment()
|
| H A D | esp4_offload.c | 138 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_transport_gso_segment() local 144 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_transport_gso_segment() 146 return segs; in xfrm4_transport_gso_segment() 154 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm4_beet_gso_segment() local 185 segs = ops->callbacks.gso_segment(skb, features); in xfrm4_beet_gso_segment() 187 return segs; in xfrm4_beet_gso_segment()
|
| /linux/net/ipv6/ |
| H A D | ip6_offload.c | 110 struct sk_buff *segs = ERR_PTR(-EINVAL); in ipv6_gso_segment() local 137 segs = ERR_PTR(-EPROTONOSUPPORT); in ipv6_gso_segment() 154 segs = ops->callbacks.gso_segment(skb, features); in ipv6_gso_segment() 155 if (!segs) in ipv6_gso_segment() 159 if (IS_ERR_OR_NULL(segs)) in ipv6_gso_segment() 162 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 164 for (skb = segs; skb; skb = skb->next) { in ipv6_gso_segment() 179 kfree_skb_list(segs); in ipv6_gso_segment() 194 return segs; in ipv6_gso_segment()
|
| H A D | esp6_offload.c | 174 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_transport_gso_segment() local 180 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_transport_gso_segment() 182 return segs; in xfrm6_transport_gso_segment() 190 struct sk_buff *segs = ERR_PTR(-EINVAL); in xfrm6_beet_gso_segment() local 225 segs = ops->callbacks.gso_segment(skb, features); in xfrm6_beet_gso_segment() 227 return segs; in xfrm6_beet_gso_segment()
|
| /linux/net/sched/ |
| H A D | sch_tbf.c | 209 struct sk_buff *segs, *nskb; in tbf_segment() local 214 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in tbf_segment() 216 if (IS_ERR_OR_NULL(segs)) in tbf_segment() 220 skb_list_walk_safe(segs, segs, nskb) { in tbf_segment() 221 skb_mark_not_on_list(segs); in tbf_segment() 222 seg_len = segs->len; in tbf_segment() 223 qdisc_skb_cb(segs)->pkt_len = seg_len; in tbf_segment() 224 qdisc_skb_cb(segs)->pkt_segs = 1; in tbf_segment() 225 ret = qdisc_enqueue(segs, in tbf_segment() [all...] |
| H A D | sch_netem.c | 429 struct sk_buff *segs; in netem_segment() local 433 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in netem_segment() 435 if (IS_ERR_OR_NULL(segs)) { in netem_segment() 440 return segs; in netem_segment() 456 struct sk_buff *segs = NULL; in netem_enqueue() 505 segs = skb->next; in netem_enqueue() 527 /* re-link segs, so that qdisc_drop_all() frees them all */ in netem_enqueue() 528 skb->next = segs; in netem_enqueue() 621 if (segs) { in netem_enqueue() 628 while (segs) { in netem_enqueue() 455 struct sk_buff *segs = NULL; netem_enqueue() local [all...] |
| H A D | sch_cake.c | 1401 u16 segs = qdisc_pkt_segs(skb); in cake_overhead() 1406 if (segs == 1) in cake_overhead() 1434 last_len = skb->len - shinfo->gso_size * (segs - 1); in cake_overhead() 1436 return (cake_calc_overhead(q, len, off) * (segs - 1) + in cake_overhead() 1785 struct sk_buff *segs, *nskb; in cake_enqueue() 1789 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in cake_enqueue() 1790 if (IS_ERR_OR_NULL(segs)) in cake_enqueue() 1793 skb_list_walk_safe(segs, segs, nskb) { in cake_enqueue() 1794 skb_mark_not_on_list(segs); in cake_enqueue() 1402 u16 segs = 1; cake_overhead() local 1792 struct sk_buff *segs, *nskb; cake_enqueue() local [all...] |
| H A D | sch_taprio.c | 588 struct sk_buff *segs, *nskb; in taprio_enqueue_segmented() local 591 segs = skb_gso_segment(skb, features & ~NETIF_F_GSO_MASK); in taprio_enqueue_segmented() 592 if (IS_ERR_OR_NULL(segs)) in taprio_enqueue_segmented() 595 skb_list_walk_safe(segs, segs, nskb) { in taprio_enqueue_segmented() 596 skb_mark_not_on_list(segs); in taprio_enqueue_segmented() 597 qdisc_skb_cb(segs)->pkt_len = segs->len; in taprio_enqueue_segmented() 598 qdisc_skb_cb(segs)->pkt_segs = 1; in taprio_enqueue_segmented() 599 slen += segs in taprio_enqueue_segmented() [all...] |
| /linux/include/uapi/linux/ |
| H A D | elf-fdpic.h | 30 struct elf32_fdpic_loadseg segs[]; member 45 struct elf64_fdpic_loadseg segs[]; member
|
| /linux/fs/ |
| H A D | binfmt_elf_fdpic.c | 764 loadmap = kzalloc(struct_size(loadmap, segs, nloads), GFP_KERNEL); in elf_fdpic_map_file() 792 seg = loadmap->segs; in elf_fdpic_map_file() 817 seg = loadmap->segs; in elf_fdpic_map_file() 838 seg = loadmap->segs; in elf_fdpic_map_file() 874 mseg = loadmap->segs; in elf_fdpic_map_file() 902 seg = loadmap->segs; in elf_fdpic_map_file() 935 seg = params->loadmap->segs; in elf_fdpic_map_file_constdisp_on_uclinux() 1022 seg = params->loadmap->segs; in elf_fdpic_map_file_by_direct_mmap() 1240 static inline void fill_elf_fdpic_header(struct elfhdr *elf, int segs) in fill_elf_fdpic_header() argument 1258 elf->e_phnum = segs; in fill_elf_fdpic_header() [all …]
|
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt_coredump.c | 98 info->segs = le16_to_cpu(*((__le16 *)(resp + in bnxt_hwrm_dbg_dma_data() 100 if (!info->segs) { in bnxt_hwrm_dbg_dma_data() 105 info->dest_buf_size = info->segs * in bnxt_hwrm_dbg_dma_data() 169 coredump->total_segs = info.segs; in bnxt_hwrm_dbg_coredump_list() 349 u32 *segs) in bnxt_get_ctx_coredump() argument 359 *segs = 0; in bnxt_get_ctx_coredump() 398 *segs += 1; in bnxt_get_ctx_coredump() 436 u32 drv_len, segs = 0; in __bnxt_get_coredump() local 438 drv_len = bnxt_get_ctx_coredump(bp, buf, offset, &segs); in __bnxt_get_coredump() 442 coredump.total_segs += segs; in __bnxt_get_coredump()
|
| /linux/include/rdma/ |
| H A D | rdmavt_mr.h | 28 struct rvt_seg segs[RVT_SEGSZ]; member 133 sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr; in rvt_update_sge() 134 sge->length = sge->mr->map[sge->m]->segs[sge->n].length; in rvt_update_sge()
|
| /linux/tools/testing/selftests/net/ |
| H A D | lwt_dst_cache_ref_loop.sh | 63 encap rpl segs 2001:db8:3::1 dev veth0 &>/dev/null 66 encap seg6 mode inline segs 2001:db8:4::1 dev veth0 &>/dev/null 186 encap rpl segs 2001:db8:2::2 \ 208 encap seg6 mode inline segs 2001:db8:2::2 \
|
| /linux/net/xfrm/ |
| H A D | xfrm_output.c | 656 struct sk_buff *segs, *nskb; in xfrm_output_gso() local 660 segs = skb_gso_segment(skb, 0); in xfrm_output_gso() 662 if (IS_ERR(segs)) in xfrm_output_gso() 663 return PTR_ERR(segs); in xfrm_output_gso() 664 if (segs == NULL) in xfrm_output_gso() 667 skb_list_walk_safe(segs, segs, nskb) { in xfrm_output_gso() 670 skb_mark_not_on_list(segs); in xfrm_output_gso() 671 err = xfrm_output2(net, sk, segs); in xfrm_output_gso()
|
| H A D | xfrm_device.c | 160 struct sk_buff *segs; in validate_xmit_xfrm() local 165 segs = skb_gso_segment(skb, esp_features); in validate_xmit_xfrm() 166 if (IS_ERR(segs)) { in validate_xmit_xfrm() 172 skb = segs; in validate_xmit_xfrm()
|
| /linux/arch/um/drivers/ |
| H A D | ubd_kern.c | 1262 int segs = 0; in ubd_submit_request() local 1268 segs = 0; in ubd_submit_request() 1270 segs = 1; in ubd_submit_request() 1272 segs = blk_rq_nr_phys_segments(req); in ubd_submit_request() 1274 io_req = ubd_alloc_req(dev, req, segs); in ubd_submit_request() 1278 io_req->desc_cnt = segs; in ubd_submit_request() 1279 if (segs) in ubd_submit_request()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_struct.h | 153 u64 segs : 2; member 228 u64 segs : 2; member
|
| H A D | otx2_txrx.c | 259 for (seg = 0; seg < sg->segs; seg++, seg_addr++) in otx2_free_rcv_seg() 325 if (cqe->sg.segs) in otx2_check_rcv_errors() 368 for (seg = 0; seg < sg->segs; seg++, seg_addr++) { in otx2_rcv_pkt_handler() 677 sg->segs = 0; in otx2_sqe_add_sg() 693 sg->segs++; in otx2_sqe_add_sg() 911 sg->segs = 0; in otx2_sqe_tso_add_sg() 924 sg->segs++; in otx2_sqe_tso_add_sg() 1304 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes() 1426 sg->segs = 1; in otx2_xdp_sqe_add_sg()
|
| /linux/arch/mips/include/asm/octeon/ |
| H A D | cvmx-pko.h | 241 uint64_t segs:6; member 246 uint64_t segs:6;
|
| /linux/fs/f2fs/ |
| H A D | gc.c | 2190 int segs = secs * SEGS_PER_SEC(sbi); in update_sb_metadata() local 2200 raw_sb->segment_count = cpu_to_le32(segment_count + segs); in update_sb_metadata() 2201 raw_sb->segment_count_main = cpu_to_le32(segment_count_main + segs); in update_sb_metadata() 2203 (long long)SEGS_TO_BLKS(sbi, segs)); in update_sb_metadata() 2210 cpu_to_le32(dev_segs + segs); in update_sb_metadata() 2218 int segs = secs * SEGS_PER_SEC(sbi); in update_fs_metadata() local 2219 long long blks = SEGS_TO_BLKS(sbi, segs); in update_fs_metadata() 2223 SM_I(sbi)->segment_count = (int)SM_I(sbi)->segment_count + segs; in update_fs_metadata() 2224 MAIN_SEGS(sbi) = (int)MAIN_SEGS(sbi) + segs; in update_fs_metadata() 2229 FREE_I(sbi)->free_segments = (int)FREE_I(sbi)->free_segments + segs; in update_fs_metadata() [all …]
|