| /linux/net/ipv4/ |
| H A D | tcp_offload.c | 19 u32 flags = skb_shinfo(gso_skb)->tx_flags & SKBTX_ANY_TSTAMP; in tcp_gso_tstamp() 20 u32 ts_seq = skb_shinfo(gso_skb)->tskey; in tcp_gso_tstamp() 24 skb_shinfo(skb)->tx_flags |= flags; in tcp_gso_tstamp() 25 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 101 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4)) in tcp4_gso_segment() 107 if (skb_shinfo(skb)->gso_type & SKB_GSO_FRAGLIST) { in tcp4_gso_segment() 110 if ((skb_pagelen(skb) - th->doff * 4 == skb_shinfo(skb)->gso_size) && in tcp4_gso_segment() 111 !(skb_shinfo(skb)->gso_type & SKB_GSO_DODGY)) in tcp4_gso_segment() 163 mss = skb_shinfo(skb)->gso_size; in tcp_gso_segment() 170 skb_shinfo(skb)->gso_segs = DIV_ROUND_UP(skb->len, mss); in tcp_gso_segment() [all …]
|
| H A D | udp_offload.c | 198 if (skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) in __skb_udp_tunnel_segment() 214 need_csum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM); in __skb_udp_tunnel_segment() 217 remcsum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_TUNNEL_REMCSUM); in __skb_udp_tunnel_segment() 250 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in __skb_udp_tunnel_segment() 282 uh->len = htons(skb_shinfo(skb)->gso_size + in __skb_udp_tunnel_segment() 466 unsigned int mss = skb_shinfo(skb)->gso_size; in __udp_gso_segment_list() 493 mss = skb_shinfo(gso_skb)->gso_size; in __udp_gso_segment() 499 !(skb_shinfo(gso_skb)->gso_type & SKB_GSO_FRAGLIST))) in __udp_gso_segment() 510 skb_shinfo(gso_skb)->gso_segs = DIV_ROUND_UP(gso_skb->len - sizeof(*uh), in __udp_gso_segment() 515 if (skb_shinfo(gso_skb)->gso_type & SKB_GSO_FRAGLIST) { in __udp_gso_segment() [all …]
|
| H A D | gre_offload.c | 45 need_csum = !!(skb_shinfo(skb)->gso_type & SKB_GSO_GRE_CSUM); in gre_gso_segment() 65 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in gre_gso_segment() 103 skb_shinfo(skb)->gso_size; in gre_gso_segment() 245 skb_shinfo(skb)->gso_type = SKB_GSO_GRE; in gre_gro_complete()
|
| /linux/net/core/ |
| H A D | skbuff.c | 388 shinfo = skb_shinfo(skb); in __finalize_skb_around() 888 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_coalesce_rx_frag() 907 skb_drop_list(&skb_shinfo(skb)->frag_list); in skb_drop_fraglist() 1051 shinfo = skb_shinfo(skb); in skb_pp_frag_ref() 1086 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_release_data() 1296 struct skb_shared_info *sh = skb_shinfo(skb); in skb_dump() 1361 for (i = 0; len && i < skb_shinfo(skb)->nr_frags; i++) { in skb_dump() 1362 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in skb_dump() 1605 atomic_inc(&(skb_shinfo(skb)->dataref)); in __skb_clone() 1628 skb_shinfo(n)->frag_list = first; in alloc_skb_for_msg() [all …]
|
| H A D | tso.c | 53 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_build_data() 54 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_build_data() 79 (tso->next_frag_idx < skb_shinfo(skb)->nr_frags)) { in tso_start() 80 skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx]; in tso_start()
|
| H A D | gro.c | 94 struct skb_shared_info *pinfo, *skbinfo = skb_shinfo(skb); in skb_gro_receive() 126 pinfo = skb_shinfo(lp); in skb_gro_receive() 206 skb_shinfo(p)->frag_list = skb; in skb_gro_receive() 233 skb_shinfo(p)->frag_list = skb; in skb_gro_receive_list() 264 skb_shinfo(skb)->gso_size = 0; in gro_complete() 401 pinfo = skb_shinfo(skb); in skb_gro_reset_offset() 416 struct skb_shared_info *pinfo = skb_shinfo(skb); in gro_pull_from_frag0() 497 NAPI_GRO_CB(skb)->count = skb_shinfo(skb)->gso_segs; in dev_gro_receive() 500 (skb_shinfo(skb)->gso_type & SKB_GSO_DODGY)) in dev_gro_receive() 551 skb_shinfo(skb)->gso_size = skb_gro_len(skb); in dev_gro_receive() [all …]
|
| /linux/drivers/net/wireless/intel/iwlwifi/ |
| H A D | iwl-utils.c | 22 unsigned int mss = skb_shinfo(skb)->gso_size; in iwl_tx_tso_segment() 27 skb_shinfo(skb)->gso_size = num_subframes * mss; in iwl_tx_tso_segment() 31 skb_shinfo(skb)->gso_size = mss; in iwl_tx_tso_segment() 32 skb_shinfo(skb)->gso_type = ipv4 ? SKB_GSO_TCPV4 : SKB_GSO_TCPV6; in iwl_tx_tso_segment() 60 skb_shinfo(tmp)->gso_size = mss; in iwl_tx_tso_segment() 61 skb_shinfo(tmp)->gso_type = ipv4 ? SKB_GSO_TCPV4 : in iwl_tx_tso_segment() 73 skb_shinfo(tmp)->gso_size = 0; in iwl_tx_tso_segment()
|
| /linux/net/ipv6/ |
| H A D | tcpv6_offload.c | 79 skb_shinfo(skb)->gso_type |= SKB_GSO_FRAGLIST | SKB_GSO_TCPV6; in tcp6_gro_complete() 80 skb_shinfo(skb)->gso_segs = NAPI_GRO_CB(skb)->count; in tcp6_gro_complete() 89 skb_shinfo(skb)->gso_type |= SKB_GSO_TCPV6; in tcp6_gro_complete() 164 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6)) in tcp6_gso_segment() 170 if (skb_shinfo(skb)->gso_type & SKB_GSO_FRAGLIST) { in tcp6_gso_segment() 173 if ((skb_pagelen(skb) - th->doff * 4 == skb_shinfo(skb)->gso_size) && in tcp6_gso_segment() 174 !(skb_shinfo(skb)->gso_type & SKB_GSO_DODGY)) in tcp6_gso_segment()
|
| H A D | ip6_offload.c | 142 skb_shinfo(skb)->gso_type & (SKB_GSO_IPXIP4 | SKB_GSO_IPXIP6)) in ipv6_gso_segment() 144 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP); in ipv6_gso_segment() 147 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP); in ipv6_gso_segment() 162 gso_partial = !!(skb_shinfo(segs)->gso_type & SKB_GSO_PARTIAL); in ipv6_gso_segment() 167 payload_len = skb_shinfo(skb)->gso_size + in ipv6_gso_segment() 395 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP4; in sit_gro_complete() 402 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP6; in ip6ip6_gro_complete() 409 skb_shinfo(skb)->gso_type |= SKB_GSO_IPXIP6; in ip4ip6_gro_complete() 417 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_IPXIP4)) in sit_gso_segment() 426 if (!(skb_shinfo(skb)->gso_type & SKB_GSO_IPXIP6)) in ip4ip6_gso_segment() [all …]
|
| /linux/drivers/net/ethernet/sfc/siena/ |
| H A D | tx.h | 28 if (skb_shinfo(skb)->gso_segs > 1 && in efx_tx_csum_type_skb() 29 !(skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) && in efx_tx_csum_type_skb() 30 (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM)) in efx_tx_csum_type_skb()
|
| /linux/drivers/net/xen-netback/ |
| H A D | netback.c | 371 skb_shinfo(skb)->destructor_arg = NULL; in xenvif_alloc_skb() 387 struct skb_shared_info *shinfo = skb_shinfo(skb); in xenvif_get_requests() 486 shinfo = skb_shinfo(nskb); in xenvif_get_requests() 507 skb_shinfo(skb)->frag_list = nskb; in xenvif_get_requests() 561 struct skb_shared_info *shinfo = skb_shinfo(skb); in xenvif_tx_check_gop() 678 shinfo = skb_shinfo(shinfo->frag_list); in xenvif_tx_check_gop() 690 struct skb_shared_info *shinfo = skb_shinfo(skb); in xenvif_fill_frags() 705 skb_shinfo(skb)->destructor_arg = in xenvif_fill_frags() 772 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4; in xenvif_set_skb_gso() 775 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6; in xenvif_set_skb_gso() [all …]
|
| /linux/Documentation/translations/zh_CN/networking/ |
| H A D | timestamping.rst | 313 节。它存储相关的序列号在 skb_shinfo(skb)->tskey。因为一个 skbuff 只有一 528 - 在 hard_start_xmit() 中,检查 (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) 530 - 如果此 skb 和请求都可能,则声明驱动程序正在执行时间戳,通过设置 skb_shinfo(skb)->tx_flags 533 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; 652 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; 660 == true``")和当前 skb 是否需要 TX 时间戳("``skb_shinfo(skb)->tx_flags 661 & SKBTX_HW_TSTAMP``")。如果为真,则设置 "``skb_shinfo(skb)->tx_flags 667 型的错误所在:驱动程序采取捷径,只检查 "``skb_shinfo(skb)->tx_flags & 672 针对 "``skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS``",还针对
|
| /linux/include/linux/ |
| H A D | virtio_net.h | 160 struct skb_shared_info *shinfo = skb_shinfo(skb); in __virtio_net_hdr_to_skb() 219 struct skb_shared_info *sinfo = skb_shinfo(skb); in virtio_net_hdr_from_skb() 334 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL_CSUM; in virtio_net_hdr_tnl_to_skb() 336 skb_shinfo(skb)->gso_type |= SKB_GSO_UDP_TUNNEL; in virtio_net_hdr_tnl_to_skb() 395 tnl_gso_type = skb_shinfo(skb)->gso_type & (SKB_GSO_UDP_TUNNEL | in virtio_net_hdr_tnl_from_skb() 411 skb_shinfo(skb)->gso_type &= ~tnl_gso_type; in virtio_net_hdr_tnl_from_skb() 413 skb_shinfo(skb)->gso_type |= tnl_gso_type; in virtio_net_hdr_tnl_from_skb() 422 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL_CSUM) in virtio_net_hdr_tnl_from_skb()
|
| H A D | skbuff_ref.h | 32 __skb_frag_ref(&skb_shinfo(skb)->frags[f]); in skb_frag_ref() 68 struct skb_shared_info *shinfo = skb_shinfo(skb); in skb_frag_unref()
|
| H A D | skbuff.h | 574 #define skb_uarg(SKB) ((struct ubuf_info *)(skb_shinfo(SKB)->destructor_arg)) 1783 #define skb_shinfo(SKB) ((struct skb_shared_info *)(skb_end_pointer(SKB))) macro 1787 return &skb_shinfo(skb)->hwtstamps; in skb_hwtstamps() 1792 bool is_zcopy = skb && skb_shinfo(skb)->flags & SKBFL_ZEROCOPY_ENABLE; in skb_zcopy() 1799 return skb_shinfo(skb)->flags & SKBFL_PURE_ZEROCOPY; in skb_zcopy_pure() 1804 return skb_shinfo(skb)->flags & SKBFL_MANAGED_FRAG_REFS; in skb_zcopy_managed() 1820 skb_shinfo(skb)->destructor_arg = uarg; in skb_zcopy_init() 1821 skb_shinfo(skb)->flags |= uarg->flags; in skb_zcopy_init() 1838 skb_shinfo(skb)->destructor_arg = (void *)((uintptr_t) val | 0x1UL); in skb_zcopy_set_nouarg() 1839 skb_shinfo(skb)->flags |= SKBFL_ZEROCOPY_FRAG; in skb_zcopy_set_nouarg() [all …]
|
| /linux/net/openvswitch/ |
| H A D | openvswitch_trace.h | 52 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 53 __entry->gso_size = skb_shinfo(skb)->gso_size; 54 __entry->gso_type = skb_shinfo(skb)->gso_type; 122 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 123 __entry->gso_size = skb_shinfo(skb)->gso_size; 124 __entry->gso_type = skb_shinfo(skb)->gso_type;
|
| /linux/include/trace/events/ |
| H A D | net.h | 57 __entry->tx_flags = skb_shinfo(skb)->tx_flags; 58 __entry->gso_size = skb_shinfo(skb)->gso_size; 59 __entry->gso_segs = skb_shinfo(skb)->gso_segs; 60 __entry->gso_type = skb_shinfo(skb)->gso_type; 229 __entry->nr_frags = skb_shinfo(skb)->nr_frags; 230 __entry->gso_size = skb_shinfo(skb)->gso_size; 231 __entry->gso_type = skb_shinfo(skb)->gso_type;
|
| /linux/net/tls/ |
| H A D | tls_strp.c | 32 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_anchor_free() 53 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in tls_strp_skb_copy() 54 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in tls_strp_skb_copy() 147 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_msg_hold() 194 struct skb_shared_info *shinfo = skb_shinfo(strp->anchor); in tls_strp_flush_anchor_copy() 219 if (unlikely(nfrag >= skb_shinfo(skb)->nr_frags)) { in tls_strp_copyin_frag() 224 frag = &skb_shinfo(skb)->frags[nfrag]; in tls_strp_copyin_frag() 305 shinfo = skb_shinfo(skb); in tls_strp_copyin_skb() 404 shinfo = skb_shinfo(strp->anchor); in tls_strp_read_copy() 440 first = skb_shinfo(strp->anchor)->frag_list; in tls_strp_check_queue_ok() [all …]
|
| /linux/drivers/net/ethernet/sfc/ |
| H A D | tx_tso.c | 291 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size; in tso_start_new_packet() 295 st->packet_space = skb_shinfo(skb)->gso_size; in tso_start_new_packet() 341 st->seqnum += skb_shinfo(skb)->gso_size; in tso_start_new_packet() 386 EFX_WARN_ON_ONCE_PARANOID(skb_shinfo(skb)->nr_frags < 1); in efx_enqueue_skb_tso() 389 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso() 408 if (++frag_i >= skb_shinfo(skb)->nr_frags) in efx_enqueue_skb_tso() 412 skb_shinfo(skb)->frags + frag_i); in efx_enqueue_skb_tso()
|
| /linux/include/net/ |
| H A D | tso.h | 23 return skb_shinfo(skb)->gso_segs * 2 + skb_shinfo(skb)->nr_frags; in tso_count_descs()
|
| /linux/net/sctp/ |
| H A D | inqueue.c | 133 chunk->skb = skb_shinfo(chunk->skb)->frag_list; in sctp_inq_pop() 168 if (skb_shinfo(chunk->skb)->frag_list) in sctp_inq_pop() 173 if (WARN_ON(!skb_shinfo(chunk->skb)->frag_list)) { in sctp_inq_pop() 179 chunk->skb = skb_shinfo(chunk->skb)->frag_list; in sctp_inq_pop()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_txrx.c | 148 if (skb_shinfo(skb)->tx_flags & SKBTX_IN_PROGRESS) { in otx2_snd_pkt_handler() 196 if (likely(!skb_shinfo(skb)->nr_frags)) { in otx2_skb_add_frag() 209 if (likely(skb_shinfo(skb)->nr_frags < MAX_SKB_FRAGS)) { in otx2_skb_add_frag() 210 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, in otx2_skb_add_frag() 714 if (skb_shinfo(skb)->gso_size) { in otx2_sqe_add_ext() 717 ext->lso_mps = skb_shinfo(skb)->gso_size; in otx2_sqe_add_ext() 720 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) { in otx2_sqe_add_ext() 729 } else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) { in otx2_sqe_add_ext() 732 } else if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) { in otx2_sqe_add_ext() 755 } else if (skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) { in otx2_sqe_add_ext() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | cxgb4_ptp.h | 57 return skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP; in cxgb4_xmit_with_hwtstamp() 62 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in cxgb4_xmit_hwtstamp_pending()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_tx.c | 162 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in mlx5e_tx_get_gso_ihs() 170 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) { in mlx5e_tx_get_gso_ihs() 210 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in mlx5e_txwqe_build_dsegs() 211 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; in mlx5e_txwqe_build_dsegs() 283 .mss = cpu_to_be16(skb_shinfo(skb)->gso_size), in mlx5e_sq_xmit_prepare() 285 .num_bytes = skb->len + (skb_shinfo(skb)->gso_segs - 1) * ihs, in mlx5e_sq_xmit_prepare() 290 stats->packets += skb_shinfo(skb)->gso_segs; in mlx5e_sq_xmit_prepare() 323 ds_cnt += !!attr->headlen + skb_shinfo(skb)->nr_frags + ds_cnt_ids; in mlx5e_sq_calc_wqe_attr() 347 if (unlikely(skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP)) in mlx5e_tx_skb_update_ts_flags() 348 skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; in mlx5e_tx_skb_update_ts_flags() [all …]
|
| /linux/drivers/net/ethernet/sun/ |
| H A D | sunvnet_common.c | 1084 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_map() 1085 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_map() 1124 docopy = skb_shinfo(skb)->nr_frags >= ncookies; in vnet_skb_shape() 1125 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in vnet_skb_shape() 1126 skb_frag_t *f = &skb_shinfo(skb)->frags[i]; in vnet_skb_shape() 1210 skb_shinfo(nskb)->gso_size = skb_shinfo(skb)->gso_size; in vnet_skb_shape() 1211 skb_shinfo(nskb)->gso_type = skb_shinfo(skb)->gso_type; in vnet_skb_shape() 1250 gso_size = skb_shinfo(skb)->gso_size; in vnet_handle_offloads() 1251 gso_type = skb_shinfo(skb)->gso_type; in vnet_handle_offloads() 1252 gso_segs = skb_shinfo(skb)->gso_segs; in vnet_handle_offloads() [all …]
|