/linux/net/tls/ |
H A D | tls_strp.c | 58 skb_frag_size(frag))); in tls_strp_skb_copy() 59 offset += skb_frag_size(frag); in tls_strp_skb_copy() 224 chunk = min_t(size_t, len, PAGE_SIZE - skb_frag_size(frag)); in tls_strp_copyin_frag() 227 skb_frag_size(frag), in tls_strp_copyin_frag() 262 chunk = min_t(size_t, chunk, PAGE_SIZE - skb_frag_size(frag)); in tls_strp_copyin_frag() 265 skb_frag_size(frag), in tls_strp_copyin_frag()
|
H A D | tls_device_fallback.c | 264 skb_frag_size(frag), skb_frag_off(frag)); in fill_sg_in() 266 remaining -= skb_frag_size(frag); in fill_sg_in()
|
/linux/net/core/ |
H A D | tso.c | 57 tso->size = skb_frag_size(frag); in tso_build_data() 83 tso->size = skb_frag_size(frag); in tso_start()
|
H A D | skbuff.c | 1337 len -= skb_frag_size(frag); in skb_dump() 1344 skb_frag_size(frag), p, p_off, p_len, in skb_dump() 1977 skb_frag_foreach_page(f, skb_frag_off(f), skb_frag_size(f), in skb_copy_ubufs() 2683 int end = offset + skb_frag_size(&skb_shinfo(skb)->frags[i]); in ___pskb_trim() 2825 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() 2893 int size = skb_frag_size(&skb_shinfo(skb)->frags[i]); in __pskb_pull_tail() 2969 end = start + skb_frag_size(f); in skb_copy_bits() 3157 skb_frag_off(f), skb_frag_size(f), in __skb_splice_bits() 3275 if (offset < skb_frag_size(frag)) in __skb_send_sock() 3278 offset -= skb_frag_size(frag); in __skb_send_sock() [all …]
|
H A D | datagram.c | 421 end = start + skb_frag_size(frag); in __skb_datagram_iter() 575 end = start + skb_frag_size(frag); in skb_copy_datagram_from_iter() 682 start == skb_frag_off(last) + skb_frag_size(last)) { in zerocopy_fill_skb_from_iter()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | xdp.c | 139 len = skb_frag_size(frag); in mlx5e_xmit_xdp_buff() 476 tmp.len = skb_frag_size(frag); in mlx5e_xmit_xdp_frame_mpwqe() 631 dseg->byte_count = cpu_to_be32(skb_frag_size(frag)); in mlx5e_xmit_xdp_frame() 691 skb_frag_size(frag), DMA_TO_DEVICE); in mlx5e_free_xdpsq_desc() 891 skb_frag_size(frag), DMA_TO_DEVICE); in mlx5e_xdp_xmit() 898 skb_frag_size(&xdptxdf.sinfo->frags[j]), in mlx5e_xdp_xmit() 915 skb_frag_size(&xdptxdf.sinfo->frags[j]), in mlx5e_xdp_xmit()
|
/linux/drivers/net/ethernet/sfc/ |
H A D | tx_tso.c | 212 skb_frag_size(frag), DMA_TO_DEVICE); in tso_get_fragment() 214 st->unmap_len = skb_frag_size(frag); in tso_get_fragment() 215 st->in_len = skb_frag_size(frag); in tso_get_fragment()
|
/linux/drivers/net/ethernet/netronome/nfp/nfdk/ |
H A D | rings.c | 44 size = skb_frag_size(frag); in nfp_nfdk_tx_ring_reset() 46 skb_frag_size(frag), DMA_TO_DEVICE); in nfp_nfdk_tx_ring_reset()
|
/linux/net/ipv4/ |
H A D | tcp_sigpool.c | 351 sg_set_page(&sg, page, skb_frag_size(f), offset_in_page(offset)); in tcp_sigpool_hash_skb_data() 352 ahash_request_set_crypt(req, &sg, NULL, skb_frag_size(f)); in tcp_sigpool_hash_skb_data()
|
/linux/drivers/net/ethernet/intel/iavf/ |
H A D | iavf_txrx.c | 2071 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 2072 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 2073 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 2074 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 2075 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 2081 int stale_size = skb_frag_size(stale); in __iavf_chk_linearize() 2083 sum += skb_frag_size(frag++); in __iavf_chk_linearize() 2223 size = skb_frag_size(frag); in iavf_tx_map()
|
H A D | iavf_txrx.h | 343 size = skb_frag_size(frag++); in iavf_xmit_descriptor_count()
|
/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_txrx.c | 1798 size = skb_frag_size(frag); in ice_tx_map() 2278 size = skb_frag_size(frag++); in ice_xmit_desc_count() 2322 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2323 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2324 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2325 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2326 sum += skb_frag_size(frag++); in __ice_chk_linearize() 2332 int stale_size = skb_frag_size(stale); in __ice_chk_linearize() 2334 sum += skb_frag_size(frag++); in __ice_chk_linearize()
|
H A D | ice_xsk.c | 592 memcpy(addr, skb_frag_page(frag), skb_frag_size(frag)); in ice_construct_skb_zc() 595 addr, 0, skb_frag_size(frag)); in ice_construct_skb_zc() 733 size = skb_frag_size(&sinfo->frags[frag]); in ice_xmit_xdp_tx_zc()
|
/linux/drivers/net/ethernet/pensando/ionic/ |
H A D | ionic_txrx.c | 355 skb_frag_size(frag), in ionic_xdp_post_frame() 359 skb_frag_size(frag)); in ionic_xdp_post_frame() 366 bi->len = skb_frag_size(frag); in ionic_xdp_post_frame() 1128 dma_addr = ionic_tx_map_frag(q, frag, 0, skb_frag_size(frag)); in ionic_tx_map_skb() 1132 buf_info->len = skb_frag_size(frag); in ionic_tx_map_skb() 1743 frag_rem = skb_frag_size(frag); in ionic_tx_descs_needed()
|
/linux/drivers/net/ethernet/aeroflex/ |
H A D | greth.c | 113 skb_frag_size(&skb_shinfo(skb)->frags[i]), true); in greth_print_tx_packet() 203 skb_frag_size(frag), in greth_clean_rings() 526 status |= skb_frag_size(frag) & GRETH_BD_LEN; in greth_start_xmit_gbit() 540 dma_addr = skb_frag_dma_map(greth->dev, frag, 0, skb_frag_size(frag), in greth_start_xmit_gbit() 721 skb_frag_size(frag), in greth_clean_tx_gbit()
|
/linux/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.c | 3515 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3516 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3517 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3518 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3519 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3525 int stale_size = skb_frag_size(stale); in __i40e_chk_linearize() 3527 sum += skb_frag_size(frag++); in __i40e_chk_linearize() 3649 size = skb_frag_size(frag); in i40e_tx_map() 3825 size = skb_frag_size(&sinfo->frags[i]); in i40e_xmit_xdp_ring()
|
H A D | i40e_xsk.c | 333 memcpy(addr, skb_frag_page(frag), skb_frag_size(frag)); in i40e_construct_skb_zc() 336 addr, 0, skb_frag_size(frag)); in i40e_construct_skb_zc()
|
/linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 290 *++addr = skb_frag_dma_map(dev, fp, 0, skb_frag_size(fp), in map_skb() 299 dma_unmap_page(dev, *--addr, skb_frag_size(fp), DMA_TO_DEVICE); in map_skb() 917 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in write_sgl() 933 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl() 934 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in write_sgl() 939 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in write_sgl()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | ktls_tx.c | 664 remaining -= skb_frag_size(frag); in tx_sync_info_get() 738 fsz = skb_frag_size(frag); in tx_post_resync_dump() 795 orig_fsz = skb_frag_size(f); in mlx5e_ktls_tx_handle_ooo()
|
/linux/drivers/net/ethernet/netronome/nfp/nfd3/ |
H A D | rings.c | 66 skb_frag_size(frag), DMA_TO_DEVICE); in nfp_nfd3_tx_ring_reset()
|
/linux/drivers/net/ethernet/intel/idpf/ |
H A D | idpf_txrx.c | 2032 size = skb_frag_size(&shinfo->frags[i]); in idpf_tx_res_count_required() 2296 size = skb_frag_size(frag); in idpf_tx_splitq_map() 2437 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2438 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2439 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2440 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2441 sum += skb_frag_size(frag++); in __idpf_chk_linearize() 2447 int stale_size = skb_frag_size(stale); in __idpf_chk_linearize() 2449 sum += skb_frag_size(frag++); in __idpf_chk_linearize()
|
/linux/drivers/net/ethernet/huawei/hinic3/ |
H A D | hinic3_tx.c | 96 skb_frag_size(frag), in hinic3_tx_map_skb() 102 dma_info[idx].len = skb_frag_size(frag); in hinic3_tx_map_skb()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | vnic_sdma.c | 70 skb_frag_size(frag), in build_vnic_ulp_payload()
|
/linux/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 257 *++addr = skb_frag_dma_map(dev, fp, 0, skb_frag_size(fp), in cxgb4_map_skb() 266 dma_unmap_page(dev, *--addr, skb_frag_size(fp), DMA_TO_DEVICE); in cxgb4_map_skb() 285 dma_unmap_page(dev, *addr++, skb_frag_size(fp), DMA_TO_DEVICE); in unmap_skb() 839 sgl->len0 = htonl(skb_frag_size(&si->frags[0])); in cxgb4_write_sgl() 855 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl() 856 to->len[1] = cpu_to_be32(skb_frag_size(&si->frags[++i])); in cxgb4_write_sgl() 861 to->len[0] = cpu_to_be32(skb_frag_size(&si->frags[i])); in cxgb4_write_sgl() 914 frag_size = skb_frag_size(frag); in cxgb4_write_partial_sgl() 920 frag_size = skb_frag_size(frag); in cxgb4_write_partial_sgl() 923 frag_size = min(len, skb_frag_size(frag) - start); in cxgb4_write_partial_sgl() [all …]
|
/linux/drivers/target/iscsi/cxgbit/ |
H A D | cxgbit_target.c | 886 skb_frag_size(dfrag), skb_frag_off(dfrag)); in cxgbit_handle_immediate_data() 1405 skb_frag_size(&ssi->frags[i])); in cxgbit_lro_skb_dump() 1449 len = skb_frag_size(&hssi->frags[hfrag_idx]); in cxgbit_lro_skb_merge() 1469 len += skb_frag_size(&hssi->frags[dfrag_idx]); in cxgbit_lro_skb_merge()
|