Home
last modified time | relevance | path

Searched refs:skb_headlen (Results 1 – 25 of 179) sorted by relevance

12345678

/linux/tools/testing/selftests/bpf/progs/
H A Ddecap_sanity.c16 static unsigned int skb_headlen(const struct sk_buff *skb) in skb_headlen() function
62 (unsigned int)skb_checksum_start_offset(kskb) >= skb_headlen(kskb)) in decap_sanity()
/linux/net/core/
H A Dskbuff.c1329 level, skb->len, skb->data_len, headroom, skb_headlen(skb), in skb_dump()
1358 seg_len = min_t(int, skb_headlen(skb), len); in skb_dump()
2219 unsigned int size = skb_headlen(skb) + headroom; in __pskb_copy_fclone()
2229 skb_put(n, skb_headlen(skb)); in __pskb_copy_fclone()
2715 int offset = skb_headlen(skb); in ___pskb_trim()
2781 if (len > skb_headlen(skb)) { in ___pskb_trim()
2807 int hdlen = (len > skb_headlen(skb)) ? skb_headlen(skb) : len; in pskb_trim_rcsum_slow()
2859 BUG_ON(skb_copy_bits(skb, skb_headlen(skb), in __pskb_pull_tail()
2899 skb_headlen(list)) in __pskb_pull_tail()
2988 int start = skb_headlen(skb); in skb_copy_bits()
[all …]
H A Dgro.c96 unsigned int headlen = skb_headlen(skb); in skb_gro_receive()
395 headlen = skb_headlen(skb); in skb_gro_reset_offset()
437 int grow = skb_gro_offset(skb) - skb_headlen(skb); in gro_try_pull_from_frag0()
648 __skb_pull(skb, skb_headlen(skb)); in napi_reuse_skb()
H A Ddatagram.c394 int start = skb_headlen(skb); in __skb_datagram_iter()
553 int start = skb_headlen(skb); in skb_copy_datagram_from_iter()
786 int copy = min_t(int, skb_headlen(skb), iov_iter_count(from)); in zerocopy_sg_from_iter()
H A Dgso.c236 if (seg_len + skb_headlen(iter) > max_len) in skb_gso_size_check()
/linux/drivers/net/ethernet/sfc/
H A Dtx_tso.c163 skb_headlen(skb)); in efx_tso_check_protocol()
180 in_len = skb_headlen(skb) - header_len; in tso_start()
199 skb_headlen(skb), DMA_TO_DEVICE); in tso_start()
201 st->header_unmap_len = skb_headlen(skb); in tso_start()
/linux/net/xfrm/
H A Dxfrm_ipcomp.c142 if (!skb_cloned(skb) && skb_headlen(skb) >= minhead) { in ipcomp_setup_req()
153 nfrags = skb_cow_data(skb, skb_headlen(skb) < minhead ? in ipcomp_setup_req()
154 minhead - skb_headlen(skb) : 0, in ipcomp_setup_req()
H A Dxfrm_iptfs.c302 skb_frag_fill_page_desc(frag, page, skb->data - addr, skb_headlen(skb)); in iptfs_skb_head_to_frag()
349 if (initial_offset >= skb_headlen(skb)) { in iptfs_skb_prepare_frag_walk()
350 initial_offset -= skb_headlen(skb); in iptfs_skb_prepare_frag_walk()
360 initial_offset -= skb_headlen(skb); in iptfs_skb_prepare_frag_walk()
1057 } else if (data + iphlen <= skb_headlen(skb) && in __input_process_payload()
2159 if (!skb_headlen(from)) { in iptfs_consume_frags()
2286 (!skb2->head_frag && skb_headlen(skb)) || in iptfs_output_queued()
/linux/drivers/infiniband/ulp/ipoib/
H A Dipoib_ib.c283 if (skb_headlen(skb)) { in ipoib_dma_map_tx()
284 mapping[0] = ib_dma_map_single(ca, skb->data, skb_headlen(skb), in ipoib_dma_map_tx()
313 ib_dma_unmap_single(ca, mapping[0], skb_headlen(skb), DMA_TO_DEVICE); in ipoib_dma_map_tx()
326 if (skb_headlen(skb)) { in ipoib_dma_unmap_tx()
327 ib_dma_unmap_single(priv->ca, mapping[0], skb_headlen(skb), in ipoib_dma_unmap_tx()
598 unsigned int usable_sge = priv->max_send_sge - !!skb_headlen(skb); in ipoib_send()
/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_tx.c83 hlen = eth_get_headlen(skb->dev, skb->data, skb_headlen(skb)); in mlx5e_calc_min_inline()
94 return min_t(u16, hlen, skb_headlen(skb)); in mlx5e_calc_min_inline()
278 .headlen = skb_headlen(skb) - ihs, in mlx5e_sq_xmit_prepare()
291 .headlen = skb_headlen(skb) - ihs, in mlx5e_sq_xmit_prepare()
H A Den_selftest.c173 if (MLX5E_TEST_PKT_SIZE - ETH_HLEN > skb_headlen(skb)) in mlx5e_test_loopback_validate()
/linux/drivers/block/aoe/
H A Daoenet.c149 sn -= skb_headlen(skb); in aoenet_rcv()
/linux/drivers/net/ethernet/stmicro/stmmac/
H A Dchain_mode.c20 unsigned int nopaged_len = skb_headlen(skb); in jumbo_frm()
/linux/drivers/net/wireless/ath/wil6210/
H A Dtxrx.c510 skb->data, skb_headlen(skb), false); in wil_vring_reap_rx()
560 skb->data, skb_headlen(skb), false); in wil_vring_reap_rx()
1836 headlen = skb_headlen(skb) - hdrlen; in __wil_tx_vring_tso()
1872 skb_headlen(skb) - headlen, in __wil_tx_vring_tso()
2042 uint len = skb_headlen(skb); in __wil_tx_ring()
2058 pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); in __wil_tx_ring()
2061 skb_headlen(skb), skb->data, &pa); in __wil_tx_ring()
2063 skb->data, skb_headlen(skb), false); in __wil_tx_ring()
/linux/drivers/net/ethernet/aeroflex/
H A Dgreth.c104 length = skb_headlen(skb); in greth_print_tx_packet()
194 skb_headlen(skb), in greth_clean_rings()
501 status |= skb_headlen(skb) & GRETH_BD_LEN; in greth_start_xmit_gbit()
508 dma_addr = dma_map_single(greth->dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); in greth_start_xmit_gbit()
712 skb_headlen(skb), in greth_clean_tx_gbit()
/linux/drivers/net/ethernet/netronome/nfp/nfdk/
H A Drings.c36 size = skb_headlen(skb); in nfp_nfdk_tx_ring_reset()
/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_txrx.h335 int count = 0, size = skb_headlen(skb); in iavf_xmit_descriptor_count()
/linux/drivers/net/ethernet/netronome/nfp/nfd3/
H A Drings.c61 skb_headlen(skb), DMA_TO_DEVICE); in nfp_nfd3_tx_ring_reset()
/linux/drivers/net/ethernet/google/gve/
H A Dgve_tx_dqo.c709 u32 len = skb_headlen(skb); in gve_tx_add_skb_no_copy_dqo()
899 num_descs = gve_num_descs_per_buf(skb_headlen(skb)); in gve_num_buffer_descs_needed()
928 cur_seg_size = skb_headlen(skb) - header_len; in gve_can_send_tso()
929 prev_frag_size = skb_headlen(skb); in gve_can_send_tso()
/linux/drivers/net/
H A Dxen-netfront.c570 skb_headlen(skb)); in xennet_count_skb_slots()
771 len = skb_headlen(skb); in xennet_start_xmit()
1180 BUG_ON(pull_to < skb_headlen(skb)); in xennet_fill_frags()
1181 __pskb_pull_tail(skb, pull_to - skb_headlen(skb)); in xennet_fill_frags()
1237 if (pull_to > skb_headlen(skb)) in handle_incoming_queue()
1238 __pskb_pull_tail(skb, pull_to - skb_headlen(skb)); in handle_incoming_queue()
/linux/drivers/net/ethernet/agere/
H A Det131x.c2453 if (skb_headlen(skb) <= 1514) { in nic_send_packet()
2457 desc[frag].len_vlan = skb_headlen(skb); in nic_send_packet()
2460 skb_headlen(skb), in nic_send_packet()
2470 desc[frag].len_vlan = skb_headlen(skb) / 2; in nic_send_packet()
2473 skb_headlen(skb) / 2, in nic_send_packet()
2483 desc[frag].len_vlan = skb_headlen(skb) / 2; in nic_send_packet()
2486 skb_headlen(skb) / 2, in nic_send_packet()
2487 skb_headlen(skb) / 2, in nic_send_packet()
/linux/drivers/net/ethernet/amd/xgbe/
H A Dxgbe-selftest.c61 if (skb_headlen(skb) < (NET_TEST_PKT_SIZE - ETH_HLEN)) in xgbe_test_loopback_validate()
/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c281 if (skb_headlen(skb) > hlen) in qede_pkt_req_lin()
1532 skb_headlen(skb), DMA_TO_DEVICE); in qede_start_xmit()
1540 BD_SET_UNMAP_ADDR_LEN(first_bd, mapping, skb_headlen(skb)); in qede_start_xmit()
1625 if (unlikely(skb_headlen(skb) > hlen)) { in qede_start_xmit()
/linux/io_uring/
H A Dzcrx.c1474 if (unlikely(offset < skb_headlen(skb))) { in io_zcrx_recv_skb()
1478 to_copy = min_t(size_t, skb_headlen(skb) - offset, len); in io_zcrx_recv_skb()
1490 if (offset != skb_headlen(skb)) in io_zcrx_recv_skb()
1494 start = skb_headlen(skb); in io_zcrx_recv_skb()
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_txrx.h508 int count = 0, size = skb_headlen(skb); in i40e_xmit_descriptor_count()

12345678