| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | decap_sanity.c | 16 static unsigned int skb_headlen(const struct sk_buff *skb) in skb_headlen() function 62 (unsigned int)skb_checksum_start_offset(kskb) >= skb_headlen(kskb)) in decap_sanity()
|
| /linux/net/core/ |
| H A D | skbuff.c | 1329 level, skb->len, skb->data_len, headroom, skb_headlen(skb), in skb_dump() 1358 seg_len = min_t(int, skb_headlen(skb), len); in skb_dump() 2219 unsigned int size = skb_headlen(skb) + headroom; in __pskb_copy_fclone() 2229 skb_put(n, skb_headlen(skb)); in __pskb_copy_fclone() 2715 int offset = skb_headlen(skb); in ___pskb_trim() 2781 if (len > skb_headlen(skb)) { in ___pskb_trim() 2807 int hdlen = (len > skb_headlen(skb)) ? skb_headlen(skb) : len; in pskb_trim_rcsum_slow() 2859 BUG_ON(skb_copy_bits(skb, skb_headlen(skb), in __pskb_pull_tail() 2899 skb_headlen(list)) in __pskb_pull_tail() 2988 int start = skb_headlen(skb); in skb_copy_bits() [all …]
|
| H A D | gro.c | 96 unsigned int headlen = skb_headlen(skb); in skb_gro_receive() 395 headlen = skb_headlen(skb); in skb_gro_reset_offset() 437 int grow = skb_gro_offset(skb) - skb_headlen(skb); in gro_try_pull_from_frag0() 648 __skb_pull(skb, skb_headlen(skb)); in napi_reuse_skb()
|
| H A D | datagram.c | 394 int start = skb_headlen(skb); in __skb_datagram_iter() 553 int start = skb_headlen(skb); in skb_copy_datagram_from_iter() 786 int copy = min_t(int, skb_headlen(skb), iov_iter_count(from)); in zerocopy_sg_from_iter()
|
| H A D | gso.c | 236 if (seg_len + skb_headlen(iter) > max_len) in skb_gso_size_check()
|
| /linux/drivers/net/ethernet/sfc/ |
| H A D | tx_tso.c | 163 skb_headlen(skb)); in efx_tso_check_protocol() 180 in_len = skb_headlen(skb) - header_len; in tso_start() 199 skb_headlen(skb), DMA_TO_DEVICE); in tso_start() 201 st->header_unmap_len = skb_headlen(skb); in tso_start()
|
| /linux/net/xfrm/ |
| H A D | xfrm_ipcomp.c | 142 if (!skb_cloned(skb) && skb_headlen(skb) >= minhead) { in ipcomp_setup_req() 153 nfrags = skb_cow_data(skb, skb_headlen(skb) < minhead ? in ipcomp_setup_req() 154 minhead - skb_headlen(skb) : 0, in ipcomp_setup_req()
|
| H A D | xfrm_iptfs.c | 302 skb_frag_fill_page_desc(frag, page, skb->data - addr, skb_headlen(skb)); in iptfs_skb_head_to_frag() 349 if (initial_offset >= skb_headlen(skb)) { in iptfs_skb_prepare_frag_walk() 350 initial_offset -= skb_headlen(skb); in iptfs_skb_prepare_frag_walk() 360 initial_offset -= skb_headlen(skb); in iptfs_skb_prepare_frag_walk() 1057 } else if (data + iphlen <= skb_headlen(skb) && in __input_process_payload() 2159 if (!skb_headlen(from)) { in iptfs_consume_frags() 2286 (!skb2->head_frag && skb_headlen(skb)) || in iptfs_output_queued()
|
| /linux/drivers/infiniband/ulp/ipoib/ |
| H A D | ipoib_ib.c | 283 if (skb_headlen(skb)) { in ipoib_dma_map_tx() 284 mapping[0] = ib_dma_map_single(ca, skb->data, skb_headlen(skb), in ipoib_dma_map_tx() 313 ib_dma_unmap_single(ca, mapping[0], skb_headlen(skb), DMA_TO_DEVICE); in ipoib_dma_map_tx() 326 if (skb_headlen(skb)) { in ipoib_dma_unmap_tx() 327 ib_dma_unmap_single(priv->ca, mapping[0], skb_headlen(skb), in ipoib_dma_unmap_tx() 598 unsigned int usable_sge = priv->max_send_sge - !!skb_headlen(skb); in ipoib_send()
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | en_tx.c | 83 hlen = eth_get_headlen(skb->dev, skb->data, skb_headlen(skb)); in mlx5e_calc_min_inline() 94 return min_t(u16, hlen, skb_headlen(skb)); in mlx5e_calc_min_inline() 278 .headlen = skb_headlen(skb) - ihs, in mlx5e_sq_xmit_prepare() 291 .headlen = skb_headlen(skb) - ihs, in mlx5e_sq_xmit_prepare()
|
| H A D | en_selftest.c | 173 if (MLX5E_TEST_PKT_SIZE - ETH_HLEN > skb_headlen(skb)) in mlx5e_test_loopback_validate()
|
| /linux/drivers/block/aoe/ |
| H A D | aoenet.c | 149 sn -= skb_headlen(skb); in aoenet_rcv()
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| H A D | chain_mode.c | 20 unsigned int nopaged_len = skb_headlen(skb); in jumbo_frm()
|
| /linux/drivers/net/wireless/ath/wil6210/ |
| H A D | txrx.c | 510 skb->data, skb_headlen(skb), false); in wil_vring_reap_rx() 560 skb->data, skb_headlen(skb), false); in wil_vring_reap_rx() 1836 headlen = skb_headlen(skb) - hdrlen; in __wil_tx_vring_tso() 1872 skb_headlen(skb) - headlen, in __wil_tx_vring_tso() 2042 uint len = skb_headlen(skb); in __wil_tx_ring() 2058 pa = dma_map_single(dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); in __wil_tx_ring() 2061 skb_headlen(skb), skb->data, &pa); in __wil_tx_ring() 2063 skb->data, skb_headlen(skb), false); in __wil_tx_ring()
|
| /linux/drivers/net/ethernet/aeroflex/ |
| H A D | greth.c | 104 length = skb_headlen(skb); in greth_print_tx_packet() 194 skb_headlen(skb), in greth_clean_rings() 501 status |= skb_headlen(skb) & GRETH_BD_LEN; in greth_start_xmit_gbit() 508 dma_addr = dma_map_single(greth->dev, skb->data, skb_headlen(skb), DMA_TO_DEVICE); in greth_start_xmit_gbit() 712 skb_headlen(skb), in greth_clean_tx_gbit()
|
| /linux/drivers/net/ethernet/netronome/nfp/nfdk/ |
| H A D | rings.c | 36 size = skb_headlen(skb); in nfp_nfdk_tx_ring_reset()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.h | 335 int count = 0, size = skb_headlen(skb); in iavf_xmit_descriptor_count()
|
| /linux/drivers/net/ethernet/netronome/nfp/nfd3/ |
| H A D | rings.c | 61 skb_headlen(skb), DMA_TO_DEVICE); in nfp_nfd3_tx_ring_reset()
|
| /linux/drivers/net/ethernet/google/gve/ |
| H A D | gve_tx_dqo.c | 709 u32 len = skb_headlen(skb); in gve_tx_add_skb_no_copy_dqo() 899 num_descs = gve_num_descs_per_buf(skb_headlen(skb)); in gve_num_buffer_descs_needed() 928 cur_seg_size = skb_headlen(skb) - header_len; in gve_can_send_tso() 929 prev_frag_size = skb_headlen(skb); in gve_can_send_tso()
|
| /linux/drivers/net/ |
| H A D | xen-netfront.c | 570 skb_headlen(skb)); in xennet_count_skb_slots() 771 len = skb_headlen(skb); in xennet_start_xmit() 1180 BUG_ON(pull_to < skb_headlen(skb)); in xennet_fill_frags() 1181 __pskb_pull_tail(skb, pull_to - skb_headlen(skb)); in xennet_fill_frags() 1237 if (pull_to > skb_headlen(skb)) in handle_incoming_queue() 1238 __pskb_pull_tail(skb, pull_to - skb_headlen(skb)); in handle_incoming_queue()
|
| /linux/drivers/net/ethernet/agere/ |
| H A D | et131x.c | 2453 if (skb_headlen(skb) <= 1514) { in nic_send_packet() 2457 desc[frag].len_vlan = skb_headlen(skb); in nic_send_packet() 2460 skb_headlen(skb), in nic_send_packet() 2470 desc[frag].len_vlan = skb_headlen(skb) / 2; in nic_send_packet() 2473 skb_headlen(skb) / 2, in nic_send_packet() 2483 desc[frag].len_vlan = skb_headlen(skb) / 2; in nic_send_packet() 2486 skb_headlen(skb) / 2, in nic_send_packet() 2487 skb_headlen(skb) / 2, in nic_send_packet()
|
| /linux/drivers/net/ethernet/amd/xgbe/ |
| H A D | xgbe-selftest.c | 61 if (skb_headlen(skb) < (NET_TEST_PKT_SIZE - ETH_HLEN)) in xgbe_test_loopback_validate()
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede_fp.c | 281 if (skb_headlen(skb) > hlen) in qede_pkt_req_lin() 1532 skb_headlen(skb), DMA_TO_DEVICE); in qede_start_xmit() 1540 BD_SET_UNMAP_ADDR_LEN(first_bd, mapping, skb_headlen(skb)); in qede_start_xmit() 1625 if (unlikely(skb_headlen(skb) > hlen)) { in qede_start_xmit()
|
| /linux/io_uring/ |
| H A D | zcrx.c | 1474 if (unlikely(offset < skb_headlen(skb))) { in io_zcrx_recv_skb() 1478 to_copy = min_t(size_t, skb_headlen(skb) - offset, len); in io_zcrx_recv_skb() 1490 if (offset != skb_headlen(skb)) in io_zcrx_recv_skb() 1494 start = skb_headlen(skb); in io_zcrx_recv_skb()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_txrx.h | 508 int count = 0, size = skb_headlen(skb); in i40e_xmit_descriptor_count()
|