| /linux/Documentation/networking/ |
| H A D | skbuff.rst | 15 Shared skbs and skb clones 19 to keep a struct sk_buff alive. skbs with a ``sk_buff.users != 1`` are referred 20 to as shared skbs (see skb_shared()). 22 skb_clone() allows for fast duplication of skbs. None of the data buffers 24 &skb_shared_info.refcount indicates the number of skbs pointing at the same 27 dataref and headerless skbs 31 :doc: dataref and headerless skbs
|
| H A D | xdp-rx-metadata.rst | 87 ``skbs``. However, TC-BPF programs can access the XDP metadata area using 91 can override some of the metadata used for building ``skbs``.
|
| /linux/kernel/bpf/ |
| H A D | cpumap.c | 140 void **skbs, u32 skb_n, in cpu_map_bpf_prog_run_skb() argument 148 struct sk_buff *skb = skbs[i]; in cpu_map_bpf_prog_run_skb() 153 skbs[pass++] = skb; in cpu_map_bpf_prog_run_skb() 251 void **skbs, struct cpu_map_ret *ret, in cpu_map_bpf_prog_run() argument 265 ret->skb_n = cpu_map_bpf_prog_run_skb(rcpu, skbs, ret->skb_n, in cpu_map_bpf_prog_run() 277 memmove(&skbs[ret->xdp_n], skbs, ret->skb_n * sizeof(*skbs)); in cpu_map_bpf_prog_run() 312 void *skbs[CPUMAP_BATCH]; in cpu_map_kthread_run() local 347 skbs[ret.skb_n++] = skb; in cpu_map_kthread_run() 364 cpu_map_bpf_prog_run(rcpu, frames, skbs, &ret, &stats); in cpu_map_kthread_run() 368 m = napi_skb_cache_get_bulk(skbs, ret.xdp_n); in cpu_map_kthread_run() [all …]
|
| /linux/drivers/net/ethernet/actions/ |
| H A D | owl-emac.c | 206 ring->skbs[i] = skb; in owl_emac_ring_prepare_rx() 252 owl_emac_dma_unmap_rx(priv, ring->skbs[i], ring->skbs_dma[i]); in owl_emac_ring_unprepare_rx() 255 dev_kfree_skb(ring->skbs[i]); in owl_emac_ring_unprepare_rx() 256 ring->skbs[i] = NULL; in owl_emac_ring_unprepare_rx() 271 owl_emac_dma_unmap_tx(priv, ring->skbs[i], ring->skbs_dma[i]); in owl_emac_ring_unprepare_tx() 274 dev_kfree_skb(ring->skbs[i]); in owl_emac_ring_unprepare_tx() 275 ring->skbs[i] = NULL; in owl_emac_ring_unprepare_tx() 288 ring->skbs = devm_kcalloc(dev, size, sizeof(struct sk_buff *), in owl_emac_ring_alloc() 290 if (!ring->skbs) in owl_emac_ring_alloc() 530 ring->skbs[tx_head] = skb; in owl_emac_setup_frame_xmit() [all …]
|
| H A D | owl-emac.h | 247 struct sk_buff **skbs; member
|
| /linux/drivers/net/ethernet/sfc/ |
| H A D | selftest.c | 87 struct sk_buff **skbs; member 431 state->skbs[i] = skb; in efx_begin_loopback() 489 skb = state->skbs[i]; in efx_end_loopback() 545 state->skbs = kzalloc_objs(state->skbs[0], state->packet_count); in efx_test_loopback() 546 if (!state->skbs) in efx_test_loopback() 567 kfree(state->skbs); in efx_test_loopback()
|
| /linux/drivers/net/ethernet/sfc/siena/ |
| H A D | selftest.c | 88 struct sk_buff **skbs; member 432 state->skbs[i] = skb; in efx_begin_loopback() 490 skb = state->skbs[i]; in efx_end_loopback() 546 state->skbs = kzalloc_objs(state->skbs[0], state->packet_count); in efx_test_loopback() 547 if (!state->skbs) in efx_test_loopback() 568 kfree(state->skbs); in efx_test_loopback()
|
| /linux/drivers/net/ethernet/sfc/falcon/ |
| H A D | selftest.c | 85 struct sk_buff **skbs; member 434 state->skbs[i] = skb; in ef4_begin_loopback() 492 skb = state->skbs[i]; in ef4_end_loopback() 548 state->skbs = kzalloc_objs(state->skbs[0], state->packet_count); in ef4_test_loopback() 549 if (!state->skbs) in ef4_test_loopback() 570 kfree(state->skbs); in ef4_test_loopback()
|
| /linux/drivers/net/ethernet/socionext/ |
| H A D | sni_ave.c | 227 struct sk_buff *skbs; member 586 skb = priv->rx.desc[entry].skbs; in ave_rxdesc_prepare() 617 priv->rx.desc[entry].skbs = skb; in ave_rxdesc_prepare() 714 if (priv->tx.desc[done_idx].skbs) { in ave_tx_complete() 717 dev_consume_skb_any(priv->tx.desc[done_idx].skbs); in ave_tx_complete() 718 priv->tx.desc[done_idx].skbs = NULL; in ave_tx_complete() 775 skb = priv->rx.desc[proc_idx].skbs; in ave_rx_receive() 776 priv->rx.desc[proc_idx].skbs = NULL; in ave_rx_receive() 1368 if (!priv->tx.desc[entry].skbs) in ave_stop() 1372 dev_kfree_skb_any(priv->tx.desc[entry].skbs); in ave_stop() [all …]
|
| /linux/net/mac80211/ |
| H A D | tx.c | 920 __skb_queue_tail(&tx->skbs, tmp); in ieee80211_fragment() 962 __skb_queue_tail(&tx->skbs, skb); in ieee80211_tx_h_fragment() 999 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_fragment() 1005 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_fragment() 1035 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_stats() 1088 skb_queue_walk(&tx->skbs, skb) { in ieee80211_tx_h_calculate_duration() 1092 if (!skb_queue_is_last(&tx->skbs, skb)) { in ieee80211_tx_h_calculate_duration() 1093 struct sk_buff *next = skb_queue_next(&tx->skbs, skb); in ieee80211_tx_h_calculate_duration() 1222 __skb_queue_head_init(&tx->skbs); in ieee80211_tx_prepare() 1682 struct sk_buff_head *skbs, in ieee80211_tx_frags() argument [all …]
|
| H A D | wep.c | 299 skb_queue_walk(&tx->skbs, skb) { in ieee80211_crypto_wep_encrypt()
|
| /linux/drivers/net/wireless/mediatek/mt7601u/ |
| H A D | dma.c | 283 struct sk_buff_head skbs; in mt7601u_tx_tasklet() local 286 __skb_queue_head_init(&skbs); in mt7601u_tx_tasklet() 295 skb_queue_splice_init(&dev->tx_skb_done, &skbs); in mt7601u_tx_tasklet() 299 while (!skb_queue_empty(&skbs)) { in mt7601u_tx_tasklet() 300 struct sk_buff *skb = __skb_dequeue(&skbs); in mt7601u_tx_tasklet()
|
| /linux/net/mctp/test/ |
| H A D | route-test.c | 771 struct sk_buff *skbs[2]; in mctp_test_route_input_sk_fail_frag() local 783 for (i = 0; i < ARRAY_SIZE(skbs); i++) { in mctp_test_route_input_sk_fail_frag() 784 skbs[i] = mctp_test_create_skb(&hdrs[i], 10); in mctp_test_route_input_sk_fail_frag() 785 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, skbs[i]); in mctp_test_route_input_sk_fail_frag() 786 skb_get(skbs[i]); in mctp_test_route_input_sk_fail_frag() 788 mctp_test_skb_set_dev(skbs[i], dev); in mctp_test_route_input_sk_fail_frag() 794 rc = mctp_dst_input(&dst, skbs[0]); in mctp_test_route_input_sk_fail_frag() 798 rc = mctp_dst_input(&dst, skbs[1]); in mctp_test_route_input_sk_fail_frag() 802 KUNIT_EXPECT_EQ(test, refcount_read(&skbs[0]->users), 1); in mctp_test_route_input_sk_fail_frag() 803 kfree_skb(skbs[0]); in mctp_test_route_input_sk_fail_frag() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| H A D | qede_fp.c | 80 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_tx_pkt() 85 bool data_split = txq->sw_tx_ring.skbs[idx].flags & QEDE_TSO_SPLIT_BD; in qede_free_tx_pkt() 125 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_tx_pkt() 126 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_tx_pkt() 137 struct sk_buff *skb = txq->sw_tx_ring.skbs[idx].skb; in qede_free_failed_tx_pkt() 173 txq->sw_tx_ring.skbs[idx].skb = NULL; in qede_free_failed_tx_pkt() 174 txq->sw_tx_ring.skbs[idx].flags = 0; in qede_free_failed_tx_pkt() 1520 txq->sw_tx_ring.skbs[idx].skb = skb; in qede_start_xmit() 1642 txq->sw_tx_ring.skbs[idx].flags |= QEDE_TSO_SPLIT_BD; in qede_start_xmit()
|
| /linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
| H A D | tx.c | 1125 struct sk_buff_head skbs; in iwlagn_rx_reply_tx() local 1146 __skb_queue_head_init(&skbs); in iwlagn_rx_reply_tx() 1175 iwl_trans_reclaim(priv->trans, txq_id, ssn, &skbs, false); in iwlagn_rx_reply_tx() 1180 skb_queue_walk(&skbs, skb) { in iwlagn_rx_reply_tx() 1251 while (!skb_queue_empty(&skbs)) { in iwlagn_rx_reply_tx() 1252 skb = __skb_dequeue(&skbs); in iwlagn_rx_reply_tx()
|
| /linux/drivers/net/ethernet/renesas/ |
| H A D | rswitch_main.c | 320 kfree(gq->skbs); in rswitch_gwca_queue_free() 321 gq->skbs = NULL; in rswitch_gwca_queue_free() 359 gq->skbs = kzalloc_objs(*gq->skbs, gq->ring_size); in rswitch_gwca_queue_alloc() 360 if (!gq->skbs) in rswitch_gwca_queue_alloc() 903 skb = gq->skbs[gq->dirty]; in rswitch_tx_free() 910 dev_kfree_skb_any(gq->skbs[gq->dirty]); in rswitch_tx_free() 911 gq->skbs[gq->dirty] = NULL; in rswitch_tx_free() 1763 gq->skbs[(gq->cur + nr_desc - 1) % gq->ring_size] = skb; in rswitch_start_xmit() 1784 gq->skbs[(gq->cur + nr_desc - 1) % gq->ring_size] = NULL; in rswitch_start_xmit()
|
| /linux/net/mptcp/ |
| H A D | protocol.c | 2192 static bool __mptcp_move_skbs(struct sock *sk, struct list_head *skbs, u32 *delta) in __mptcp_move_skbs() argument 2194 struct sk_buff *skb = list_first_entry(skbs, struct sk_buff, list); in __mptcp_move_skbs() 2209 if (list_empty(skbs)) in __mptcp_move_skbs() 2212 skb = list_first_entry(skbs, struct sk_buff, list); in __mptcp_move_skbs() 2221 static bool mptcp_can_spool_backlog(struct sock *sk, struct list_head *skbs) in mptcp_can_spool_backlog() argument 2236 INIT_LIST_HEAD(skbs); in mptcp_can_spool_backlog() 2237 list_splice_init(&msk->backlog_list, skbs); in mptcp_can_spool_backlog() 2242 struct list_head *skbs) in mptcp_backlog_spooled() argument 2247 list_splice(skbs, &msk->backlog_list); in mptcp_backlog_spooled() 2252 struct list_head skbs; in mptcp_move_skbs() local [all …]
|
| /linux/drivers/net/ethernet/mediatek/ |
| H A D | mtk_star_emac.c | 241 struct sk_buff *skbs[MTK_STAR_RING_NUM_DESCS]; member 325 desc_data->skb = ring->skbs[ring->tail]; in mtk_star_ring_pop_tail() 328 ring->skbs[ring->tail] = NULL; in mtk_star_ring_pop_tail() 349 ring->skbs[ring->head] = desc_data->skb; in mtk_star_ring_push_head() 718 ring->skbs[i] = skb; in mtk_star_prepare_rx_skbs() 738 desc_data.skb = ring->skbs[i]; in mtk_star_ring_free_skbs()
|
| /linux/tools/testing/selftests/net/packetdrill/ |
| H A D | tcp_zerocopy_maxfrags.pkt | 95 // verify that it is split in skbs with 17 frags
|
| /linux/drivers/net/ethernet/natsemi/ |
| H A D | ns83820.c | 401 struct sk_buff *skbs[NR_RX_DESC]; member 524 BUG_ON(NULL != dev->rx_info.skbs[next_empty]); in ns83820_add_rx_skb() 525 dev->rx_info.skbs[next_empty] = skb; in ns83820_add_rx_skb() 790 struct sk_buff *skb = dev->rx_info.skbs[i]; in ns83820_cleanup_rx() 791 dev->rx_info.skbs[i] = NULL; in ns83820_cleanup_rx() 856 skb = info->skbs[next_rx]; 857 info->skbs[next_rx] = NULL;
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| H A D | 3945.c | 283 skb = txq->skbs[txq->q.read_ptr]; in il3945_tx_queue_reclaim() 285 txq->skbs[txq->q.read_ptr] = NULL; in il3945_tx_queue_reclaim() 332 info = IEEE80211_SKB_CB(txq->skbs[txq->q.read_ptr]); in il3945_hdl_tx() 667 if (txq->skbs) { in il3945_hw_txq_free_tfd() 668 struct sk_buff *skb = txq->skbs[txq->q.read_ptr]; in il3945_hw_txq_free_tfd() 673 txq->skbs[txq->q.read_ptr] = NULL; in il3945_hw_txq_free_tfd()
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| H A D | chtls_main.c | 423 struct sk_buff **skbs, const __be64 *rsp) in chtls_recv() argument 425 struct sk_buff *skb = *skbs; in chtls_recv()
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | cxgb3_offload.c | 480 static int rx_offload_blackhole(struct t3cdev *dev, struct sk_buff **skbs, in rx_offload_blackhole() argument 484 dev_kfree_skb_any(skbs[n]); in rx_offload_blackhole() 980 static int process_rx(struct t3cdev *dev, struct sk_buff **skbs, int n) in process_rx() argument 983 struct sk_buff *skb = *skbs++; in process_rx()
|
| H A D | sge.c | 1877 struct sk_buff *skbs[], int n) in deliver_partial_bundle() argument 1881 tdev->recv(tdev, skbs, n); in deliver_partial_bundle() 1904 struct sk_buff *skb, *tmp, *skbs[RX_BUNDLE_SIZE]; in ofld_poll() local 1926 skbs[ngathered] = skb; in ofld_poll() 1929 adapter->tdev.recv(&adapter->tdev, skbs, in ofld_poll() 1940 deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); in ofld_poll()
|
| /linux/net/bluetooth/ |
| H A D | l2cap_core.c | 61 struct sk_buff_head *skbs, u8 event); 1937 struct sk_buff_head *skbs) in l2cap_streaming_send() argument 1942 BT_DBG("chan %p, skbs %p", chan, skbs); in l2cap_streaming_send() 1944 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_streaming_send() 2782 struct sk_buff_head *skbs, u8 event) in l2cap_tx_state_xmit() argument 2784 BT_DBG("chan %p, control %p, skbs %p, event %d", chan, control, skbs, in l2cap_tx_state_xmit() 2790 chan->tx_send_head = skb_peek(skbs); in l2cap_tx_state_xmit() 2792 skb_queue_splice_tail_init(skbs, &chan->tx_q); in l2cap_tx_state_xmit() 2854 struct sk_buff_head *skbs, u8 event) in l2cap_tx_state_wait_f() argument 2856 BT_DBG("chan %p, control %p, skbs %p, event %d", chan, control, skbs, in l2cap_tx_state_wait_f() [all …]
|