| /linux/net/sched/ |
| H A D | sch_choke.c | 148 static bool choke_match_flow(struct sk_buff *skb1, in choke_match_flow() argument 153 if (skb1->protocol != skb2->protocol) in choke_match_flow() 156 if (!choke_skb_cb(skb1)->keys_valid) { in choke_match_flow() 157 choke_skb_cb(skb1)->keys_valid = 1; in choke_match_flow() 158 skb_flow_dissect_flow_keys(skb1, &temp, 0); in choke_match_flow() 159 make_flow_keys_digest(&choke_skb_cb(skb1)->keys, &temp); in choke_match_flow() 168 return !memcmp(&choke_skb_cb(skb1)->keys, in choke_match_flow() 170 sizeof(choke_skb_cb(skb1)->keys)); in choke_match_flow()
|
| /linux/net/llc/ |
| H A D | llc_sap.c | 369 struct sk_buff *skb1; in llc_do_mcast() local 373 skb1 = skb_clone(skb, GFP_ATOMIC); in llc_do_mcast() 374 if (!skb1) { in llc_do_mcast() 379 llc_sap_rcv(sap, skb1, stack[i]); in llc_do_mcast()
|
| /linux/net/core/ |
| H A D | skbuff.c | 725 fclones = container_of(skb, struct sk_buff_fclones, skb1); in __alloc_skb() 1141 fclones = container_of(skb, struct sk_buff_fclones, skb1); in kfree_skbmem() 2092 skb1); in skb_clone() 4128 struct sk_buff* skb1, in skb_split_inside_header() argument 4133 skb_copy_from_linear_data_offset(skb, len, skb_put(skb1, pos - len), in skb_split_inside_header() 4137 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; in skb_split_inside_header() 4139 skb_shinfo(skb1)->nr_frags = skb_shinfo(skb)->nr_frags; in skb_split_inside_header() 4140 skb1->unreadable = skb->unreadable; in skb_split_inside_header() 4142 skb1->data_len = skb->data_len; in skb_split_inside_header() 4143 skb1->len += skb1->data_len; in skb_split_inside_header() [all …]
|
| /linux/net/batman-adv/ |
| H A D | send.c | 1042 struct sk_buff *skb1; in batadv_send_outstanding_bcast_packet() local 1061 skb1 = skb_clone(forw_packet->skb, GFP_ATOMIC); in batadv_send_outstanding_bcast_packet() 1062 if (!skb1) in batadv_send_outstanding_bcast_packet() 1065 batadv_send_broadcast_skb(skb1, forw_packet->if_outgoing); in batadv_send_outstanding_bcast_packet()
|
| /linux/net/mptcp/ |
| H A D | protocol.c | 248 struct sk_buff *skb1; in mptcp_data_queue_ofo() local 296 skb1 = rb_to_skb(parent); in mptcp_data_queue_ofo() 297 if (before64(seq, MPTCP_SKB_CB(skb1)->map_seq)) { in mptcp_data_queue_ofo() 301 if (before64(seq, MPTCP_SKB_CB(skb1)->end_seq)) { in mptcp_data_queue_ofo() 302 if (!after64(end_seq, MPTCP_SKB_CB(skb1)->end_seq)) { in mptcp_data_queue_ofo() 308 if (after64(seq, MPTCP_SKB_CB(skb1)->map_seq)) { in mptcp_data_queue_ofo() 318 rb_replace_node(&skb1->rbnode, &skb->rbnode, in mptcp_data_queue_ofo() 320 mptcp_drop(sk, skb1); in mptcp_data_queue_ofo() 324 } else if (mptcp_ooo_try_coalesce(msk, skb1, skb)) { in mptcp_data_queue_ofo() 338 while ((skb1 = skb_rb_next(skb)) != NULL) { in mptcp_data_queue_ofo() [all …]
|
| /linux/net/ipv4/ |
| H A D | tcp_input.c | 5352 struct sk_buff *skb1; in tcp_data_queue_ofo() local 5416 skb1 = rb_to_skb(parent); in tcp_data_queue_ofo() 5417 if (before(seq, TCP_SKB_CB(skb1)->seq)) { in tcp_data_queue_ofo() 5421 if (before(seq, TCP_SKB_CB(skb1)->end_seq)) { in tcp_data_queue_ofo() 5422 if (!after(end_seq, TCP_SKB_CB(skb1)->end_seq)) { in tcp_data_queue_ofo() 5432 if (after(seq, TCP_SKB_CB(skb1)->seq)) { in tcp_data_queue_ofo() 5434 tcp_dsack_set(sk, seq, TCP_SKB_CB(skb1)->end_seq); in tcp_data_queue_ofo() 5439 rb_replace_node(&skb1->rbnode, &skb->rbnode, in tcp_data_queue_ofo() 5442 TCP_SKB_CB(skb1)->seq, in tcp_data_queue_ofo() 5443 TCP_SKB_CB(skb1)->end_seq); in tcp_data_queue_ofo() [all …]
|
| H A D | icmp.c | 398 struct sk_buff *skb1; in icmp_push_reply() local 403 skb_queue_walk(&sk->sk_write_queue, skb1) { in icmp_push_reply() 404 csum = csum_add(csum, skb1->csum); in icmp_push_reply()
|
| /linux/drivers/net/wireless/ath/ath6kl/ |
| H A D | txrx.c | 1315 struct sk_buff *skb1 = NULL; in ath6kl_rx() local 1572 skb1 = skb_copy(skb, GFP_ATOMIC); in ath6kl_rx() 1583 skb1 = skb; in ath6kl_rx() 1590 if (skb1) in ath6kl_rx() 1591 ath6kl_data_tx(skb1, vif->ndev); in ath6kl_rx()
|
| /linux/drivers/net/vxlan/ |
| H A D | vxlan_mdb.c | 1693 struct sk_buff *skb1; in vxlan_mdb_xmit() local 1705 skb1 = skb_clone(skb, GFP_ATOMIC); in vxlan_mdb_xmit() 1706 if (skb1) in vxlan_mdb_xmit() 1707 vxlan_xmit_one(skb1, vxlan->dev, src_vni, in vxlan_mdb_xmit()
|
| H A D | vxlan_core.c | 2820 struct sk_buff *skb1; in vxlan_xmit() local 2826 skb1 = skb_clone(skb, GFP_ATOMIC); in vxlan_xmit() 2827 if (skb1) in vxlan_xmit() 2828 vxlan_xmit_one(skb1, dev, vni, rdst, did_rsc); in vxlan_xmit()
|
| /linux/include/linux/ |
| H A D | skbuff.h | 1395 struct sk_buff skb1; member 1416 fclones = container_of(skb, struct sk_buff_fclones, skb1); in skb_fclone_busy() 1694 static inline int skb_cmp_decrypted(const struct sk_buff *skb1, in skb_cmp_decrypted() argument 1698 return skb2->decrypted - skb1->decrypted; in skb_cmp_decrypted() 1807 static inline bool skb_pure_zcopy_same(const struct sk_buff *skb1, in skb_pure_zcopy_same() argument 1810 return skb_zcopy_pure(skb1) == skb_zcopy_pure(skb2); in skb_pure_zcopy_same() 4272 void skb_split(struct sk_buff *skb, struct sk_buff *skb1, const u32 len);
|
| /linux/drivers/net/ethernet/qlogic/ |
| H A D | qla3xxx.c | 2042 struct sk_buff *skb1 = NULL, *skb2; in ql_process_macip_rx_intr() local 2056 skb1 = lrg_buf_cb1->skb; in ql_process_macip_rx_intr() 2058 if (*((u16 *) skb1->data) != 0xFFFF) in ql_process_macip_rx_intr() 2078 skb_copy_from_linear_data_offset(skb1, VLAN_ID_LEN, in ql_process_macip_rx_intr()
|
| /linux/Documentation/networking/ |
| H A D | snmp_counter.rst | 725 10 to 15, skb1 has seq 10 to 13, skb2 has seq 14 to 20. The seq 14 and 726 15 in skb2 would be moved to skb1. This operation is 'shift'. If a 727 SACK block acknowledges seq 10 to 20, skb1 has seq 10 to 13, skb2 has 728 seq 14 to 20. All data in skb2 will be moved to skb1, and skb2 will be
|