/linux/net/sched/ |
H A D | sch_choke.c | 148 static bool choke_match_flow(struct sk_buff *skb1, in choke_match_flow() argument 153 if (skb1->protocol != skb2->protocol) in choke_match_flow() 156 if (!choke_skb_cb(skb1)->keys_valid) { in choke_match_flow() 157 choke_skb_cb(skb1)->keys_valid = 1; in choke_match_flow() 158 skb_flow_dissect_flow_keys(skb1, &temp, 0); in choke_match_flow() 159 make_flow_keys_digest(&choke_skb_cb(skb1)->keys, &temp); in choke_match_flow() 168 return !memcmp(&choke_skb_cb(skb1)->keys, in choke_match_flow() 170 sizeof(choke_skb_cb(skb1)->keys)); in choke_match_flow()
|
/linux/net/llc/ |
H A D | llc_sap.c | 369 struct sk_buff *skb1; in llc_do_mcast() local 373 skb1 = skb_clone(skb, GFP_ATOMIC); in llc_do_mcast() 374 if (!skb1) { in llc_do_mcast() 379 llc_sap_rcv(sap, skb1, stack[i]); in llc_do_mcast()
|
/linux/net/core/ |
H A D | skbuff.c | 627 fclones = container_of(skb, struct sk_buff_fclones, skb1); in __alloc_skb() 1062 fclones = container_of(skb, struct sk_buff_fclones, skb1); in kfree_skbmem() 1979 skb1); in skb_clone() 3989 struct sk_buff* skb1, in skb_split_inside_header() argument 3994 skb_copy_from_linear_data_offset(skb, len, skb_put(skb1, pos - len), in skb_split_inside_header() 3998 skb_shinfo(skb1)->frags[i] = skb_shinfo(skb)->frags[i]; in skb_split_inside_header() 4000 skb_shinfo(skb1)->nr_frags = skb_shinfo(skb)->nr_frags; in skb_split_inside_header() 4001 skb1->unreadable = skb->unreadable; in skb_split_inside_header() 4003 skb1->data_len = skb->data_len; in skb_split_inside_header() 4004 skb1->len += skb1->data_len; in skb_split_inside_header() [all …]
|
/linux/net/batman-adv/ |
H A D | send.c | 1057 struct sk_buff *skb1; in batadv_send_outstanding_bcast_packet() local 1076 skb1 = skb_clone(forw_packet->skb, GFP_ATOMIC); in batadv_send_outstanding_bcast_packet() 1077 if (!skb1) in batadv_send_outstanding_bcast_packet() 1080 batadv_send_broadcast_skb(skb1, forw_packet->if_outgoing); in batadv_send_outstanding_bcast_packet()
|
/linux/net/ipv4/ |
H A D | tcp_input.c | 5017 struct sk_buff *skb1; in tcp_data_queue_ofo() local 5080 skb1 = rb_to_skb(parent); in tcp_data_queue_ofo() 5081 if (before(seq, TCP_SKB_CB(skb1)->seq)) { in tcp_data_queue_ofo() 5085 if (before(seq, TCP_SKB_CB(skb1)->end_seq)) { in tcp_data_queue_ofo() 5086 if (!after(end_seq, TCP_SKB_CB(skb1)->end_seq)) { in tcp_data_queue_ofo() 5096 if (after(seq, TCP_SKB_CB(skb1)->seq)) { in tcp_data_queue_ofo() 5098 tcp_dsack_set(sk, seq, TCP_SKB_CB(skb1)->end_seq); in tcp_data_queue_ofo() 5103 rb_replace_node(&skb1->rbnode, &skb->rbnode, in tcp_data_queue_ofo() 5106 TCP_SKB_CB(skb1)->seq, in tcp_data_queue_ofo() 5107 TCP_SKB_CB(skb1)->end_seq); in tcp_data_queue_ofo() [all …]
|
H A D | icmp.c | 382 struct sk_buff *skb1; in icmp_push_reply() local 387 skb_queue_walk(&sk->sk_write_queue, skb1) { in icmp_push_reply() 388 csum = csum_add(csum, skb1->csum); in icmp_push_reply()
|
/linux/net/mptcp/ |
H A D | protocol.c | 215 struct sk_buff *skb1; in mptcp_data_queue_ofo() local 263 skb1 = rb_to_skb(parent); in mptcp_data_queue_ofo() 264 if (before64(seq, MPTCP_SKB_CB(skb1)->map_seq)) { in mptcp_data_queue_ofo() 268 if (before64(seq, MPTCP_SKB_CB(skb1)->end_seq)) { in mptcp_data_queue_ofo() 269 if (!after64(end_seq, MPTCP_SKB_CB(skb1)->end_seq)) { in mptcp_data_queue_ofo() 275 if (after64(seq, MPTCP_SKB_CB(skb1)->map_seq)) { in mptcp_data_queue_ofo() 285 rb_replace_node(&skb1->rbnode, &skb->rbnode, in mptcp_data_queue_ofo() 287 mptcp_drop(sk, skb1); in mptcp_data_queue_ofo() 291 } else if (mptcp_ooo_try_coalesce(msk, skb1, skb)) { in mptcp_data_queue_ofo() 305 while ((skb1 = skb_rb_next(skb)) != NULL) { in mptcp_data_queue_ofo() [all …]
|
/linux/drivers/atm/ |
H A D | iphase.c | 641 struct sk_buff *skb = NULL, *skb1 = NULL; in ia_tx_poll() local 666 skb1 = skb_dequeue(&iavcc->txing_skb); in ia_tx_poll() 667 while (skb1 && (skb1 != skb)) { in ia_tx_poll() 668 if (!(IA_SKB_STATE(skb1) & IA_TX_DONE)) { in ia_tx_poll() 672 if ((vcc->pop) && (skb1->len != 0)) in ia_tx_poll() 674 vcc->pop(vcc, skb1); in ia_tx_poll() 676 (long)skb1);) in ia_tx_poll() 679 dev_kfree_skb_any(skb1); in ia_tx_poll() 680 skb1 = skb_dequeue(&iavcc->txing_skb); in ia_tx_poll() 682 if (!skb1) { in ia_tx_poll()
|
/linux/drivers/net/wireless/ath/ath6kl/ |
H A D | txrx.c | 1315 struct sk_buff *skb1 = NULL; in ath6kl_rx() local 1572 skb1 = skb_copy(skb, GFP_ATOMIC); in ath6kl_rx() 1583 skb1 = skb; in ath6kl_rx() 1590 if (skb1) in ath6kl_rx() 1591 ath6kl_data_tx(skb1, vif->ndev); in ath6kl_rx()
|
/linux/drivers/net/ethernet/qlogic/ |
H A D | qla3xxx.c | 2042 struct sk_buff *skb1 = NULL, *skb2; in ql_process_macip_rx_intr() local 2056 skb1 = lrg_buf_cb1->skb; in ql_process_macip_rx_intr() 2058 if (*((u16 *) skb1->data) != 0xFFFF) in ql_process_macip_rx_intr() 2078 skb_copy_from_linear_data_offset(skb1, VLAN_ID_LEN, in ql_process_macip_rx_intr()
|
/linux/Documentation/networking/ |
H A D | snmp_counter.rst | 725 10 to 15, skb1 has seq 10 to 13, skb2 has seq 14 to 20. The seq 14 and 726 15 in skb2 would be moved to skb1. This operation is 'shift'. If a 727 SACK block acknowledges seq 10 to 20, skb1 has seq 10 to 13, skb2 has 728 seq 14 to 20. All data in skb2 will be moved to skb1, and skb2 will be
|
/linux/drivers/net/vxlan/ |
H A D | vxlan_core.c | 2811 struct sk_buff *skb1; in vxlan_xmit() local 2817 skb1 = skb_clone(skb, GFP_ATOMIC); in vxlan_xmit() 2818 if (skb1) in vxlan_xmit() 2819 vxlan_xmit_one(skb1, dev, vni, rdst, did_rsc); in vxlan_xmit()
|