/linux/net/tipc/ |
H A D | msg.c | 256 bool tipc_msg_validate(struct sk_buff **_skb) in tipc_msg_validate() argument 258 struct sk_buff *skb = *_skb; in tipc_msg_validate() 267 kfree_skb(*_skb); in tipc_msg_validate() 268 *_skb = skb; in tipc_msg_validate() 315 struct sk_buff *_skb; in tipc_msg_fragment() local 335 _skb = tipc_buf_acquire(INT_H_SIZE + eat, GFP_ATOMIC); in tipc_msg_fragment() 336 if (!_skb) in tipc_msg_fragment() 338 skb_orphan(_skb); in tipc_msg_fragment() 339 __skb_queue_tail(frags, _skb); in tipc_msg_fragment() 341 skb_copy_to_linear_data(_skb, hdr, INT_H_SIZE); in tipc_msg_fragment() [all …]
|
H A D | bcast.c | 181 struct sk_buff *skb, *_skb; in tipc_bcbase_xmit() local 201 _skb = pskb_copy_for_clone(skb, GFP_ATOMIC); in tipc_bcbase_xmit() 202 if (!_skb) in tipc_bcbase_xmit() 204 __skb_queue_tail(&_xmitq, _skb); in tipc_bcbase_xmit() 324 struct sk_buff *_skb; in tipc_mcast_send_sync() local 338 _skb = tipc_buf_acquire(MCAST_H_SIZE, GFP_KERNEL); in tipc_mcast_send_sync() 339 if (!_skb) in tipc_mcast_send_sync() 346 skb_copy_to_linear_data(_skb, hdr, MCAST_H_SIZE); in tipc_mcast_send_sync() 347 skb_orphan(_skb); in tipc_mcast_send_sync() 350 _hdr = buf_msg(_skb); in tipc_mcast_send_sync() [all …]
|
H A D | msg.h | 1156 bool tipc_msg_validate(struct sk_buff **_skb); 1246 struct sk_buff *_skb, *tmp, *skb = NULL; in tipc_skb_dequeue() local 1249 skb_queue_walk_safe(list, _skb, tmp) { in tipc_skb_dequeue() 1250 if (msg_destport(buf_msg(_skb)) == dport) { in tipc_skb_dequeue() 1251 __skb_unlink(_skb, list); in tipc_skb_dequeue() 1252 skb = _skb; in tipc_skb_dequeue()
|
H A D | link.c | 1019 struct sk_buff *skb, *_skb; in tipc_link_xmit() local 1066 _skb = skb_clone(skb, GFP_ATOMIC); in tipc_link_xmit() 1067 if (!_skb) { in tipc_link_xmit() 1074 __skb_queue_tail(xmitq, _skb); in tipc_link_xmit() 1153 struct sk_buff *skb, *_skb; in tipc_link_advance_backlog() local 1164 _skb = skb_clone(skb, GFP_ATOMIC); in tipc_link_advance_backlog() 1165 if (!_skb) in tipc_link_advance_backlog() 1176 __skb_queue_tail(xmitq, _skb); in tipc_link_advance_backlog() 1542 struct sk_buff *skb, *_skb, *tmp; in tipc_link_advance_transmq() local 1627 _skb = pskb_copy(skb, GFP_ATOMIC); in tipc_link_advance_transmq() [all …]
|
H A D | udp_media.c | 263 struct sk_buff *_skb; in tipc_udp_send_msg() local 265 _skb = pskb_copy(skb, GFP_ATOMIC); in tipc_udp_send_msg() 266 if (!_skb) { in tipc_udp_send_msg() 271 err = tipc_udp_xmit(net, _skb, ub, src, &rcast->addr, in tipc_udp_send_msg()
|
H A D | group.c | 476 struct sk_buff *_skb, *tmp; in tipc_group_sort_msg() local 481 skb_queue_walk_safe(defq, _skb, tmp) { in tipc_group_sort_msg() 482 _hdr = buf_msg(_skb); in tipc_group_sort_msg() 485 __skb_queue_before(defq, _skb, skb); in tipc_group_sort_msg()
|
H A D | bearer.c | 766 struct sk_buff *skb, *_skb; in tipc_clone_to_loopback() local 769 skb_queue_walk(pkts, _skb) { in tipc_clone_to_loopback() 770 skb = pskb_copy(_skb, GFP_ATOMIC); in tipc_clone_to_loopback()
|
H A D | socket.c | 1194 struct sk_buff *skb, *_skb; in tipc_sk_mcast_rcv() local 1245 _skb = __pskb_copy(skb, hlen, GFP_ATOMIC); in tipc_sk_mcast_rcv() 1246 if (_skb) { in tipc_sk_mcast_rcv() 1247 msg_set_destport(buf_msg(_skb), portid); in tipc_sk_mcast_rcv() 1248 __skb_queue_tail(&tmpq, _skb); in tipc_sk_mcast_rcv()
|
/linux/drivers/net/ipvlan/ |
H A D | ipvlan.h | 107 #define IPVL_SKB_CB(_skb) ((struct ipvl_skb_cb *)&((_skb)->cb[0])) argument
|
/linux/include/linux/soc/mediatek/ |
H A D | mtk_wed.h | 290 #define mtk_wed_device_ppe_check(_dev, _skb, _reason, _hash) \ argument 291 (_dev)->ops->ppe_check(_dev, _skb, _reason, _hash) 321 #define mtk_wed_device_ppe_check(_dev, _skb, _reason, _hash) do {} while (0) argument
|