Searched refs:skb_out (Results 1 – 9 of 9) sorted by relevance
| /linux/net/hsr/ |
| H A D | hsr_netlink.c | 316 struct sk_buff *skb_out; in hsr_get_node_status() local 347 skb_out = genlmsg_new(NLMSG_GOODSIZE, GFP_ATOMIC); in hsr_get_node_status() 348 if (!skb_out) { in hsr_get_node_status() 353 msg_head = genlmsg_put(skb_out, NETLINK_CB(skb_in).portid, in hsr_get_node_status() 361 res = nla_put_u32(skb_out, HSR_A_IFINDEX, hsr_dev->ifindex); in hsr_get_node_status() 378 res = nla_put(skb_out, HSR_A_NODE_ADDR, ETH_ALEN, in hsr_get_node_status() 384 res = nla_put(skb_out, HSR_A_NODE_ADDR_B, ETH_ALEN, in hsr_get_node_status() 389 res = nla_put_u32(skb_out, HSR_A_ADDR_B_IFINDEX, in hsr_get_node_status() 395 res = nla_put_u32(skb_out, HSR_A_IF1_AGE, hsr_node_if1_age); in hsr_get_node_status() 398 res = nla_put_u16(skb_out, HSR_A_IF1_SEQ, hsr_node_if1_seq); in hsr_get_node_status() [all …]
|
| /linux/net/batman-adv/ |
| H A D | fragmentation.c | 251 struct sk_buff *skb_out; in batadv_frag_merge_packets() local 260 skb_out = entry->skb; in batadv_frag_merge_packets() 263 packet = (struct batadv_frag_packet *)skb_out->data; in batadv_frag_merge_packets() 267 if (pskb_expand_head(skb_out, 0, size - skb_out->len, GFP_ATOMIC) < 0) { in batadv_frag_merge_packets() 268 kfree_skb(skb_out); in batadv_frag_merge_packets() 269 skb_out = NULL; in batadv_frag_merge_packets() 277 skb_pull(skb_out, hdr_size); in batadv_frag_merge_packets() 278 skb_out->ip_summed = CHECKSUM_NONE; in batadv_frag_merge_packets() 279 memmove(skb_out->data - ETH_HLEN, skb_mac_header(skb_out), ETH_HLEN); in batadv_frag_merge_packets() 280 skb_set_mac_header(skb_out, -ETH_HLEN); in batadv_frag_merge_packets() [all …]
|
| /linux/drivers/net/usb/ |
| H A D | cdc_ncm.c | 1215 struct sk_buff *skb_out; in cdc_ncm_fill_tx_frame() local 1241 skb_out = ctx->tx_curr_skb; in cdc_ncm_fill_tx_frame() 1244 if (!skb_out) { in cdc_ncm_fill_tx_frame() 1247 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame() 1253 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame() 1262 if (skb_out == NULL) { in cdc_ncm_fill_tx_frame() 1272 skb_out = alloc_skb(ctx->tx_curr_size, GFP_ATOMIC); in cdc_ncm_fill_tx_frame() 1275 if (!skb_out) in cdc_ncm_fill_tx_frame() 1281 nth.nth16 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth16)); in cdc_ncm_fill_tx_frame() 1287 nth.nth32 = skb_put_zero(skb_out, sizeof(struct usb_cdc_ncm_nth32)); in cdc_ncm_fill_tx_frame() [all …]
|
| H A D | cdc_mbim.c | 223 struct sk_buff *skb_out; in cdc_mbim_tx_fixup() local 294 skb_out = cdc_ncm_fill_tx_frame(dev, skb, sign); in cdc_mbim_tx_fixup() 296 return skb_out; in cdc_mbim_tx_fixup()
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_net.c | 430 int skb_out; in rxe_skb_tx_dtor() local 432 skb_out = atomic_dec_return(&qp->skb_out); in rxe_skb_tx_dtor() 434 skb_out < RXE_INFLIGHT_SKBS_PER_QP_LOW)) in rxe_skb_tx_dtor() 450 atomic_inc(&pkt->qp->skb_out); in rxe_send() 473 atomic_inc(&pkt->qp->skb_out); in rxe_loopback()
|
| H A D | rxe_req.c | 722 if (unlikely(atomic_read(&qp->skb_out) > in rxe_requester()
|
| H A D | rxe_qp.c | 231 atomic_set(&qp->skb_out, 0); in rxe_qp_init_misc()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_rx.c | 213 goto skb_out; in rx_alloc_pkts() 221 goto skb_out; in rx_alloc_pkts() 229 skb_out: in rx_alloc_pkts()
|
| /linux/net/key/ |
| H A D | af_key.c | 2759 struct sk_buff *skb_out; in key_notify_policy_flush() local 2762 skb_out = alloc_skb(sizeof(struct sadb_msg) + 16, GFP_ATOMIC); in key_notify_policy_flush() 2763 if (!skb_out) in key_notify_policy_flush() 2765 hdr = skb_put(skb_out, sizeof(struct sadb_msg)); in key_notify_policy_flush() 2774 pfkey_broadcast(skb_out, GFP_ATOMIC, BROADCAST_ALL, NULL, c->net); in key_notify_policy_flush()
|