| /linux/net/x25/ |
| H A D | x25_out.c | 49 struct sk_buff *skbn; in x25_output() local 67 skbn = sock_alloc_send_skb(sk, frontlen + max_len, in x25_output() 70 if (!skbn) { in x25_output() 81 skb_reserve(skbn, frontlen); in x25_output() 86 skb_copy_from_linear_data(skb, skb_put(skbn, len), len); in x25_output() 90 skb_push(skbn, header_len); in x25_output() 91 skb_copy_to_linear_data(skbn, header, header_len); in x25_output() 95 skbn->data[3] |= X25_EXT_M_BIT; in x25_output() 97 skbn->data[2] |= X25_STD_M_BIT; in x25_output() 100 skb_queue_tail(&sk->sk_write_queue, skbn); in x25_output() [all …]
|
| H A D | x25_forward.c | 23 struct sk_buff *skbn; in x25_forward_call() local 71 if ( (skbn = skb_clone(skb, GFP_ATOMIC)) == NULL){ in x25_forward_call() 74 x25_transmit_link(skbn, neigh_new); in x25_forward_call() 94 struct sk_buff *skbn; in x25_forward_data() local 114 if ( (skbn = pskb_copy(skb, GFP_ATOMIC)) == NULL){ in x25_forward_data() 118 x25_transmit_link(skbn, nb); in x25_forward_data()
|
| H A D | x25_in.c | 34 struct sk_buff *skbo, *skbn = skb; in x25_queue_rx_frame() local 51 skbn = alloc_skb(len, GFP_ATOMIC); in x25_queue_rx_frame() 52 if (!skbn) in x25_queue_rx_frame() 57 skb_reset_transport_header(skbn); in x25_queue_rx_frame() 60 skb_copy_from_linear_data(skbo, skb_put(skbn, skbo->len), in x25_queue_rx_frame() 69 skb_put(skbn, skbo->len), in x25_queue_rx_frame() 77 skb_set_owner_r(skbn, sk); in x25_queue_rx_frame() 78 skb_queue_tail(&sk->sk_receive_queue, skbn); in x25_queue_rx_frame()
|
| H A D | x25_link.c | 67 struct sk_buff *skbn; in x25_link_control() local 125 while ((skbn = skb_dequeue(&nb->queue)) != NULL) in x25_link_control() 126 x25_send_frame(skbn, nb); in x25_link_control()
|
| /linux/net/lapb/ |
| H A D | lapb_out.c | 68 struct sk_buff *skb, *skbn; in lapb_kick() local 85 skbn = skb_copy(skb, GFP_ATOMIC); in lapb_kick() 86 if (!skbn) { in lapb_kick() 92 skb_set_owner_w(skbn, skb->sk); in lapb_kick() 97 lapb_send_iframe(lapb, skbn, LAPB_POLLOFF); in lapb_kick()
|
| /linux/drivers/net/wwan/ |
| H A D | mhi_wwan_mbim.c | 308 struct sk_buff *skbn; in mhi_mbim_rx() local 319 skbn = netdev_alloc_skb(link->ndev, dgram_len); in mhi_mbim_rx() 320 if (!skbn) in mhi_mbim_rx() 323 skb_put(skbn, dgram_len); in mhi_mbim_rx() 324 skb_copy_bits(skb, dgram_offset, skbn->data, dgram_len); in mhi_mbim_rx() 326 switch (skbn->data[0] & 0xf0) { in mhi_mbim_rx() 328 skbn->protocol = htons(ETH_P_IP); in mhi_mbim_rx() 331 skbn->protocol = htons(ETH_P_IPV6); in mhi_mbim_rx() 336 dev_kfree_skb_any(skbn); in mhi_mbim_rx() 345 u64_stats_add(&link->rx_bytes, skbn->len); in mhi_mbim_rx() [all …]
|
| /linux/drivers/net/ethernet/qualcomm/rmnet/ |
| H A D | rmnet_handlers.c | 112 struct sk_buff *skbn; in rmnet_map_ingress_handler() local 124 while ((skbn = rmnet_map_deaggregate(skb, port)) != NULL) in rmnet_map_ingress_handler() 125 __rmnet_map_ingress_handler(skbn, port); in rmnet_map_ingress_handler()
|
| /linux/drivers/net/usb/ |
| H A D | qmi_wwan.c | 167 struct sk_buff *skbn; in qmimux_rx_fixup() local 191 skbn = netdev_alloc_skb(net, pkt_len + LL_MAX_HEADER); in qmimux_rx_fixup() 192 if (!skbn) in qmimux_rx_fixup() 199 skb_reset_mac_header(skbn); in qmimux_rx_fixup() 203 skbn->protocol = htons(ETH_P_IP); in qmimux_rx_fixup() 206 skbn->protocol = htons(ETH_P_IPV6); in qmimux_rx_fixup() 210 kfree_skb(skbn); in qmimux_rx_fixup() 214 skb_reserve(skbn, LL_MAX_HEADER); in qmimux_rx_fixup() 215 skb_put_data(skbn, skb->data + offset + qmimux_hdr_sz, pkt_len); in qmimux_rx_fixup() 216 if (netif_rx(skbn) != NET_RX_SUCCESS) { in qmimux_rx_fixup()
|
| /linux/net/qrtr/ |
| H A D | af_qrtr.c | 878 struct sk_buff *skbn; in qrtr_bcast_enqueue() local 882 skbn = pskb_copy(skb, GFP_KERNEL); in qrtr_bcast_enqueue() 883 if (!skbn) in qrtr_bcast_enqueue() 885 skb_set_owner_w(skbn, skb->sk); in qrtr_bcast_enqueue() 886 qrtr_node_enqueue(node, skbn, type, from, to); in qrtr_bcast_enqueue()
|