| /linux/drivers/bluetooth/ |
| H A D | btrsi.c | 51 struct sk_buff *new_skb = NULL; in rsi_hci_send_pkt() local 67 new_skb = skb_realloc_headroom(skb, RSI_HEADROOM_FOR_BT_HAL); in rsi_hci_send_pkt() 68 if (unlikely(!new_skb)) in rsi_hci_send_pkt() 70 bt_cb(new_skb)->pkt_type = hci_skb_pkt_type(skb); in rsi_hci_send_pkt() 72 skb = new_skb; in rsi_hci_send_pkt()
|
| /linux/drivers/net/ethernet/sunplus/ |
| H A D | spl2sw_int.c | 22 struct sk_buff *skb, *new_skb; in spl2sw_rx_poll() local 74 new_skb = netdev_alloc_skb(NULL, comm->rx_desc_buff_size); in spl2sw_rx_poll() 75 if (unlikely(!new_skb)) { in spl2sw_rx_poll() 84 sinfo->mapping = dma_map_single(&comm->pdev->dev, new_skb->data, in spl2sw_rx_poll() 88 dev_kfree_skb_irq(new_skb); in spl2sw_rx_poll() 97 sinfo->skb = new_skb; in spl2sw_rx_poll()
|
| /linux/drivers/net/ethernet/sgi/ |
| H A D | ioc3-eth.c | 126 struct sk_buff *new_skb; in ioc3_alloc_skb() local 130 new_skb = alloc_skb(RX_BUF_SIZE + IOC3_DMA_XFER_LEN - 1, GFP_ATOMIC); in ioc3_alloc_skb() 131 if (!new_skb) in ioc3_alloc_skb() 135 offset = aligned_rx_skb_addr((unsigned long)new_skb->data); in ioc3_alloc_skb() 137 skb_reserve(new_skb, offset); in ioc3_alloc_skb() 139 d = dma_map_single(ip->dma_dev, new_skb->data, in ioc3_alloc_skb() 143 dev_kfree_skb_any(new_skb); in ioc3_alloc_skb() 147 *rxb = (struct ioc3_erxbuf *)new_skb->data; in ioc3_alloc_skb() 148 skb_reserve(new_skb, RX_OFFSET); in ioc3_alloc_skb() 149 *skb = new_skb; in ioc3_alloc_skb() [all …]
|
| /linux/drivers/net/ethernet/natsemi/ |
| H A D | sonic.c | 501 struct sk_buff **new_skb, dma_addr_t *new_addr) in sonic_alloc_rb() argument 503 *new_skb = netdev_alloc_skb(dev, SONIC_RBSIZE + 2); in sonic_alloc_rb() 504 if (!*new_skb) in sonic_alloc_rb() 508 skb_reserve(*new_skb, 2); in sonic_alloc_rb() 510 *new_addr = dma_map_single(lp->device, skb_put(*new_skb, SONIC_RBSIZE), in sonic_alloc_rb() 513 dev_kfree_skb(*new_skb); in sonic_alloc_rb() 514 *new_skb = NULL; in sonic_alloc_rb() 567 struct sk_buff *new_skb; in sonic_rx() local 579 if (sonic_alloc_rb(dev, lp, &new_skb, &new_laddr)) { in sonic_rx() 596 lp->rx_skb[i] = new_skb; in sonic_rx()
|
| /linux/drivers/net/usb/ |
| H A D | sierra_net.c | 774 struct sk_buff *new_skb; in sierra_net_skb_clone() local 777 new_skb = skb_clone(skb, GFP_ATOMIC); in sierra_net_skb_clone() 783 if (new_skb) { in sierra_net_skb_clone() 784 skb_trim(new_skb, len); in sierra_net_skb_clone() 791 return new_skb; in sierra_net_skb_clone() 799 struct sk_buff *new_skb; in sierra_net_rx_fixup() local 841 new_skb = sierra_net_skb_clone(dev, skb, hh.payload_len.word); in sierra_net_rx_fixup() 842 if (new_skb) in sierra_net_rx_fixup() 843 usbnet_skb_return(dev, new_skb); in sierra_net_rx_fixup()
|
| H A D | aqc111.c | 1075 struct sk_buff *new_skb = NULL; in aqc111_rx_fixup() local 1145 new_skb = netdev_alloc_skb_ip_align(dev->net, pkt_len); in aqc111_rx_fixup() 1147 if (!new_skb) in aqc111_rx_fixup() 1150 skb_put(new_skb, pkt_len); in aqc111_rx_fixup() 1151 memcpy(new_skb->data, skb->data, pkt_len); in aqc111_rx_fixup() 1152 skb_pull(new_skb, AQ_RX_HW_PAD); in aqc111_rx_fixup() 1155 aqc111_rx_checksum(new_skb, pkt_desc); in aqc111_rx_fixup() 1159 __vlan_hwaccel_put_tag(new_skb, htons(ETH_P_8021Q), in aqc111_rx_fixup() 1163 usbnet_skb_return(dev, new_skb); in aqc111_rx_fixup() 1172 new_skb = NULL; in aqc111_rx_fixup() [all …]
|
| /linux/net/sctp/ |
| H A D | inqueue.c | 134 goto new_skb; in sctp_inq_pop() 138 goto new_skb; in sctp_inq_pop() 188 new_skb: in sctp_inq_pop()
|
| /linux/drivers/net/ethernet/xilinx/ |
| H A D | xilinx_emaclite.c | 1000 struct sk_buff *new_skb; in xemaclite_send() local 1006 new_skb = orig_skb; in xemaclite_send() 1009 if (xemaclite_send_data(lp, (u8 *)new_skb->data, len) != 0) { in xemaclite_send() 1015 lp->deferred_skb = new_skb; in xemaclite_send() 1017 skb_tx_timestamp(new_skb); in xemaclite_send() 1023 skb_tx_timestamp(new_skb); in xemaclite_send() 1026 dev_consume_skb_any(new_skb); in xemaclite_send()
|
| /linux/net/nfc/ |
| H A D | digital_dep.c | 204 struct sk_buff *new_skb; in digital_send_dep_data_prep() local 209 new_skb = digital_skb_alloc(ddev, ddev->remote_payload_max); in digital_send_dep_data_prep() 210 if (!new_skb) { in digital_send_dep_data_prep() 217 skb_put_data(new_skb, skb->data, ddev->remote_payload_max); in digital_send_dep_data_prep() 224 new_skb = skb; in digital_send_dep_data_prep() 227 return new_skb; in digital_send_dep_data_prep() 238 struct sk_buff *new_skb; in digital_recv_dep_data_gather() local 253 new_skb = skb_copy_expand(ddev->chaining_skb, in digital_recv_dep_data_gather() 258 if (!new_skb) { in digital_recv_dep_data_gather() 264 ddev->chaining_skb = new_skb; in digital_recv_dep_data_gather()
|
| H A D | llcp_core.c | 1416 struct sk_buff *new_skb; in nfc_llcp_recv_agf() 1440 new_skb = nfc_alloc_recv_skb(pdu_len, GFP_KERNEL); in nfc_llcp_recv_agf() 1441 if (new_skb == NULL) { in nfc_llcp_recv_agf() 1446 skb_put_data(new_skb, skb->data, pdu_len); in nfc_llcp_recv_agf() 1448 nfc_llcp_rx_skb(local, new_skb); in nfc_llcp_recv_agf() 1450 kfree_skb(new_skb); in nfc_llcp_recv_agf() 1414 struct sk_buff *new_skb; nfc_llcp_recv_agf() local
|
| /linux/drivers/net/wireless/marvell/mwifiex/ |
| H A D | uap_txrx.c | 87 struct sk_buff *new_skb; in mwifiex_uap_queue_bridged_pkt() local 163 new_skb = in mwifiex_uap_queue_bridged_pkt() 165 if (unlikely(!new_skb)) { in mwifiex_uap_queue_bridged_pkt() 174 skb = new_skb; in mwifiex_uap_queue_bridged_pkt()
|
| H A D | main.c | 872 struct sk_buff *new_skb; in mwifiex_hard_start_xmit() local 897 new_skb = in mwifiex_hard_start_xmit() 899 if (unlikely(!new_skb)) { in mwifiex_hard_start_xmit() 907 skb = new_skb; in mwifiex_hard_start_xmit()
|
| /linux/drivers/net/ethernet/freescale/dpaa/ |
| H A D | dpaa_eth.c | 2140 struct sk_buff *new_skb, *skb = *s; in dpaa_a050385_wa_skb() local 2172 new_skb = netdev_alloc_skb(net_dev, skb->len + DPAA_A050385_ALIGN - 1 + in dpaa_a050385_wa_skb() 2174 if (!new_skb) in dpaa_a050385_wa_skb() 2178 skb_reserve(new_skb, priv->tx_headroom - NET_SKB_PAD); in dpaa_a050385_wa_skb() 2181 start = PTR_ALIGN(new_skb->data, DPAA_A050385_ALIGN); in dpaa_a050385_wa_skb() 2182 if (start - new_skb->data) in dpaa_a050385_wa_skb() 2183 skb_reserve(new_skb, start - new_skb->data); in dpaa_a050385_wa_skb() 2185 skb_put(new_skb, skb->len); in dpaa_a050385_wa_skb() 2186 skb_copy_bits(skb, 0, new_skb->data, skb->len); in dpaa_a050385_wa_skb() 2187 skb_copy_header(new_skb, skb); in dpaa_a050385_wa_skb() [all …]
|
| /linux/drivers/net/ethernet/mediatek/ |
| H A D | mtk_star_emac.c | 1277 struct sk_buff *curr_skb, *new_skb; in mtk_star_rx() local 1291 new_skb = curr_skb; in mtk_star_rx() 1298 new_skb = mtk_star_alloc_skb(ndev); in mtk_star_rx() 1299 if (!new_skb) { in mtk_star_rx() 1301 new_skb = curr_skb; in mtk_star_rx() 1305 new_dma_addr = mtk_star_dma_map_rx(priv, new_skb); in mtk_star_rx() 1308 dev_kfree_skb(new_skb); in mtk_star_rx() 1309 new_skb = curr_skb; in mtk_star_rx() 1332 desc_data.len = skb_tailroom(new_skb); in mtk_star_rx() 1333 desc_data.skb = new_skb; in mtk_star_rx()
|
| /linux/drivers/net/ethernet/sun/ |
| H A D | sunbmac.c | 825 struct sk_buff *new_skb; in bigmac_rx() local 828 new_skb = big_mac_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); in bigmac_rx() 829 if (new_skb == NULL) { in bigmac_rx() 837 bp->rx_skbs[elem] = new_skb; in bigmac_rx() 838 skb_put(new_skb, ETH_FRAME_LEN); in bigmac_rx() 839 skb_reserve(new_skb, 34); in bigmac_rx() 842 new_skb->data, in bigmac_rx()
|
| H A D | sungem.c | 806 struct sk_buff *new_skb; in gem_rx() local 808 new_skb = gem_alloc_skb(dev, RX_BUF_ALLOC_SIZE(gp), GFP_ATOMIC); in gem_rx() 809 if (new_skb == NULL) { in gem_rx() 815 gp->rx_skbs[entry] = new_skb; in gem_rx() 816 skb_put(new_skb, (gp->rx_buf_sz + RX_OFFSET)); in gem_rx() 818 virt_to_page(new_skb->data), in gem_rx() 819 offset_in_page(new_skb->data), in gem_rx() 822 skb_reserve(new_skb, RX_OFFSET); in gem_rx()
|
| H A D | sunhme.c | 1795 struct sk_buff *new_skb; in happy_meal_rx() local 1799 new_skb = happy_meal_alloc_skb(RX_BUF_ALLOC_SIZE, GFP_ATOMIC); in happy_meal_rx() 1800 if (new_skb == NULL) { in happy_meal_rx() 1804 skb_put(new_skb, (ETH_FRAME_LEN + RX_OFFSET + 4)); in happy_meal_rx() 1805 mapping = dma_map_single(hp->dma_dev, new_skb->data, in happy_meal_rx() 1809 dev_kfree_skb_any(new_skb); in happy_meal_rx() 1815 hp->rx_skbs[elem] = new_skb; in happy_meal_rx() 1819 skb_reserve(new_skb, RX_OFFSET); in happy_meal_rx()
|
| /linux/drivers/net/ethernet/3com/ |
| H A D | typhoon.c | 1632 struct sk_buff *skb, *new_skb; in typhoon_rx() local 1662 (new_skb = netdev_alloc_skb(tp->dev, pkt_len + 2)) != NULL) { in typhoon_rx() 1663 skb_reserve(new_skb, 2); in typhoon_rx() 1666 skb_copy_to_linear_data(new_skb, skb->data, pkt_len); in typhoon_rx() 1670 skb_put(new_skb, pkt_len); in typhoon_rx() 1673 new_skb = skb; in typhoon_rx() 1674 skb_put(new_skb, pkt_len); in typhoon_rx() 1679 new_skb->protocol = eth_type_trans(new_skb, tp->dev); in typhoon_rx() 1686 new_skb->ip_summed = CHECKSUM_UNNECESSARY; in typhoon_rx() 1688 skb_checksum_none_assert(new_skb); in typhoon_rx() [all …]
|
| /linux/drivers/net/ethernet/actions/ |
| H A D | owl-emac.c | 752 struct sk_buff *curr_skb, *new_skb; in owl_emac_rx_process() local 815 new_skb = owl_emac_alloc_skb(netdev); in owl_emac_rx_process() 816 if (unlikely(!new_skb)) in owl_emac_rx_process() 819 new_dma = owl_emac_dma_map_rx(priv, new_skb); in owl_emac_rx_process() 821 dev_kfree_skb(new_skb); in owl_emac_rx_process() 844 new_skb = curr_skb; in owl_emac_rx_process() 850 ring->skbs[ring->head] = new_skb; in owl_emac_rx_process()
|
| /linux/drivers/net/wireless/ath/ath6kl/ |
| H A D | txrx.c | 993 struct sk_buff *new_skb; in aggr_slice_amsdu() local 1014 new_skb = aggr_get_free_skb(p_aggr); in aggr_slice_amsdu() 1015 if (!new_skb) { in aggr_slice_amsdu() 1020 memcpy(new_skb->data, framep, frame_8023_len); in aggr_slice_amsdu() 1021 skb_put(new_skb, frame_8023_len); in aggr_slice_amsdu() 1022 if (ath6kl_wmi_dot3_2_dix(new_skb)) { in aggr_slice_amsdu() 1024 dev_kfree_skb(new_skb); in aggr_slice_amsdu() 1028 skb_queue_tail(&rxtid->q, new_skb); in aggr_slice_amsdu()
|
| /linux/drivers/net/ethernet/amd/ |
| H A D | amd8111e.c | 687 struct sk_buff *skb, *new_skb; in amd8111e_rx_poll() local 733 new_skb = netdev_alloc_skb(dev, lp->rx_buff_len); in amd8111e_rx_poll() 734 if (!new_skb) { in amd8111e_rx_poll() 743 skb_reserve(new_skb, 2); in amd8111e_rx_poll() 748 lp->rx_skbuff[rx_index] = new_skb; in amd8111e_rx_poll() 750 new_skb->data, in amd8111e_rx_poll()
|
| /linux/drivers/net/ethernet/hisilicon/hns/ |
| H A D | hns_enet.c | 214 struct sk_buff *new_skb = NULL; in hns_nic_maybe_stop_tx() local 224 new_skb = skb_copy(skb, GFP_ATOMIC); in hns_nic_maybe_stop_tx() 225 if (!new_skb) in hns_nic_maybe_stop_tx() 229 *out_skb = new_skb; in hns_nic_maybe_stop_tx() 247 struct sk_buff *new_skb = NULL; in hns_nic_maybe_stop_tso() local 265 new_skb = skb_copy(skb, GFP_ATOMIC); in hns_nic_maybe_stop_tso() 266 if (!new_skb) in hns_nic_maybe_stop_tso() 269 *out_skb = new_skb; in hns_nic_maybe_stop_tso()
|
| /linux/drivers/net/ethernet/rdc/ |
| H A D | r6040.c | 513 struct sk_buff *skb_ptr, *new_skb; in r6040_rx() local 545 new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE); in r6040_rx() 546 if (!new_skb) { in r6040_rx() 565 descptr->skb_ptr = new_skb; in r6040_rx()
|
| /linux/drivers/net/ethernet/realtek/ |
| H A D | 8139cp.c | 475 struct sk_buff *skb, *new_skb; in cp_rx_poll() local 510 new_skb = napi_alloc_skb(napi, buflen); in cp_rx_poll() 511 if (!new_skb) { in cp_rx_poll() 516 new_mapping = dma_map_single(&cp->pdev->dev, new_skb->data, buflen, in cp_rx_poll() 520 kfree_skb(new_skb); in cp_rx_poll() 535 cp->rx_skb[rx_tail] = new_skb; in cp_rx_poll()
|
| /linux/drivers/net/ethernet/ibm/ |
| H A D | ibmveth.c | 1470 struct sk_buff *skb, *new_skb; in ibmveth_poll() local 1492 new_skb = NULL; in ibmveth_poll() 1494 new_skb = netdev_alloc_skb(netdev, length); in ibmveth_poll() 1496 if (new_skb) { in ibmveth_poll() 1497 skb_copy_to_linear_data(new_skb, in ibmveth_poll() 1505 skb = new_skb; in ibmveth_poll()
|