Lines Matching +full:tx +full:- +full:rings +full:- +full:empty
5 * This program is dual-licensed; you may select either version 2 of
22 #include "dwc-xlgmac.h"
23 #include "dwc-xlgmac-reg.h"
30 return (ring->dma_desc_count - (ring->cur - ring->dirty)); in xlgmac_tx_avail_desc()
35 return (ring->cur - ring->dirty); in xlgmac_rx_dirty_desc()
43 struct xlgmac_pdata *pdata = channel->pdata; in xlgmac_maybe_stop_tx_queue()
46 netif_info(pdata, drv, pdata->netdev, in xlgmac_maybe_stop_tx_queue()
47 "Tx queue stopped, not enough descriptors available\n"); in xlgmac_maybe_stop_tx_queue()
48 netif_stop_subqueue(pdata->netdev, channel->queue_index); in xlgmac_maybe_stop_tx_queue()
49 ring->tx.queue_stopped = 1; in xlgmac_maybe_stop_tx_queue()
54 if (ring->tx.xmit_more) in xlgmac_maybe_stop_tx_queue()
55 pdata->hw_ops.tx_start_xmit(channel, ring); in xlgmac_maybe_stop_tx_queue()
67 pkt_info->vlan_ctag = skb_vlan_tag_get(skb); in xlgmac_prep_vlan()
75 if (!XLGMAC_GET_REG_BITS(pkt_info->attributes, in xlgmac_prep_tso()
84 pkt_info->header_len = skb_tcp_all_headers(skb); in xlgmac_prep_tso()
85 pkt_info->tcp_header_len = tcp_hdrlen(skb); in xlgmac_prep_tso()
86 pkt_info->tcp_payload_len = skb->len - pkt_info->header_len; in xlgmac_prep_tso()
87 pkt_info->mss = skb_shinfo(skb)->gso_size; in xlgmac_prep_tso()
89 XLGMAC_PR("header_len=%u\n", pkt_info->header_len); in xlgmac_prep_tso()
91 pkt_info->tcp_header_len, pkt_info->tcp_payload_len); in xlgmac_prep_tso()
92 XLGMAC_PR("mss=%u\n", pkt_info->mss); in xlgmac_prep_tso()
97 pkt_info->tx_packets = skb_shinfo(skb)->gso_segs; in xlgmac_prep_tso()
98 pkt_info->tx_bytes += (pkt_info->tx_packets - 1) * pkt_info->header_len; in xlgmac_prep_tso()
105 if (skb->ip_summed != CHECKSUM_PARTIAL) in xlgmac_is_tso()
124 pkt_info->skb = skb; in xlgmac_prep_tx_pkt()
127 pkt_info->desc_count = 0; in xlgmac_prep_tx_pkt()
129 pkt_info->tx_packets = 1; in xlgmac_prep_tx_pkt()
130 pkt_info->tx_bytes = skb->len; in xlgmac_prep_tx_pkt()
134 if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) { in xlgmac_prep_tx_pkt()
136 pkt_info->desc_count++; in xlgmac_prep_tx_pkt()
140 pkt_info->desc_count++; in xlgmac_prep_tx_pkt()
142 pkt_info->attributes = XLGMAC_SET_REG_BITS( in xlgmac_prep_tx_pkt()
143 pkt_info->attributes, in xlgmac_prep_tx_pkt()
147 pkt_info->attributes = XLGMAC_SET_REG_BITS( in xlgmac_prep_tx_pkt()
148 pkt_info->attributes, in xlgmac_prep_tx_pkt()
152 } else if (skb->ip_summed == CHECKSUM_PARTIAL) in xlgmac_prep_tx_pkt()
153 pkt_info->attributes = XLGMAC_SET_REG_BITS( in xlgmac_prep_tx_pkt()
154 pkt_info->attributes, in xlgmac_prep_tx_pkt()
161 if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag) in xlgmac_prep_tx_pkt()
165 pkt_info->desc_count++; in xlgmac_prep_tx_pkt()
168 pkt_info->attributes = XLGMAC_SET_REG_BITS( in xlgmac_prep_tx_pkt()
169 pkt_info->attributes, in xlgmac_prep_tx_pkt()
176 pkt_info->desc_count++; in xlgmac_prep_tx_pkt()
177 len -= min_t(unsigned int, len, XLGMAC_TX_MAX_BUF_SIZE); in xlgmac_prep_tx_pkt()
180 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { in xlgmac_prep_tx_pkt()
181 frag = &skb_shinfo(skb)->frags[i]; in xlgmac_prep_tx_pkt()
183 pkt_info->desc_count++; in xlgmac_prep_tx_pkt()
184 len -= min_t(unsigned int, len, XLGMAC_TX_MAX_BUF_SIZE); in xlgmac_prep_tx_pkt()
195 return -EINVAL; in xlgmac_calc_rx_buf_size()
201 rx_buf_size = (rx_buf_size + XLGMAC_RX_BUF_ALIGN - 1) & in xlgmac_calc_rx_buf_size()
202 ~(XLGMAC_RX_BUF_ALIGN - 1); in xlgmac_calc_rx_buf_size()
209 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_enable_rx_tx_ints()
214 channel = pdata->channel_head; in xlgmac_enable_rx_tx_ints()
215 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_enable_rx_tx_ints()
216 if (channel->tx_ring && channel->rx_ring) in xlgmac_enable_rx_tx_ints()
218 else if (channel->tx_ring) in xlgmac_enable_rx_tx_ints()
220 else if (channel->rx_ring) in xlgmac_enable_rx_tx_ints()
225 hw_ops->enable_int(channel, int_id); in xlgmac_enable_rx_tx_ints()
231 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_disable_rx_tx_ints()
236 channel = pdata->channel_head; in xlgmac_disable_rx_tx_ints()
237 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_disable_rx_tx_ints()
238 if (channel->tx_ring && channel->rx_ring) in xlgmac_disable_rx_tx_ints()
240 else if (channel->tx_ring) in xlgmac_disable_rx_tx_ints()
242 else if (channel->rx_ring) in xlgmac_disable_rx_tx_ints()
247 hw_ops->disable_int(channel, int_id); in xlgmac_disable_rx_tx_ints()
259 hw_ops = &pdata->hw_ops; in xlgmac_isr()
263 * this register to be non-zero in xlgmac_isr()
265 dma_isr = readl(pdata->mac_regs + DMA_ISR); in xlgmac_isr()
269 netif_dbg(pdata, intr, pdata->netdev, "DMA_ISR=%#010x\n", dma_isr); in xlgmac_isr()
271 for (i = 0; i < pdata->channel_count; i++) { in xlgmac_isr()
275 channel = pdata->channel_head + i; in xlgmac_isr()
278 netif_dbg(pdata, intr, pdata->netdev, "DMA_CH%u_ISR=%#010x\n", in xlgmac_isr()
289 if (!pdata->per_channel_irq && (ti || ri)) { in xlgmac_isr()
290 if (napi_schedule_prep(&pdata->napi)) { in xlgmac_isr()
291 /* Disable Tx and Rx interrupts */ in xlgmac_isr()
294 pdata->stats.napi_poll_isr++; in xlgmac_isr()
296 __napi_schedule_irqoff(&pdata->napi); in xlgmac_isr()
302 pdata->stats.tx_process_stopped++; in xlgmac_isr()
306 pdata->stats.rx_process_stopped++; in xlgmac_isr()
310 pdata->stats.tx_buffer_unavailable++; in xlgmac_isr()
314 pdata->stats.rx_buffer_unavailable++; in xlgmac_isr()
319 pdata->stats.fatal_bus_error++; in xlgmac_isr()
320 schedule_work(&pdata->restart_work); in xlgmac_isr()
329 mac_isr = readl(pdata->mac_regs + MAC_ISR); in xlgmac_isr()
333 hw_ops->tx_mmc_int(pdata); in xlgmac_isr()
337 hw_ops->rx_mmc_int(pdata); in xlgmac_isr()
350 if (napi_schedule_prep(&channel->napi)) { in xlgmac_dma_isr()
351 /* Disable Tx and Rx interrupts */ in xlgmac_dma_isr()
352 disable_irq_nosync(channel->dma_irq); in xlgmac_dma_isr()
355 __napi_schedule_irqoff(&channel->napi); in xlgmac_dma_isr()
365 struct xlgmac_pdata *pdata = channel->pdata; in xlgmac_tx_timer()
368 napi = (pdata->per_channel_irq) ? &channel->napi : &pdata->napi; in xlgmac_tx_timer()
371 /* Disable Tx and Rx interrupts */ in xlgmac_tx_timer()
372 if (pdata->per_channel_irq) in xlgmac_tx_timer()
373 disable_irq_nosync(channel->dma_irq); in xlgmac_tx_timer()
377 pdata->stats.napi_poll_txtimer++; in xlgmac_tx_timer()
382 channel->tx_timer_active = 0; in xlgmac_tx_timer()
390 channel = pdata->channel_head; in xlgmac_init_timers()
391 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_init_timers()
392 if (!channel->tx_ring) in xlgmac_init_timers()
395 timer_setup(&channel->tx_timer, xlgmac_tx_timer, 0); in xlgmac_init_timers()
404 channel = pdata->channel_head; in xlgmac_stop_timers()
405 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_stop_timers()
406 if (!channel->tx_ring) in xlgmac_stop_timers()
409 timer_delete_sync(&channel->tx_timer); in xlgmac_stop_timers()
418 if (pdata->per_channel_irq) { in xlgmac_napi_enable()
419 channel = pdata->channel_head; in xlgmac_napi_enable()
420 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_napi_enable()
422 netif_napi_add(pdata->netdev, &channel->napi, in xlgmac_napi_enable()
425 napi_enable(&channel->napi); in xlgmac_napi_enable()
429 netif_napi_add(pdata->netdev, &pdata->napi, in xlgmac_napi_enable()
432 napi_enable(&pdata->napi); in xlgmac_napi_enable()
441 if (pdata->per_channel_irq) { in xlgmac_napi_disable()
442 channel = pdata->channel_head; in xlgmac_napi_disable()
443 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_napi_disable()
444 napi_disable(&channel->napi); in xlgmac_napi_disable()
447 netif_napi_del(&channel->napi); in xlgmac_napi_disable()
450 napi_disable(&pdata->napi); in xlgmac_napi_disable()
453 netif_napi_del(&pdata->napi); in xlgmac_napi_disable()
459 struct net_device *netdev = pdata->netdev; in xlgmac_request_irqs()
464 ret = devm_request_irq(pdata->dev, pdata->dev_irq, xlgmac_isr, in xlgmac_request_irqs()
465 IRQF_SHARED, netdev->name, pdata); in xlgmac_request_irqs()
468 pdata->dev_irq); in xlgmac_request_irqs()
472 if (!pdata->per_channel_irq) in xlgmac_request_irqs()
475 channel = pdata->channel_head; in xlgmac_request_irqs()
476 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_request_irqs()
477 snprintf(channel->dma_irq_name, in xlgmac_request_irqs()
478 sizeof(channel->dma_irq_name) - 1, in xlgmac_request_irqs()
479 "%s-TxRx-%u", netdev_name(netdev), in xlgmac_request_irqs()
480 channel->queue_index); in xlgmac_request_irqs()
482 ret = devm_request_irq(pdata->dev, channel->dma_irq, in xlgmac_request_irqs()
484 channel->dma_irq_name, channel); in xlgmac_request_irqs()
487 channel->dma_irq); in xlgmac_request_irqs()
496 for (i--, channel--; i < pdata->channel_count; i--, channel--) in xlgmac_request_irqs()
497 devm_free_irq(pdata->dev, channel->dma_irq, channel); in xlgmac_request_irqs()
499 devm_free_irq(pdata->dev, pdata->dev_irq, pdata); in xlgmac_request_irqs()
509 devm_free_irq(pdata->dev, pdata->dev_irq, pdata); in xlgmac_free_irqs()
511 if (!pdata->per_channel_irq) in xlgmac_free_irqs()
514 channel = pdata->channel_head; in xlgmac_free_irqs()
515 for (i = 0; i < pdata->channel_count; i++, channel++) in xlgmac_free_irqs()
516 devm_free_irq(pdata->dev, channel->dma_irq, channel); in xlgmac_free_irqs()
521 struct xlgmac_desc_ops *desc_ops = &pdata->desc_ops; in xlgmac_free_tx_data()
527 channel = pdata->channel_head; in xlgmac_free_tx_data()
528 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_free_tx_data()
529 ring = channel->tx_ring; in xlgmac_free_tx_data()
533 for (j = 0; j < ring->dma_desc_count; j++) { in xlgmac_free_tx_data()
535 desc_ops->unmap_desc_data(pdata, desc_data); in xlgmac_free_tx_data()
542 struct xlgmac_desc_ops *desc_ops = &pdata->desc_ops; in xlgmac_free_rx_data()
548 channel = pdata->channel_head; in xlgmac_free_rx_data()
549 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_free_rx_data()
550 ring = channel->rx_ring; in xlgmac_free_rx_data()
554 for (j = 0; j < ring->dma_desc_count; j++) { in xlgmac_free_rx_data()
556 desc_ops->unmap_desc_data(pdata, desc_data); in xlgmac_free_rx_data()
563 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_start()
564 struct net_device *netdev = pdata->netdev; in xlgmac_start()
567 hw_ops->init(pdata); in xlgmac_start()
574 hw_ops->enable_tx(pdata); in xlgmac_start()
575 hw_ops->enable_rx(pdata); in xlgmac_start()
582 hw_ops->exit(pdata); in xlgmac_start()
589 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_stop()
590 struct net_device *netdev = pdata->netdev; in xlgmac_stop()
597 hw_ops->disable_tx(pdata); in xlgmac_stop()
598 hw_ops->disable_rx(pdata); in xlgmac_stop()
601 hw_ops->exit(pdata); in xlgmac_stop()
603 channel = pdata->channel_head; in xlgmac_stop()
604 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_stop()
605 if (!channel->tx_ring) in xlgmac_stop()
608 txq = netdev_get_tx_queue(netdev, channel->queue_index); in xlgmac_stop()
616 if (!netif_running(pdata->netdev)) in xlgmac_restart_dev()
646 desc_ops = &pdata->desc_ops; in xlgmac_open()
650 /* Calculate the Rx buffer size before allocating rings */ in xlgmac_open()
651 ret = xlgmac_calc_rx_buf_size(netdev, netdev->mtu); in xlgmac_open()
654 pdata->rx_buf_size = ret; in xlgmac_open()
656 /* Allocate the channels and rings */ in xlgmac_open()
657 ret = desc_ops->alloc_channels_and_rings(pdata); in xlgmac_open()
661 INIT_WORK(&pdata->restart_work, xlgmac_restart); in xlgmac_open()
671 desc_ops->free_channels_and_rings(pdata); in xlgmac_open()
681 desc_ops = &pdata->desc_ops; in xlgmac_close()
686 /* Free the channels and rings */ in xlgmac_close()
687 desc_ops->free_channels_and_rings(pdata); in xlgmac_close()
696 netdev_warn(netdev, "tx timeout, device restarting\n"); in xlgmac_tx_timeout()
697 schedule_work(&pdata->restart_work); in xlgmac_tx_timeout()
711 desc_ops = &pdata->desc_ops; in xlgmac_xmit()
712 hw_ops = &pdata->hw_ops; in xlgmac_xmit()
714 XLGMAC_PR("skb->len = %d\n", skb->len); in xlgmac_xmit()
716 channel = pdata->channel_head + skb->queue_mapping; in xlgmac_xmit()
717 txq = netdev_get_tx_queue(netdev, channel->queue_index); in xlgmac_xmit()
718 ring = channel->tx_ring; in xlgmac_xmit()
719 tx_pkt_info = &ring->pkt_info; in xlgmac_xmit()
721 if (skb->len == 0) { in xlgmac_xmit()
723 "empty skb received from stack\n"); in xlgmac_xmit()
728 /* Prepare preliminary packet info for TX */ in xlgmac_xmit()
734 tx_pkt_info->desc_count); in xlgmac_xmit()
747 if (!desc_ops->map_tx_skb(channel, skb)) { in xlgmac_xmit()
753 netdev_tx_sent_queue(txq, tx_pkt_info->tx_bytes); in xlgmac_xmit()
756 hw_ops->dev_xmit(channel); in xlgmac_xmit()
771 struct xlgmac_stats *pstats = &pdata->stats; in xlgmac_get_stats64()
773 pdata->hw_ops.read_mmc_stats(pdata); in xlgmac_get_stats64()
775 s->rx_packets = pstats->rxframecount_gb; in xlgmac_get_stats64()
776 s->rx_bytes = pstats->rxoctetcount_gb; in xlgmac_get_stats64()
777 s->rx_errors = pstats->rxframecount_gb - in xlgmac_get_stats64()
778 pstats->rxbroadcastframes_g - in xlgmac_get_stats64()
779 pstats->rxmulticastframes_g - in xlgmac_get_stats64()
780 pstats->rxunicastframes_g; in xlgmac_get_stats64()
781 s->multicast = pstats->rxmulticastframes_g; in xlgmac_get_stats64()
782 s->rx_length_errors = pstats->rxlengtherror; in xlgmac_get_stats64()
783 s->rx_crc_errors = pstats->rxcrcerror; in xlgmac_get_stats64()
784 s->rx_fifo_errors = pstats->rxfifooverflow; in xlgmac_get_stats64()
786 s->tx_packets = pstats->txframecount_gb; in xlgmac_get_stats64()
787 s->tx_bytes = pstats->txoctetcount_gb; in xlgmac_get_stats64()
788 s->tx_errors = pstats->txframecount_gb - pstats->txframecount_g; in xlgmac_get_stats64()
789 s->tx_dropped = netdev->stats.tx_dropped; in xlgmac_get_stats64()
795 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_set_mac_address()
798 if (!is_valid_ether_addr(saddr->sa_data)) in xlgmac_set_mac_address()
799 return -EADDRNOTAVAIL; in xlgmac_set_mac_address()
801 eth_hw_addr_set(netdev, saddr->sa_data); in xlgmac_set_mac_address()
803 hw_ops->set_mac_address(pdata, netdev->dev_addr); in xlgmac_set_mac_address()
812 return -ENODEV; in xlgmac_ioctl()
826 pdata->rx_buf_size = ret; in xlgmac_change_mtu()
827 WRITE_ONCE(netdev->mtu, mtu); in xlgmac_change_mtu()
839 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_vlan_rx_add_vid()
841 set_bit(vid, pdata->active_vlans); in xlgmac_vlan_rx_add_vid()
842 hw_ops->update_vlan_hash_table(pdata); in xlgmac_vlan_rx_add_vid()
852 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_vlan_rx_kill_vid()
854 clear_bit(vid, pdata->active_vlans); in xlgmac_vlan_rx_kill_vid()
855 hw_ops->update_vlan_hash_table(pdata); in xlgmac_vlan_rx_kill_vid()
867 if (pdata->per_channel_irq) { in xlgmac_poll_controller()
868 channel = pdata->channel_head; in xlgmac_poll_controller()
869 for (i = 0; i < pdata->channel_count; i++, channel++) in xlgmac_poll_controller()
870 xlgmac_dma_isr(channel->dma_irq, channel); in xlgmac_poll_controller()
872 disable_irq(pdata->dev_irq); in xlgmac_poll_controller()
873 xlgmac_isr(pdata->dev_irq, pdata); in xlgmac_poll_controller()
874 enable_irq(pdata->dev_irq); in xlgmac_poll_controller()
884 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_set_features()
887 rxhash = pdata->netdev_features & NETIF_F_RXHASH; in xlgmac_set_features()
888 rxcsum = pdata->netdev_features & NETIF_F_RXCSUM; in xlgmac_set_features()
889 rxvlan = pdata->netdev_features & NETIF_F_HW_VLAN_CTAG_RX; in xlgmac_set_features()
890 rxvlan_filter = pdata->netdev_features & NETIF_F_HW_VLAN_CTAG_FILTER; in xlgmac_set_features()
893 ret = hw_ops->enable_rss(pdata); in xlgmac_set_features()
895 ret = hw_ops->disable_rss(pdata); in xlgmac_set_features()
900 hw_ops->enable_rx_csum(pdata); in xlgmac_set_features()
902 hw_ops->disable_rx_csum(pdata); in xlgmac_set_features()
905 hw_ops->enable_rx_vlan_stripping(pdata); in xlgmac_set_features()
907 hw_ops->disable_rx_vlan_stripping(pdata); in xlgmac_set_features()
910 hw_ops->enable_rx_vlan_filtering(pdata); in xlgmac_set_features()
912 hw_ops->disable_rx_vlan_filtering(pdata); in xlgmac_set_features()
914 pdata->netdev_features = features; in xlgmac_set_features()
922 struct xlgmac_hw_ops *hw_ops = &pdata->hw_ops; in xlgmac_set_rx_mode()
924 hw_ops->config_rx_mode(pdata); in xlgmac_set_rx_mode()
953 struct xlgmac_pdata *pdata = channel->pdata; in xlgmac_rx_refresh()
954 struct xlgmac_ring *ring = channel->rx_ring; in xlgmac_rx_refresh()
959 desc_ops = &pdata->desc_ops; in xlgmac_rx_refresh()
960 hw_ops = &pdata->hw_ops; in xlgmac_rx_refresh()
962 while (ring->dirty != ring->cur) { in xlgmac_rx_refresh()
963 desc_data = XLGMAC_GET_DESC_DATA(ring, ring->dirty); in xlgmac_rx_refresh()
966 desc_ops->unmap_desc_data(pdata, desc_data); in xlgmac_rx_refresh()
968 if (desc_ops->map_rx_buffer(pdata, ring, desc_data)) in xlgmac_rx_refresh()
971 hw_ops->rx_desc_reset(pdata, desc_data, ring->dirty); in xlgmac_rx_refresh()
973 ring->dirty++; in xlgmac_rx_refresh()
982 desc_data = XLGMAC_GET_DESC_DATA(ring, ring->dirty - 1); in xlgmac_rx_refresh()
983 writel(lower_32_bits(desc_data->dma_desc_addr), in xlgmac_rx_refresh()
996 skb = napi_alloc_skb(napi, desc_data->rx.hdr.dma_len); in xlgmac_create_skb()
1003 dma_sync_single_range_for_cpu(pdata->dev, desc_data->rx.hdr.dma_base, in xlgmac_create_skb()
1004 desc_data->rx.hdr.dma_off, in xlgmac_create_skb()
1005 desc_data->rx.hdr.dma_len, in xlgmac_create_skb()
1008 packet = page_address(desc_data->rx.hdr.pa.pages) + in xlgmac_create_skb()
1009 desc_data->rx.hdr.pa.pages_offset; in xlgmac_create_skb()
1010 copy_len = (desc_data->rx.hdr_len) ? desc_data->rx.hdr_len : len; in xlgmac_create_skb()
1011 copy_len = min(desc_data->rx.hdr.dma_len, copy_len); in xlgmac_create_skb()
1015 len -= copy_len; in xlgmac_create_skb()
1018 dma_sync_single_range_for_cpu(pdata->dev, in xlgmac_create_skb()
1019 desc_data->rx.buf.dma_base, in xlgmac_create_skb()
1020 desc_data->rx.buf.dma_off, in xlgmac_create_skb()
1021 desc_data->rx.buf.dma_len, in xlgmac_create_skb()
1024 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, in xlgmac_create_skb()
1025 desc_data->rx.buf.pa.pages, in xlgmac_create_skb()
1026 desc_data->rx.buf.pa.pages_offset, in xlgmac_create_skb()
1027 len, desc_data->rx.buf.dma_len); in xlgmac_create_skb()
1028 desc_data->rx.buf.pa.pages = NULL; in xlgmac_create_skb()
1036 struct xlgmac_pdata *pdata = channel->pdata; in xlgmac_tx_poll()
1037 struct xlgmac_ring *ring = channel->tx_ring; in xlgmac_tx_poll()
1038 struct net_device *netdev = pdata->netdev; in xlgmac_tx_poll()
1048 desc_ops = &pdata->desc_ops; in xlgmac_tx_poll()
1049 hw_ops = &pdata->hw_ops; in xlgmac_tx_poll()
1051 /* Nothing to do if there isn't a Tx ring for this channel */ in xlgmac_tx_poll()
1055 cur = ring->cur; in xlgmac_tx_poll()
1057 /* Be sure we get ring->cur before accessing descriptor data */ in xlgmac_tx_poll()
1060 txq = netdev_get_tx_queue(netdev, channel->queue_index); in xlgmac_tx_poll()
1063 (ring->dirty != cur)) { in xlgmac_tx_poll()
1064 desc_data = XLGMAC_GET_DESC_DATA(ring, ring->dirty); in xlgmac_tx_poll()
1065 dma_desc = desc_data->dma_desc; in xlgmac_tx_poll()
1067 if (!hw_ops->tx_complete(dma_desc)) in xlgmac_tx_poll()
1076 xlgmac_dump_tx_desc(pdata, ring, ring->dirty, 1, 0); in xlgmac_tx_poll()
1078 if (hw_ops->is_last_desc(dma_desc)) { in xlgmac_tx_poll()
1079 tx_packets += desc_data->tx.packets; in xlgmac_tx_poll()
1080 tx_bytes += desc_data->tx.bytes; in xlgmac_tx_poll()
1083 /* Free the SKB and reset the descriptor for re-use */ in xlgmac_tx_poll()
1084 desc_ops->unmap_desc_data(pdata, desc_data); in xlgmac_tx_poll()
1085 hw_ops->tx_desc_reset(desc_data); in xlgmac_tx_poll()
1088 ring->dirty++; in xlgmac_tx_poll()
1096 if ((ring->tx.queue_stopped == 1) && in xlgmac_tx_poll()
1098 ring->tx.queue_stopped = 0; in xlgmac_tx_poll()
1109 struct xlgmac_pdata *pdata = channel->pdata; in xlgmac_rx_poll()
1110 struct xlgmac_ring *ring = channel->rx_ring; in xlgmac_rx_poll()
1111 struct net_device *netdev = pdata->netdev; in xlgmac_rx_poll()
1123 hw_ops = &pdata->hw_ops; in xlgmac_rx_poll()
1132 napi = (pdata->per_channel_irq) ? &channel->napi : &pdata->napi; in xlgmac_rx_poll()
1134 desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); in xlgmac_rx_poll()
1135 pkt_info = &ring->pkt_info; in xlgmac_rx_poll()
1138 if (!received && desc_data->state_saved) { in xlgmac_rx_poll()
1139 skb = desc_data->state.skb; in xlgmac_rx_poll()
1140 error = desc_data->state.error; in xlgmac_rx_poll()
1141 len = desc_data->state.len; in xlgmac_rx_poll()
1150 desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); in xlgmac_rx_poll()
1155 if (hw_ops->dev_read(channel)) in xlgmac_rx_poll()
1159 ring->cur++; in xlgmac_rx_poll()
1162 pkt_info->attributes, in xlgmac_rx_poll()
1166 pkt_info->attributes, in xlgmac_rx_poll()
1170 pkt_info->attributes, in xlgmac_rx_poll()
1178 if (error || pkt_info->errors) { in xlgmac_rx_poll()
1179 if (pkt_info->errors) in xlgmac_rx_poll()
1188 dma_desc_len = desc_data->rx.len - len; in xlgmac_rx_poll()
1198 pdata->dev, in xlgmac_rx_poll()
1199 desc_data->rx.buf.dma_base, in xlgmac_rx_poll()
1200 desc_data->rx.buf.dma_off, in xlgmac_rx_poll()
1201 desc_data->rx.buf.dma_len, in xlgmac_rx_poll()
1205 skb, skb_shinfo(skb)->nr_frags, in xlgmac_rx_poll()
1206 desc_data->rx.buf.pa.pages, in xlgmac_rx_poll()
1207 desc_data->rx.buf.pa.pages_offset, in xlgmac_rx_poll()
1209 desc_data->rx.buf.dma_len); in xlgmac_rx_poll()
1210 desc_data->rx.buf.pa.pages = NULL; in xlgmac_rx_poll()
1221 max_len = netdev->mtu + ETH_HLEN; in xlgmac_rx_poll()
1222 if (!(netdev->features & NETIF_F_HW_VLAN_CTAG_RX) && in xlgmac_rx_poll()
1223 (skb->protocol == htons(ETH_P_8021Q))) in xlgmac_rx_poll()
1226 if (skb->len > max_len) { in xlgmac_rx_poll()
1237 if (XLGMAC_GET_REG_BITS(pkt_info->attributes, in xlgmac_rx_poll()
1240 skb->ip_summed = CHECKSUM_UNNECESSARY; in xlgmac_rx_poll()
1242 if (XLGMAC_GET_REG_BITS(pkt_info->attributes, in xlgmac_rx_poll()
1246 pkt_info->vlan_ctag); in xlgmac_rx_poll()
1247 pdata->stats.rx_vlan_packets++; in xlgmac_rx_poll()
1250 if (XLGMAC_GET_REG_BITS(pkt_info->attributes, in xlgmac_rx_poll()
1253 skb_set_hash(skb, pkt_info->rss_hash, in xlgmac_rx_poll()
1254 pkt_info->rss_hash_type); in xlgmac_rx_poll()
1256 skb->dev = netdev; in xlgmac_rx_poll()
1257 skb->protocol = eth_type_trans(skb, netdev); in xlgmac_rx_poll()
1258 skb_record_rx_queue(skb, channel->queue_index); in xlgmac_rx_poll()
1268 desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); in xlgmac_rx_poll()
1269 desc_data->state_saved = 1; in xlgmac_rx_poll()
1270 desc_data->state.skb = skb; in xlgmac_rx_poll()
1271 desc_data->state.len = len; in xlgmac_rx_poll()
1272 desc_data->state.error = error; in xlgmac_rx_poll()
1289 /* Cleanup Tx ring first */ in xlgmac_one_poll()
1300 /* Enable Tx and Rx interrupts */ in xlgmac_one_poll()
1301 enable_irq(channel->dma_irq); in xlgmac_one_poll()
1322 ring_budget = budget / pdata->rx_ring_count; in xlgmac_all_poll()
1326 channel = pdata->channel_head; in xlgmac_all_poll()
1327 for (i = 0; i < pdata->channel_count; i++, channel++) { in xlgmac_all_poll()
1328 /* Cleanup Tx ring first */ in xlgmac_all_poll()
1332 if (ring_budget > (budget - processed)) in xlgmac_all_poll()
1333 ring_budget = budget - processed; in xlgmac_all_poll()
1343 /* Enable Tx and Rx interrupts */ in xlgmac_all_poll()