Lines Matching full:np
52 static void dl2k_enable_int(struct netdev_private *np) in dl2k_enable_int() argument
54 void __iomem *ioaddr = np->ioaddr; in dl2k_enable_int()
113 struct netdev_private *np; in rio_probe1() local
134 dev = alloc_etherdev (sizeof (*np)); in rio_probe1()
139 np = netdev_priv(dev); in rio_probe1()
142 np->rmon_enable = true; in rio_probe1()
148 np->eeprom_addr = ioaddr; in rio_probe1()
150 if (np->rmon_enable) { in rio_probe1()
157 np->ioaddr = ioaddr; in rio_probe1()
158 np->chip_id = chip_idx; in rio_probe1()
159 np->pdev = pdev; in rio_probe1()
161 spin_lock_init(&np->stats_lock); in rio_probe1()
162 spin_lock_init (&np->tx_lock); in rio_probe1()
163 spin_lock_init (&np->rx_lock); in rio_probe1()
166 np->an_enable = 1; in rio_probe1()
167 np->tx_coalesce = 1; in rio_probe1()
170 np->an_enable = 0; in rio_probe1()
174 np->an_enable = 2; in rio_probe1()
177 np->speed = 100; in rio_probe1()
178 np->full_duplex = 1; in rio_probe1()
181 np->speed = 100; in rio_probe1()
182 np->full_duplex = 0; in rio_probe1()
185 np->speed = 10; in rio_probe1()
186 np->full_duplex = 1; in rio_probe1()
189 np->speed = 10; in rio_probe1()
190 np->full_duplex = 0; in rio_probe1()
193 np->speed=1000; in rio_probe1()
194 np->full_duplex=1; in rio_probe1()
197 np->speed = 1000; in rio_probe1()
198 np->full_duplex = 0; in rio_probe1()
200 np->an_enable = 1; in rio_probe1()
204 np->jumbo = 1; in rio_probe1()
207 np->jumbo = 0; in rio_probe1()
211 np->vlan = (vlan[card_idx] > 0 && vlan[card_idx] < 4096) ? in rio_probe1()
214 np->rx_coalesce = rx_coalesce; in rio_probe1()
215 np->rx_timeout = rx_timeout; in rio_probe1()
216 np->coalesce = 1; in rio_probe1()
218 np->tx_flow = (tx_flow == 0) ? 0 : 1; in rio_probe1()
219 np->rx_flow = (rx_flow == 0) ? 0 : 1; in rio_probe1()
234 dev->max_mtu = np->jumbo ? MAX_JUMBO : PACKET_SIZE; in rio_probe1()
242 np->tx_ring = ring_space; in rio_probe1()
243 np->tx_ring_dma = ring_dma; in rio_probe1()
249 np->rx_ring = ring_space; in rio_probe1()
250 np->rx_ring_dma = ring_dma; in rio_probe1()
261 np->phy_media = (dr16(ASICCtrl) & PhyMedia) ? 1 : 0; in rio_probe1()
262 np->link_status = 0; in rio_probe1()
264 if (np->phy_media) { in rio_probe1()
266 if (np->an_enable == 2) { in rio_probe1()
267 np->an_enable = 1; in rio_probe1()
272 if (np->speed == 1000) in rio_probe1()
273 np->an_enable = 1; in rio_probe1()
283 dev->name, np->name, dev->dev_addr, irq); in rio_probe1()
287 if (np->coalesce) in rio_probe1()
291 np->rx_coalesce, np->rx_timeout*640); in rio_probe1()
292 if (np->vlan) in rio_probe1()
293 printk(KERN_INFO "vlan(id):\t%d\n", np->vlan); in rio_probe1()
297 dma_free_coherent(&pdev->dev, RX_TOTAL_SIZE, np->rx_ring, in rio_probe1()
298 np->rx_ring_dma); in rio_probe1()
300 dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, np->tx_ring, in rio_probe1()
301 np->tx_ring_dma); in rio_probe1()
303 if (np->rmon_enable) in rio_probe1()
304 pci_iounmap(pdev, np->ioaddr); in rio_probe1()
305 pci_iounmap(pdev, np->eeprom_addr); in rio_probe1()
318 struct netdev_private *np = netdev_priv(dev); in find_miiphy() local
321 np->phy_addr = 1; in find_miiphy()
326 np->phy_addr = i; in find_miiphy()
340 struct netdev_private *np = netdev_priv(dev); in parse_eeprom() local
341 void __iomem *ioaddr = np->ioaddr; in parse_eeprom()
351 ((__le16 *) sromdata)[i] = cpu_to_le16(read_eeprom(np, i)); in parse_eeprom()
353 if (np->pdev->vendor == PCI_VENDOR_ID_DLINK) { /* D-Link Only */ in parse_eeprom()
366 if (np->chip_id == CHIP_IP1000A) { in parse_eeprom()
367 np->led_mode = le16_to_cpu(psrom->led_mode); in parse_eeprom()
371 if (np->pdev->vendor != PCI_VENDOR_ID_DLINK) { in parse_eeprom()
391 np->duplex_polarity = psib[i]; in parse_eeprom()
395 np->wake_polarity = psib[i]; in parse_eeprom()
399 memcpy (np->name, &(psib[i]), j); in parse_eeprom()
418 struct netdev_private *np = netdev_priv(dev); in rio_set_led_mode() local
419 void __iomem *ioaddr = np->ioaddr; in rio_set_led_mode()
422 if (np->chip_id != CHIP_IP1000A) in rio_set_led_mode()
428 if (np->led_mode & 0x01) in rio_set_led_mode()
430 if (np->led_mode & 0x02) in rio_set_led_mode()
432 if (np->led_mode & 0x08) in rio_set_led_mode()
445 struct netdev_private *np = netdev_priv(dev); in free_list() local
451 skb = np->rx_skbuff[i]; in free_list()
453 dma_unmap_single(&np->pdev->dev, in free_list()
454 desc_to_dma(&np->rx_ring[i]), in free_list()
457 np->rx_skbuff[i] = NULL; in free_list()
459 np->rx_ring[i].status = 0; in free_list()
460 np->rx_ring[i].fraginfo = 0; in free_list()
463 skb = np->tx_skbuff[i]; in free_list()
465 dma_unmap_single(&np->pdev->dev, in free_list()
466 desc_to_dma(&np->tx_ring[i]), in free_list()
469 np->tx_skbuff[i] = NULL; in free_list()
474 static void rio_reset_ring(struct netdev_private *np) in rio_reset_ring() argument
478 np->cur_rx = 0; in rio_reset_ring()
479 np->cur_tx = 0; in rio_reset_ring()
480 np->old_rx = 0; in rio_reset_ring()
481 np->old_tx = 0; in rio_reset_ring()
484 np->tx_ring[i].status = cpu_to_le64(TFDDone); in rio_reset_ring()
487 np->rx_ring[i].status = 0; in rio_reset_ring()
493 struct netdev_private *np = netdev_priv(dev); in alloc_list() local
496 rio_reset_ring(np); in alloc_list()
497 np->rx_buf_sz = (dev->mtu <= 1500 ? PACKET_SIZE : dev->mtu + 32); in alloc_list()
501 np->tx_skbuff[i] = NULL; in alloc_list()
502 np->tx_ring[i].next_desc = cpu_to_le64(np->tx_ring_dma + in alloc_list()
513 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in alloc_list()
514 np->rx_skbuff[i] = skb; in alloc_list()
518 addr = dma_map_single(&np->pdev->dev, skb->data, in alloc_list()
519 np->rx_buf_sz, DMA_FROM_DEVICE); in alloc_list()
520 if (dma_mapping_error(&np->pdev->dev, addr)) in alloc_list()
523 np->rx_ring[i].next_desc = cpu_to_le64(np->rx_ring_dma + in alloc_list()
527 np->rx_ring[i].fraginfo = cpu_to_le64(addr); in alloc_list()
528 np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); in alloc_list()
534 dev_kfree_skb(np->rx_skbuff[i]); in alloc_list()
535 np->rx_skbuff[i] = NULL; in alloc_list()
543 struct netdev_private *np = netdev_priv(dev); in rio_hw_init() local
544 void __iomem *ioaddr = np->ioaddr; in rio_hw_init()
558 if (np->chip_id == CHIP_IP1000A && in rio_hw_init()
559 (np->pdev->revision == 0x40 || np->pdev->revision == 0x41)) { in rio_hw_init()
561 mii_write(dev, np->phy_addr, 31, 0x0001); in rio_hw_init()
562 mii_write(dev, np->phy_addr, 27, 0x01e0); in rio_hw_init()
563 mii_write(dev, np->phy_addr, 31, 0x0002); in rio_hw_init()
564 mii_write(dev, np->phy_addr, 27, 0xeb8e); in rio_hw_init()
565 mii_write(dev, np->phy_addr, 31, 0x0000); in rio_hw_init()
566 mii_write(dev, np->phy_addr, 30, 0x005e); in rio_hw_init()
568 mii_write(dev, np->phy_addr, MII_CTRL1000, 0x0700); in rio_hw_init()
571 if (np->phy_media) in rio_hw_init()
577 if (np->jumbo != 0) in rio_hw_init()
581 dw32(RFDListPtr0, np->rx_ring_dma); in rio_hw_init()
592 if (np->coalesce) { in rio_hw_init()
593 dw32(RxDMAIntCtrl, np->rx_coalesce | np->rx_timeout << 16); in rio_hw_init()
600 if (!np->rmon_enable) in rio_hw_init()
606 if (np->vlan) { in rio_hw_init()
610 dw16(VLANId, np->vlan); in rio_hw_init()
612 dw32(VLANTag, 0x8100 << 16 | np->vlan); in rio_hw_init()
622 macctrl |= (np->vlan) ? AutoVLANuntagging : 0; in rio_hw_init()
623 macctrl |= (np->full_duplex) ? DuplexSelect : 0; in rio_hw_init()
624 macctrl |= (np->tx_flow) ? TxFlowControlEnable : 0; in rio_hw_init()
625 macctrl |= (np->rx_flow) ? RxFlowControlEnable : 0; in rio_hw_init()
631 struct netdev_private *np = netdev_priv(dev); in rio_hw_stop() local
632 void __iomem *ioaddr = np->ioaddr; in rio_hw_stop()
643 struct netdev_private *np = netdev_priv(dev); in rio_open() local
644 const int irq = np->pdev->irq; in rio_open()
660 timer_setup(&np->timer, rio_timer, 0); in rio_open()
661 np->timer.expires = jiffies + 1 * HZ; in rio_open()
662 add_timer(&np->timer); in rio_open()
666 dl2k_enable_int(np); in rio_open()
673 struct netdev_private *np = timer_container_of(np, t, timer); in rio_timer() local
674 struct net_device *dev = pci_get_drvdata(np->pdev); in rio_timer()
679 spin_lock_irqsave(&np->rx_lock, flags); in rio_timer()
681 if (np->cur_rx - np->old_rx >= RX_RING_SIZE) { in rio_timer()
684 for (; np->cur_rx - np->old_rx > 0; np->old_rx++) { in rio_timer()
686 entry = np->old_rx % RX_RING_SIZE; in rio_timer()
688 if (np->rx_skbuff[entry] == NULL) { in rio_timer()
690 np->rx_buf_sz); in rio_timer()
692 np->rx_ring[entry].fraginfo = 0; in rio_timer()
698 np->rx_skbuff[entry] = skb; in rio_timer()
699 np->rx_ring[entry].fraginfo = in rio_timer()
700 cpu_to_le64 (dma_map_single(&np->pdev->dev, skb->data, in rio_timer()
701 np->rx_buf_sz, DMA_FROM_DEVICE)); in rio_timer()
703 np->rx_ring[entry].fraginfo |= in rio_timer()
704 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer()
705 np->rx_ring[entry].status = 0; in rio_timer()
708 spin_unlock_irqrestore (&np->rx_lock, flags); in rio_timer()
709 np->timer.expires = jiffies + next_tick; in rio_timer()
710 add_timer(&np->timer); in rio_timer()
716 struct netdev_private *np = netdev_priv(dev); in rio_tx_timeout() local
717 void __iomem *ioaddr = np->ioaddr; in rio_tx_timeout()
729 struct netdev_private *np = netdev_priv(dev); in start_xmit() local
730 void __iomem *ioaddr = np->ioaddr; in start_xmit()
735 if (np->link_status == 0) { /* Link Down */ in start_xmit()
739 entry = np->cur_tx % TX_RING_SIZE; in start_xmit()
740 np->tx_skbuff[entry] = skb; in start_xmit()
741 txdesc = &np->tx_ring[entry]; in start_xmit()
750 if (np->vlan) { in start_xmit()
752 ((u64)np->vlan << 32) | in start_xmit()
755 txdesc->fraginfo = cpu_to_le64 (dma_map_single(&np->pdev->dev, skb->data, in start_xmit()
761 if (entry % np->tx_coalesce == 0 || np->speed == 10) in start_xmit()
775 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; in start_xmit()
776 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in start_xmit()
777 < TX_QUEUE_LEN - 1 && np->speed != 10) { in start_xmit()
785 dw32(TFDListPtr0, np->tx_ring_dma + in start_xmit()
797 struct netdev_private *np = netdev_priv(dev); in rio_interrupt() local
798 void __iomem *ioaddr = np->ioaddr; in rio_interrupt()
828 if (np->cur_tx != np->old_tx) in rio_interrupt()
836 struct netdev_private *np = netdev_priv(dev); in rio_free_tx() local
837 int entry = np->old_tx % TX_RING_SIZE; in rio_free_tx()
841 spin_lock(&np->tx_lock); in rio_free_tx()
843 spin_lock_irqsave(&np->tx_lock, flag); in rio_free_tx()
846 while (entry != np->cur_tx) { in rio_free_tx()
849 if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) in rio_free_tx()
851 skb = np->tx_skbuff[entry]; in rio_free_tx()
852 dma_unmap_single(&np->pdev->dev, in rio_free_tx()
853 desc_to_dma(&np->tx_ring[entry]), skb->len, in rio_free_tx()
860 np->tx_skbuff[entry] = NULL; in rio_free_tx()
864 spin_unlock(&np->tx_lock); in rio_free_tx()
866 spin_unlock_irqrestore(&np->tx_lock, flag); in rio_free_tx()
867 np->old_tx = entry; in rio_free_tx()
873 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in rio_free_tx()
874 < TX_QUEUE_LEN - 1 || np->speed == 10)) { in rio_free_tx()
882 struct netdev_private *np = netdev_priv(dev); in tx_error() local
883 void __iomem *ioaddr = np->ioaddr; in tx_error()
906 dw32(TFDListPtr0, np->tx_ring_dma + in tx_error()
907 np->old_tx * sizeof (struct netdev_desc)); in tx_error()
927 spin_lock(&np->stats_lock); in tx_error()
933 spin_unlock(&np->stats_lock); in tx_error()
942 struct netdev_private *np = netdev_priv(dev); in receive_packet() local
943 int entry = np->cur_rx % RX_RING_SIZE; in receive_packet()
948 struct netdev_desc *desc = &np->rx_ring[entry]; in receive_packet()
969 if (frame_status & RxAlignmentError && np->speed != 1000) in receive_packet()
981 dma_unmap_single(&np->pdev->dev, in receive_packet()
983 np->rx_buf_sz, in receive_packet()
985 skb_put (skb = np->rx_skbuff[entry], pkt_len); in receive_packet()
986 np->rx_skbuff[entry] = NULL; in receive_packet()
988 dma_sync_single_for_cpu(&np->pdev->dev, in receive_packet()
990 np->rx_buf_sz, in receive_packet()
993 np->rx_skbuff[entry]->data, in receive_packet()
996 dma_sync_single_for_device(&np->pdev->dev, in receive_packet()
998 np->rx_buf_sz, in receive_packet()
1004 if (np->pdev->pci_rev_id >= 0x0c && in receive_packet()
1013 spin_lock(&np->rx_lock); in receive_packet()
1014 np->cur_rx = entry; in receive_packet()
1016 entry = np->old_rx; in receive_packet()
1017 while (entry != np->cur_rx) { in receive_packet()
1020 if (np->rx_skbuff[entry] == NULL) { in receive_packet()
1021 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in receive_packet()
1023 np->rx_ring[entry].fraginfo = 0; in receive_packet()
1030 np->rx_skbuff[entry] = skb; in receive_packet()
1031 np->rx_ring[entry].fraginfo = in receive_packet()
1032 cpu_to_le64(dma_map_single(&np->pdev->dev, skb->data, in receive_packet()
1033 np->rx_buf_sz, DMA_FROM_DEVICE)); in receive_packet()
1035 np->rx_ring[entry].fraginfo |= in receive_packet()
1036 cpu_to_le64((u64)np->rx_buf_sz << 48); in receive_packet()
1037 np->rx_ring[entry].status = 0; in receive_packet()
1040 np->old_rx = entry; in receive_packet()
1041 spin_unlock(&np->rx_lock); in receive_packet()
1048 struct netdev_private *np = netdev_priv(dev); in rio_error() local
1049 void __iomem *ioaddr = np->ioaddr; in rio_error()
1056 if (np->phy_media) in rio_error()
1060 if (np->speed == 1000) in rio_error()
1061 np->tx_coalesce = tx_coalesce; in rio_error()
1063 np->tx_coalesce = 1; in rio_error()
1065 macctrl |= (np->vlan) ? AutoVLANuntagging : 0; in rio_error()
1066 macctrl |= (np->full_duplex) ? DuplexSelect : 0; in rio_error()
1067 macctrl |= (np->tx_flow) ? in rio_error()
1069 macctrl |= (np->rx_flow) ? in rio_error()
1072 np->link_status = 1; in rio_error()
1076 np->link_status = 0; in rio_error()
1100 struct netdev_private *np = netdev_priv(dev); in get_stats() local
1101 void __iomem *ioaddr = np->ioaddr; in get_stats()
1105 spin_lock_irqsave(&np->stats_lock, flags); in get_stats()
1146 if (np->rmon_enable) in get_stats()
1156 spin_unlock_irqrestore(&np->stats_lock, flags); in get_stats()
1164 struct netdev_private *np = netdev_priv(dev); in clear_stats() local
1165 void __iomem *ioaddr = np->ioaddr; in clear_stats()
1201 if (np->rmon_enable) in clear_stats()
1215 struct netdev_private *np = netdev_priv(dev); in set_multicast() local
1216 void __iomem *ioaddr = np->ioaddr; in set_multicast()
1249 if (np->vlan) { in set_multicast()
1261 struct netdev_private *np = netdev_priv(dev); in rio_get_drvinfo() local
1264 strscpy(info->bus_info, pci_name(np->pdev), sizeof(info->bus_info)); in rio_get_drvinfo()
1270 struct netdev_private *np = netdev_priv(dev); in rio_get_link_ksettings() local
1273 if (np->phy_media) { in rio_get_link_ksettings()
1290 if (np->link_status) { in rio_get_link_ksettings()
1291 cmd->base.speed = np->speed; in rio_get_link_ksettings()
1292 cmd->base.duplex = np->full_duplex ? DUPLEX_FULL : DUPLEX_HALF; in rio_get_link_ksettings()
1297 if (np->an_enable) in rio_get_link_ksettings()
1302 cmd->base.phy_address = np->phy_addr; in rio_get_link_ksettings()
1315 struct netdev_private *np = netdev_priv(dev); in rio_set_link_ksettings() local
1321 if (np->an_enable) { in rio_set_link_ksettings()
1324 np->an_enable = 1; in rio_set_link_ksettings()
1329 np->an_enable = 0; in rio_set_link_ksettings()
1330 if (np->speed == 1000) { in rio_set_link_ksettings()
1337 np->speed = 10; in rio_set_link_ksettings()
1338 np->full_duplex = (duplex == DUPLEX_FULL); in rio_set_link_ksettings()
1341 np->speed = 100; in rio_set_link_ksettings()
1342 np->full_duplex = (duplex == DUPLEX_FULL); in rio_set_link_ksettings()
1355 struct netdev_private *np = netdev_priv(dev); in rio_get_link() local
1356 return np->link_status; in rio_get_link()
1370 struct netdev_private *np = netdev_priv(dev); in rio_ioctl() local
1373 phy_addr = np->phy_addr; in rio_ioctl()
1396 static int read_eeprom(struct netdev_private *np, int eep_addr) in read_eeprom() argument
1398 void __iomem *ioaddr = np->eeprom_addr; in read_eeprom()
1418 struct netdev_private *np = netdev_priv(dev); in mii_sendbit() local
1419 void __iomem *ioaddr = np->ioaddr; in mii_sendbit()
1431 struct netdev_private *np = netdev_priv(dev); in mii_getbit() local
1432 void __iomem *ioaddr = np->ioaddr; in mii_getbit()
1501 struct netdev_private *np; in mii_wait_link() local
1503 np = netdev_priv(dev); in mii_wait_link()
1504 phy_addr = np->phy_addr; in mii_wait_link()
1522 struct netdev_private *np; in mii_get_media() local
1524 np = netdev_priv(dev); in mii_get_media()
1525 phy_addr = np->phy_addr; in mii_get_media()
1528 if (np->an_enable) { in mii_get_media()
1538 np->speed = 1000; in mii_get_media()
1539 np->full_duplex = 1; in mii_get_media()
1542 np->speed = 1000; in mii_get_media()
1543 np->full_duplex = 0; in mii_get_media()
1546 np->speed = 100; in mii_get_media()
1547 np->full_duplex = 1; in mii_get_media()
1550 np->speed = 100; in mii_get_media()
1551 np->full_duplex = 0; in mii_get_media()
1554 np->speed = 10; in mii_get_media()
1555 np->full_duplex = 1; in mii_get_media()
1558 np->speed = 10; in mii_get_media()
1559 np->full_duplex = 0; in mii_get_media()
1563 np->tx_flow &= 1; in mii_get_media()
1564 np->rx_flow &= 1; in mii_get_media()
1566 np->tx_flow = 0; in mii_get_media()
1567 np->rx_flow &= 1; in mii_get_media()
1588 if (np->tx_flow) in mii_get_media()
1592 if (np->rx_flow) in mii_get_media()
1608 struct netdev_private *np; in mii_set_media() local
1609 np = netdev_priv(dev); in mii_set_media()
1610 phy_addr = np->phy_addr; in mii_set_media()
1613 if (np->an_enable) { in mii_set_media()
1665 if (np->speed == 100) { in mii_set_media()
1668 } else if (np->speed == 10) { in mii_set_media()
1671 if (np->full_duplex) { in mii_set_media()
1695 struct netdev_private *np; in mii_get_media_pcs() local
1697 np = netdev_priv(dev); in mii_get_media_pcs()
1698 phy_addr = np->phy_addr; in mii_get_media_pcs()
1701 if (np->an_enable) { in mii_get_media_pcs()
1708 np->speed = 1000; in mii_get_media_pcs()
1711 np->full_duplex = 1; in mii_get_media_pcs()
1714 np->full_duplex = 0; in mii_get_media_pcs()
1717 np->tx_flow &= 1; in mii_get_media_pcs()
1718 np->rx_flow &= 1; in mii_get_media_pcs()
1720 np->tx_flow = 0; in mii_get_media_pcs()
1721 np->rx_flow &= 1; in mii_get_media_pcs()
1733 if (np->tx_flow) in mii_get_media_pcs()
1737 if (np->rx_flow) in mii_get_media_pcs()
1752 struct netdev_private *np; in mii_set_media_pcs() local
1753 np = netdev_priv(dev); in mii_set_media_pcs()
1754 phy_addr = np->phy_addr; in mii_set_media_pcs()
1757 if (np->an_enable) { in mii_set_media_pcs()
1781 if (np->full_duplex) { in mii_set_media_pcs()
1801 struct netdev_private *np = netdev_priv(dev); in rio_close() local
1802 struct pci_dev *pdev = np->pdev; in rio_close()
1809 timer_delete_sync(&np->timer); in rio_close()
1822 struct netdev_private *np = netdev_priv(dev); in rio_remove1() local
1825 dma_free_coherent(&pdev->dev, RX_TOTAL_SIZE, np->rx_ring, in rio_remove1()
1826 np->rx_ring_dma); in rio_remove1()
1827 dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, np->tx_ring, in rio_remove1()
1828 np->tx_ring_dma); in rio_remove1()
1829 if (np->rmon_enable) in rio_remove1()
1830 pci_iounmap(pdev, np->ioaddr); in rio_remove1()
1831 pci_iounmap(pdev, np->eeprom_addr); in rio_remove1()
1842 struct netdev_private *np = netdev_priv(dev); in rio_suspend() local
1848 timer_delete_sync(&np->timer); in rio_suspend()
1857 struct netdev_private *np = netdev_priv(dev); in rio_resume() local
1862 rio_reset_ring(np); in rio_resume()
1864 np->timer.expires = jiffies + 1 * HZ; in rio_resume()
1865 add_timer(&np->timer); in rio_resume()
1867 dl2k_enable_int(np); in rio_resume()