Lines Matching full:np
52 static void dl2k_enable_int(struct netdev_private *np) in dl2k_enable_int() argument
54 void __iomem *ioaddr = np->ioaddr; in dl2k_enable_int()
106 struct netdev_private *np; in rio_probe1() local
127 dev = alloc_etherdev (sizeof (*np)); in rio_probe1()
132 np = netdev_priv(dev); in rio_probe1()
138 np->eeprom_addr = ioaddr; in rio_probe1()
146 np->ioaddr = ioaddr; in rio_probe1()
147 np->chip_id = chip_idx; in rio_probe1()
148 np->pdev = pdev; in rio_probe1()
150 spin_lock_init(&np->stats_lock); in rio_probe1()
151 spin_lock_init (&np->tx_lock); in rio_probe1()
152 spin_lock_init (&np->rx_lock); in rio_probe1()
155 np->an_enable = 1; in rio_probe1()
156 np->tx_coalesce = 1; in rio_probe1()
159 np->an_enable = 0; in rio_probe1()
163 np->an_enable = 2; in rio_probe1()
166 np->speed = 100; in rio_probe1()
167 np->full_duplex = 1; in rio_probe1()
170 np->speed = 100; in rio_probe1()
171 np->full_duplex = 0; in rio_probe1()
174 np->speed = 10; in rio_probe1()
175 np->full_duplex = 1; in rio_probe1()
178 np->speed = 10; in rio_probe1()
179 np->full_duplex = 0; in rio_probe1()
182 np->speed=1000; in rio_probe1()
183 np->full_duplex=1; in rio_probe1()
186 np->speed = 1000; in rio_probe1()
187 np->full_duplex = 0; in rio_probe1()
189 np->an_enable = 1; in rio_probe1()
193 np->jumbo = 1; in rio_probe1()
196 np->jumbo = 0; in rio_probe1()
200 np->vlan = (vlan[card_idx] > 0 && vlan[card_idx] < 4096) ? in rio_probe1()
203 np->rx_coalesce = rx_coalesce; in rio_probe1()
204 np->rx_timeout = rx_timeout; in rio_probe1()
205 np->coalesce = 1; in rio_probe1()
207 np->tx_flow = (tx_flow == 0) ? 0 : 1; in rio_probe1()
208 np->rx_flow = (rx_flow == 0) ? 0 : 1; in rio_probe1()
223 dev->max_mtu = np->jumbo ? MAX_JUMBO : PACKET_SIZE; in rio_probe1()
231 np->tx_ring = ring_space; in rio_probe1()
232 np->tx_ring_dma = ring_dma; in rio_probe1()
238 np->rx_ring = ring_space; in rio_probe1()
239 np->rx_ring_dma = ring_dma; in rio_probe1()
250 np->phy_media = (dr16(ASICCtrl) & PhyMedia) ? 1 : 0; in rio_probe1()
251 np->link_status = 0; in rio_probe1()
253 if (np->phy_media) { in rio_probe1()
255 if (np->an_enable == 2) { in rio_probe1()
256 np->an_enable = 1; in rio_probe1()
261 if (np->speed == 1000) in rio_probe1()
262 np->an_enable = 1; in rio_probe1()
272 dev->name, np->name, dev->dev_addr, irq); in rio_probe1()
276 if (np->coalesce) in rio_probe1()
280 np->rx_coalesce, np->rx_timeout*640); in rio_probe1()
281 if (np->vlan) in rio_probe1()
282 printk(KERN_INFO "vlan(id):\t%d\n", np->vlan); in rio_probe1()
286 dma_free_coherent(&pdev->dev, RX_TOTAL_SIZE, np->rx_ring, in rio_probe1()
287 np->rx_ring_dma); in rio_probe1()
289 dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, np->tx_ring, in rio_probe1()
290 np->tx_ring_dma); in rio_probe1()
293 pci_iounmap(pdev, np->ioaddr); in rio_probe1()
295 pci_iounmap(pdev, np->eeprom_addr); in rio_probe1()
308 struct netdev_private *np = netdev_priv(dev); in find_miiphy() local
311 np->phy_addr = 1; in find_miiphy()
316 np->phy_addr = i; in find_miiphy()
330 struct netdev_private *np = netdev_priv(dev); in parse_eeprom() local
331 void __iomem *ioaddr = np->ioaddr; in parse_eeprom()
341 ((__le16 *) sromdata)[i] = cpu_to_le16(read_eeprom(np, i)); in parse_eeprom()
343 if (np->pdev->vendor == PCI_VENDOR_ID_DLINK) { /* D-Link Only */ in parse_eeprom()
356 if (np->chip_id == CHIP_IP1000A) { in parse_eeprom()
357 np->led_mode = le16_to_cpu(psrom->led_mode); in parse_eeprom()
361 if (np->pdev->vendor != PCI_VENDOR_ID_DLINK) { in parse_eeprom()
381 np->duplex_polarity = psib[i]; in parse_eeprom()
385 np->wake_polarity = psib[i]; in parse_eeprom()
389 memcpy (np->name, &(psib[i]), j); in parse_eeprom()
408 struct netdev_private *np = netdev_priv(dev); in rio_set_led_mode() local
409 void __iomem *ioaddr = np->ioaddr; in rio_set_led_mode()
412 if (np->chip_id != CHIP_IP1000A) in rio_set_led_mode()
418 if (np->led_mode & 0x01) in rio_set_led_mode()
420 if (np->led_mode & 0x02) in rio_set_led_mode()
422 if (np->led_mode & 0x08) in rio_set_led_mode()
435 struct netdev_private *np = netdev_priv(dev); in free_list() local
441 skb = np->rx_skbuff[i]; in free_list()
443 dma_unmap_single(&np->pdev->dev, in free_list()
444 desc_to_dma(&np->rx_ring[i]), in free_list()
447 np->rx_skbuff[i] = NULL; in free_list()
449 np->rx_ring[i].status = 0; in free_list()
450 np->rx_ring[i].fraginfo = 0; in free_list()
453 skb = np->tx_skbuff[i]; in free_list()
455 dma_unmap_single(&np->pdev->dev, in free_list()
456 desc_to_dma(&np->tx_ring[i]), in free_list()
459 np->tx_skbuff[i] = NULL; in free_list()
464 static void rio_reset_ring(struct netdev_private *np) in rio_reset_ring() argument
468 np->cur_rx = 0; in rio_reset_ring()
469 np->cur_tx = 0; in rio_reset_ring()
470 np->old_rx = 0; in rio_reset_ring()
471 np->old_tx = 0; in rio_reset_ring()
474 np->tx_ring[i].status = cpu_to_le64(TFDDone); in rio_reset_ring()
477 np->rx_ring[i].status = 0; in rio_reset_ring()
483 struct netdev_private *np = netdev_priv(dev); in alloc_list() local
486 rio_reset_ring(np); in alloc_list()
487 np->rx_buf_sz = (dev->mtu <= 1500 ? PACKET_SIZE : dev->mtu + 32); in alloc_list()
491 np->tx_skbuff[i] = NULL; in alloc_list()
492 np->tx_ring[i].next_desc = cpu_to_le64(np->tx_ring_dma + in alloc_list()
502 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in alloc_list()
503 np->rx_skbuff[i] = skb; in alloc_list()
509 np->rx_ring[i].next_desc = cpu_to_le64(np->rx_ring_dma + in alloc_list()
513 np->rx_ring[i].fraginfo = in alloc_list()
514 cpu_to_le64(dma_map_single(&np->pdev->dev, skb->data, in alloc_list()
515 np->rx_buf_sz, DMA_FROM_DEVICE)); in alloc_list()
516 np->rx_ring[i].fraginfo |= cpu_to_le64((u64)np->rx_buf_sz << 48); in alloc_list()
524 struct netdev_private *np = netdev_priv(dev); in rio_hw_init() local
525 void __iomem *ioaddr = np->ioaddr; in rio_hw_init()
539 if (np->chip_id == CHIP_IP1000A && in rio_hw_init()
540 (np->pdev->revision == 0x40 || np->pdev->revision == 0x41)) { in rio_hw_init()
542 mii_write(dev, np->phy_addr, 31, 0x0001); in rio_hw_init()
543 mii_write(dev, np->phy_addr, 27, 0x01e0); in rio_hw_init()
544 mii_write(dev, np->phy_addr, 31, 0x0002); in rio_hw_init()
545 mii_write(dev, np->phy_addr, 27, 0xeb8e); in rio_hw_init()
546 mii_write(dev, np->phy_addr, 31, 0x0000); in rio_hw_init()
547 mii_write(dev, np->phy_addr, 30, 0x005e); in rio_hw_init()
549 mii_write(dev, np->phy_addr, MII_CTRL1000, 0x0700); in rio_hw_init()
552 if (np->phy_media) in rio_hw_init()
558 if (np->jumbo != 0) in rio_hw_init()
562 dw32(RFDListPtr0, np->rx_ring_dma); in rio_hw_init()
574 if (np->coalesce) { in rio_hw_init()
575 dw32(RxDMAIntCtrl, np->rx_coalesce | np->rx_timeout << 16); in rio_hw_init()
587 if (np->vlan) { in rio_hw_init()
591 dw16(VLANId, np->vlan); in rio_hw_init()
593 dw32(VLANTag, 0x8100 << 16 | np->vlan); in rio_hw_init()
603 macctrl |= (np->vlan) ? AutoVLANuntagging : 0; in rio_hw_init()
604 macctrl |= (np->full_duplex) ? DuplexSelect : 0; in rio_hw_init()
605 macctrl |= (np->tx_flow) ? TxFlowControlEnable : 0; in rio_hw_init()
606 macctrl |= (np->rx_flow) ? RxFlowControlEnable : 0; in rio_hw_init()
612 struct netdev_private *np = netdev_priv(dev); in rio_hw_stop() local
613 void __iomem *ioaddr = np->ioaddr; in rio_hw_stop()
624 struct netdev_private *np = netdev_priv(dev); in rio_open() local
625 const int irq = np->pdev->irq; in rio_open()
641 timer_setup(&np->timer, rio_timer, 0); in rio_open()
642 np->timer.expires = jiffies + 1 * HZ; in rio_open()
643 add_timer(&np->timer); in rio_open()
647 dl2k_enable_int(np); in rio_open()
654 struct netdev_private *np = from_timer(np, t, timer); in rio_timer() local
655 struct net_device *dev = pci_get_drvdata(np->pdev); in rio_timer()
660 spin_lock_irqsave(&np->rx_lock, flags); in rio_timer()
662 if (np->cur_rx - np->old_rx >= RX_RING_SIZE) { in rio_timer()
665 for (; np->cur_rx - np->old_rx > 0; np->old_rx++) { in rio_timer()
667 entry = np->old_rx % RX_RING_SIZE; in rio_timer()
669 if (np->rx_skbuff[entry] == NULL) { in rio_timer()
671 np->rx_buf_sz); in rio_timer()
673 np->rx_ring[entry].fraginfo = 0; in rio_timer()
679 np->rx_skbuff[entry] = skb; in rio_timer()
680 np->rx_ring[entry].fraginfo = in rio_timer()
681 cpu_to_le64 (dma_map_single(&np->pdev->dev, skb->data, in rio_timer()
682 np->rx_buf_sz, DMA_FROM_DEVICE)); in rio_timer()
684 np->rx_ring[entry].fraginfo |= in rio_timer()
685 cpu_to_le64((u64)np->rx_buf_sz << 48); in rio_timer()
686 np->rx_ring[entry].status = 0; in rio_timer()
689 spin_unlock_irqrestore (&np->rx_lock, flags); in rio_timer()
690 np->timer.expires = jiffies + next_tick; in rio_timer()
691 add_timer(&np->timer); in rio_timer()
697 struct netdev_private *np = netdev_priv(dev); in rio_tx_timeout() local
698 void __iomem *ioaddr = np->ioaddr; in rio_tx_timeout()
710 struct netdev_private *np = netdev_priv(dev); in start_xmit() local
711 void __iomem *ioaddr = np->ioaddr; in start_xmit()
716 if (np->link_status == 0) { /* Link Down */ in start_xmit()
720 entry = np->cur_tx % TX_RING_SIZE; in start_xmit()
721 np->tx_skbuff[entry] = skb; in start_xmit()
722 txdesc = &np->tx_ring[entry]; in start_xmit()
731 if (np->vlan) { in start_xmit()
733 ((u64)np->vlan << 32) | in start_xmit()
736 txdesc->fraginfo = cpu_to_le64 (dma_map_single(&np->pdev->dev, skb->data, in start_xmit()
742 if (entry % np->tx_coalesce == 0 || np->speed == 10) in start_xmit()
756 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; in start_xmit()
757 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in start_xmit()
758 < TX_QUEUE_LEN - 1 && np->speed != 10) { in start_xmit()
766 dw32(TFDListPtr0, np->tx_ring_dma + in start_xmit()
778 struct netdev_private *np = netdev_priv(dev); in rio_interrupt() local
779 void __iomem *ioaddr = np->ioaddr; in rio_interrupt()
809 if (np->cur_tx != np->old_tx) in rio_interrupt()
817 struct netdev_private *np = netdev_priv(dev); in rio_free_tx() local
818 int entry = np->old_tx % TX_RING_SIZE; in rio_free_tx()
822 spin_lock(&np->tx_lock); in rio_free_tx()
824 spin_lock_irqsave(&np->tx_lock, flag); in rio_free_tx()
827 while (entry != np->cur_tx) { in rio_free_tx()
830 if (!(np->tx_ring[entry].status & cpu_to_le64(TFDDone))) in rio_free_tx()
832 skb = np->tx_skbuff[entry]; in rio_free_tx()
833 dma_unmap_single(&np->pdev->dev, in rio_free_tx()
834 desc_to_dma(&np->tx_ring[entry]), skb->len, in rio_free_tx()
841 np->tx_skbuff[entry] = NULL; in rio_free_tx()
845 spin_unlock(&np->tx_lock); in rio_free_tx()
847 spin_unlock_irqrestore(&np->tx_lock, flag); in rio_free_tx()
848 np->old_tx = entry; in rio_free_tx()
854 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in rio_free_tx()
855 < TX_QUEUE_LEN - 1 || np->speed == 10)) { in rio_free_tx()
863 struct netdev_private *np = netdev_priv(dev); in tx_error() local
864 void __iomem *ioaddr = np->ioaddr; in tx_error()
887 dw32(TFDListPtr0, np->tx_ring_dma + in tx_error()
888 np->old_tx * sizeof (struct netdev_desc)); in tx_error()
908 spin_lock(&np->stats_lock); in tx_error()
914 spin_unlock(&np->stats_lock); in tx_error()
923 struct netdev_private *np = netdev_priv(dev); in receive_packet() local
924 int entry = np->cur_rx % RX_RING_SIZE; in receive_packet()
929 struct netdev_desc *desc = &np->rx_ring[entry]; in receive_packet()
950 if (frame_status & RxAlignmentError && np->speed != 1000) in receive_packet()
959 dma_unmap_single(&np->pdev->dev, in receive_packet()
961 np->rx_buf_sz, in receive_packet()
963 skb_put (skb = np->rx_skbuff[entry], pkt_len); in receive_packet()
964 np->rx_skbuff[entry] = NULL; in receive_packet()
966 dma_sync_single_for_cpu(&np->pdev->dev, in receive_packet()
968 np->rx_buf_sz, in receive_packet()
971 np->rx_skbuff[entry]->data, in receive_packet()
974 dma_sync_single_for_device(&np->pdev->dev, in receive_packet()
976 np->rx_buf_sz, in receive_packet()
982 if (np->pdev->pci_rev_id >= 0x0c && in receive_packet()
991 spin_lock(&np->rx_lock); in receive_packet()
992 np->cur_rx = entry; in receive_packet()
994 entry = np->old_rx; in receive_packet()
995 while (entry != np->cur_rx) { in receive_packet()
998 if (np->rx_skbuff[entry] == NULL) { in receive_packet()
999 skb = netdev_alloc_skb_ip_align(dev, np->rx_buf_sz); in receive_packet()
1001 np->rx_ring[entry].fraginfo = 0; in receive_packet()
1008 np->rx_skbuff[entry] = skb; in receive_packet()
1009 np->rx_ring[entry].fraginfo = in receive_packet()
1010 cpu_to_le64(dma_map_single(&np->pdev->dev, skb->data, in receive_packet()
1011 np->rx_buf_sz, DMA_FROM_DEVICE)); in receive_packet()
1013 np->rx_ring[entry].fraginfo |= in receive_packet()
1014 cpu_to_le64((u64)np->rx_buf_sz << 48); in receive_packet()
1015 np->rx_ring[entry].status = 0; in receive_packet()
1018 np->old_rx = entry; in receive_packet()
1019 spin_unlock(&np->rx_lock); in receive_packet()
1026 struct netdev_private *np = netdev_priv(dev); in rio_error() local
1027 void __iomem *ioaddr = np->ioaddr; in rio_error()
1034 if (np->phy_media) in rio_error()
1038 if (np->speed == 1000) in rio_error()
1039 np->tx_coalesce = tx_coalesce; in rio_error()
1041 np->tx_coalesce = 1; in rio_error()
1043 macctrl |= (np->vlan) ? AutoVLANuntagging : 0; in rio_error()
1044 macctrl |= (np->full_duplex) ? DuplexSelect : 0; in rio_error()
1045 macctrl |= (np->tx_flow) ? in rio_error()
1047 macctrl |= (np->rx_flow) ? in rio_error()
1050 np->link_status = 1; in rio_error()
1054 np->link_status = 0; in rio_error()
1078 struct netdev_private *np = netdev_priv(dev); in get_stats() local
1079 void __iomem *ioaddr = np->ioaddr; in get_stats()
1086 spin_lock_irqsave(&np->stats_lock, flags); in get_stats()
1137 spin_unlock_irqrestore(&np->stats_lock, flags); in get_stats()
1145 struct netdev_private *np = netdev_priv(dev); in clear_stats() local
1146 void __iomem *ioaddr = np->ioaddr; in clear_stats()
1200 struct netdev_private *np = netdev_priv(dev); in set_multicast() local
1201 void __iomem *ioaddr = np->ioaddr; in set_multicast()
1234 if (np->vlan) { in set_multicast()
1246 struct netdev_private *np = netdev_priv(dev); in rio_get_drvinfo() local
1249 strscpy(info->bus_info, pci_name(np->pdev), sizeof(info->bus_info)); in rio_get_drvinfo()
1255 struct netdev_private *np = netdev_priv(dev); in rio_get_link_ksettings() local
1258 if (np->phy_media) { in rio_get_link_ksettings()
1275 if (np->link_status) { in rio_get_link_ksettings()
1276 cmd->base.speed = np->speed; in rio_get_link_ksettings()
1277 cmd->base.duplex = np->full_duplex ? DUPLEX_FULL : DUPLEX_HALF; in rio_get_link_ksettings()
1282 if (np->an_enable) in rio_get_link_ksettings()
1287 cmd->base.phy_address = np->phy_addr; in rio_get_link_ksettings()
1300 struct netdev_private *np = netdev_priv(dev); in rio_set_link_ksettings() local
1306 if (np->an_enable) { in rio_set_link_ksettings()
1309 np->an_enable = 1; in rio_set_link_ksettings()
1314 np->an_enable = 0; in rio_set_link_ksettings()
1315 if (np->speed == 1000) { in rio_set_link_ksettings()
1322 np->speed = 10; in rio_set_link_ksettings()
1323 np->full_duplex = (duplex == DUPLEX_FULL); in rio_set_link_ksettings()
1326 np->speed = 100; in rio_set_link_ksettings()
1327 np->full_duplex = (duplex == DUPLEX_FULL); in rio_set_link_ksettings()
1340 struct netdev_private *np = netdev_priv(dev); in rio_get_link() local
1341 return np->link_status; in rio_get_link()
1355 struct netdev_private *np = netdev_priv(dev); in rio_ioctl() local
1358 phy_addr = np->phy_addr; in rio_ioctl()
1381 static int read_eeprom(struct netdev_private *np, int eep_addr) in read_eeprom() argument
1383 void __iomem *ioaddr = np->eeprom_addr; in read_eeprom()
1403 struct netdev_private *np = netdev_priv(dev); in mii_sendbit() local
1404 void __iomem *ioaddr = np->ioaddr; in mii_sendbit()
1416 struct netdev_private *np = netdev_priv(dev); in mii_getbit() local
1417 void __iomem *ioaddr = np->ioaddr; in mii_getbit()
1486 struct netdev_private *np; in mii_wait_link() local
1488 np = netdev_priv(dev); in mii_wait_link()
1489 phy_addr = np->phy_addr; in mii_wait_link()
1507 struct netdev_private *np; in mii_get_media() local
1509 np = netdev_priv(dev); in mii_get_media()
1510 phy_addr = np->phy_addr; in mii_get_media()
1513 if (np->an_enable) { in mii_get_media()
1523 np->speed = 1000; in mii_get_media()
1524 np->full_duplex = 1; in mii_get_media()
1527 np->speed = 1000; in mii_get_media()
1528 np->full_duplex = 0; in mii_get_media()
1531 np->speed = 100; in mii_get_media()
1532 np->full_duplex = 1; in mii_get_media()
1535 np->speed = 100; in mii_get_media()
1536 np->full_duplex = 0; in mii_get_media()
1539 np->speed = 10; in mii_get_media()
1540 np->full_duplex = 1; in mii_get_media()
1543 np->speed = 10; in mii_get_media()
1544 np->full_duplex = 0; in mii_get_media()
1548 np->tx_flow &= 1; in mii_get_media()
1549 np->rx_flow &= 1; in mii_get_media()
1551 np->tx_flow = 0; in mii_get_media()
1552 np->rx_flow &= 1; in mii_get_media()
1573 if (np->tx_flow) in mii_get_media()
1577 if (np->rx_flow) in mii_get_media()
1593 struct netdev_private *np; in mii_set_media() local
1594 np = netdev_priv(dev); in mii_set_media()
1595 phy_addr = np->phy_addr; in mii_set_media()
1598 if (np->an_enable) { in mii_set_media()
1650 if (np->speed == 100) { in mii_set_media()
1653 } else if (np->speed == 10) { in mii_set_media()
1656 if (np->full_duplex) { in mii_set_media()
1680 struct netdev_private *np; in mii_get_media_pcs() local
1682 np = netdev_priv(dev); in mii_get_media_pcs()
1683 phy_addr = np->phy_addr; in mii_get_media_pcs()
1686 if (np->an_enable) { in mii_get_media_pcs()
1693 np->speed = 1000; in mii_get_media_pcs()
1696 np->full_duplex = 1; in mii_get_media_pcs()
1699 np->full_duplex = 0; in mii_get_media_pcs()
1702 np->tx_flow &= 1; in mii_get_media_pcs()
1703 np->rx_flow &= 1; in mii_get_media_pcs()
1705 np->tx_flow = 0; in mii_get_media_pcs()
1706 np->rx_flow &= 1; in mii_get_media_pcs()
1718 if (np->tx_flow) in mii_get_media_pcs()
1722 if (np->rx_flow) in mii_get_media_pcs()
1737 struct netdev_private *np; in mii_set_media_pcs() local
1738 np = netdev_priv(dev); in mii_set_media_pcs()
1739 phy_addr = np->phy_addr; in mii_set_media_pcs()
1742 if (np->an_enable) { in mii_set_media_pcs()
1766 if (np->full_duplex) { in mii_set_media_pcs()
1786 struct netdev_private *np = netdev_priv(dev); in rio_close() local
1787 struct pci_dev *pdev = np->pdev; in rio_close()
1794 del_timer_sync (&np->timer); in rio_close()
1807 struct netdev_private *np = netdev_priv(dev); in rio_remove1() local
1810 dma_free_coherent(&pdev->dev, RX_TOTAL_SIZE, np->rx_ring, in rio_remove1()
1811 np->rx_ring_dma); in rio_remove1()
1812 dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, np->tx_ring, in rio_remove1()
1813 np->tx_ring_dma); in rio_remove1()
1815 pci_iounmap(pdev, np->ioaddr); in rio_remove1()
1817 pci_iounmap(pdev, np->eeprom_addr); in rio_remove1()
1828 struct netdev_private *np = netdev_priv(dev); in rio_suspend() local
1834 del_timer_sync(&np->timer); in rio_suspend()
1843 struct netdev_private *np = netdev_priv(dev); in rio_resume() local
1848 rio_reset_ring(np); in rio_resume()
1850 np->timer.expires = jiffies + 1 * HZ; in rio_resume()
1851 add_timer(&np->timer); in rio_resume()
1853 dl2k_enable_int(np); in rio_resume()