Home
last modified time | relevance | path

Searched refs:cur_tx (Results 1 – 25 of 55) sorted by relevance

123

/drivers/net/ethernet/amd/
Dariadne.c92 int cur_tx, cur_rx; /* The next free ring entry */ member
125 priv->cur_rx = priv->cur_tx = 0; in ariadne_init_ring()
308 while (dirty_tx < priv->cur_tx) { in ariadne_interrupt()
345 if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) { in ariadne_interrupt()
347 dirty_tx, priv->cur_tx, in ariadne_interrupt()
354 dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) { in ariadne_interrupt()
576 entry = priv->cur_tx % TX_RING_SIZE; in ariadne_start_xmit()
596 priv->cur_tx++; in ariadne_start_xmit()
597 if ((priv->cur_tx >= TX_RING_SIZE) && in ariadne_start_xmit()
601 priv->cur_tx, priv->dirty_tx); in ariadne_start_xmit()
[all …]
Datarilance.c222 int cur_rx, cur_tx; /* The next free ring entry */ member
690 lp->cur_rx = lp->cur_tx = 0; in lance_init_ring()
749 lp->dirty_tx, lp->cur_tx, in lance_tx_timeout()
816 entry = lp->cur_tx & TX_RING_MOD_MASK; in lance_start_xmit()
830 lp->cur_tx++; in lance_start_xmit()
831 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { in lance_start_xmit()
832 lp->cur_tx -= TX_RING_SIZE; in lance_start_xmit()
886 while( dirty_tx < lp->cur_tx) { in lance_interrupt()
922 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt()
925 dirty_tx, lp->cur_tx, lp->tx_full )); in lance_interrupt()
[all …]
Dlance.c248 int cur_rx, cur_tx; /* The next free ring entry */ member
867 lp->cur_rx = lp->cur_tx = 0; in lance_init_ring()
931 lp->dirty_tx, lp->cur_tx, netif_queue_stopped(dev) ? " (full)" : "", in lance_tx_timeout()
971 entry = lp->cur_tx & TX_RING_MOD_MASK; in lance_start_xmit()
1006 lp->cur_tx++; in lance_start_xmit()
1012 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) in lance_start_xmit()
1051 while (dirty_tx < lp->cur_tx) { in lance_interrupt()
1095 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt()
1097 dirty_tx, lp->cur_tx, in lance_interrupt()
1105 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) in lance_interrupt()
/drivers/net/ethernet/packetengines/
Dyellowfin.c326 unsigned int cur_tx, dirty_tx; member
689 yp->cur_tx, yp->dirty_tx, in yellowfin_tx_timeout()
714 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) in yellowfin_tx_timeout()
728 yp->cur_rx = yp->cur_tx = 0; in yellowfin_init_ring()
820 entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit()
850 yp->cur_tx++; in yellowfin_start_xmit()
858 yp->cur_tx++; in yellowfin_start_xmit()
860 unsigned next_entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit()
875 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) in yellowfin_start_xmit()
882 yp->cur_tx, entry); in yellowfin_start_xmit()
[all …]
Dhamachi.c495 unsigned int cur_tx, dirty_tx; member
998 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++) { in hamachi_tx()
1112 hmp->cur_rx = hmp->cur_tx = 0; in hamachi_tx_timeout()
1166 hmp->cur_rx = hmp->cur_tx = 0; in hamachi_init_ring()
1221 printk(KERN_WARNING "%s: Hamachi transmit queue full at slot %d.\n",dev->name, hmp->cur_tx); in hamachi_start_xmit()
1235 entry = hmp->cur_tx % TX_RING_SIZE; in hamachi_start_xmit()
1257 hmp->cur_tx++; in hamachi_start_xmit()
1276 if ((hmp->cur_tx - hmp->dirty_tx) < (TX_RING_SIZE - 4)) in hamachi_start_xmit()
1285 dev->name, hmp->cur_tx, entry); in hamachi_start_xmit()
1330 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++){ in hamachi_interrupt()
[all …]
/drivers/net/arcnet/
Darcnet.c473 lp->cur_tx = lp->next_tx = -1; in arcnet_open()
609 lp->hw.status(dev), lp->cur_tx, lp->next_tx, skb->len, skb->protocol); in arcnet_send_packet()
690 lp->hw.status(dev), lp->intmask, lp->next_tx, lp->cur_tx); in go_tx()
692 if (lp->cur_tx != -1 || lp->next_tx == -1) in go_tx()
698 lp->cur_tx = lp->next_tx; in go_tx()
702 lp->hw.command(dev, TXcmd | (lp->cur_tx << 3)); in go_tx()
728 lp->hw.command(dev, NOTXcmd | (lp->cur_tx << 3)); in arcnet_timeout()
745 if (lp->cur_tx == -1) in arcnet_timeout()
850 if (lp->cur_tx != -1 && !lp->timed_out) { in arcnet_interrupt()
882 if (lp->cur_tx != -1) in arcnet_interrupt()
[all …]
/drivers/net/ethernet/
Dfealnx.c403 struct fealnx_desc *cur_tx; member
1157 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), in enable_rxtx()
1280 np->cur_tx = &np->tx_ring[0]; in init_ring()
1385 np->cur_tx = &np->tx_ring[0]; in reset_tx_descriptors()
1494 long tx_status = np->cur_tx->status; in intr_handler()
1495 long tx_control = np->cur_tx->control; in intr_handler()
1500 next = np->cur_tx->next_desc_logical; in intr_handler()
1537 pci_unmap_single(np->pci_dev, np->cur_tx->buffer, in intr_handler()
1538 np->cur_tx->skbuff->len, PCI_DMA_TODEVICE); in intr_handler()
1539 dev_kfree_skb_irq(np->cur_tx->skbuff); in intr_handler()
[all …]
Dethoc.c208 unsigned int cur_tx; member
309 dev->cur_tx = 0; in ethoc_init_ring()
526 if (bd.stat & TX_BD_READY || (priv->dty_tx == priv->cur_tx)) { in ethoc_tx()
537 (priv->dty_tx == priv->cur_tx)) in ethoc_tx()
545 if ((priv->cur_tx - priv->dty_tx) <= (priv->num_tx / 2)) in ethoc_tx()
869 entry = priv->cur_tx % priv->num_tx; in ethoc_start_xmit()
871 priv->cur_tx++; in ethoc_start_xmit()
889 if (priv->cur_tx == (priv->dty_tx + priv->num_tx)) { in ethoc_start_xmit()
/drivers/net/ethernet/smsc/
Depic100.c268 unsigned int cur_tx, dirty_tx; member
790 ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx); in epic_restart()
881 ep->dirty_tx, ep->cur_tx); in epic_tx_timeout()
905 ep->dirty_tx = ep->cur_tx = 0; in epic_init_ring()
960 free_count = ep->cur_tx - ep->dirty_tx; in epic_start_xmit()
961 entry = ep->cur_tx % TX_RING_SIZE; in epic_start_xmit()
982 ep->cur_tx++; in epic_start_xmit()
1021 unsigned int dirty_tx, cur_tx; in epic_tx() local
1027 cur_tx = ep->cur_tx; in epic_tx()
1028 for (dirty_tx = ep->dirty_tx; cur_tx - dirty_tx > 0; dirty_tx++) { in epic_tx()
[all …]
/drivers/net/ethernet/dec/tulip/
Dtulip_core.c334 tp->cur_rx = tp->cur_tx = 0; in tulip_up()
367 tp->tx_buffers[tp->cur_tx].skb = NULL; in tulip_up()
368 tp->tx_buffers[tp->cur_tx].mapping = mapping; in tulip_up()
371 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); in tulip_up()
372 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); in tulip_up()
373 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); in tulip_up()
375 tp->cur_tx++; in tulip_up()
677 entry = tp->cur_tx % TX_RING_SIZE; in tulip_start_xmit()
685 if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */ in tulip_start_xmit()
687 } else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) { in tulip_start_xmit()
[all …]
Dwinbond-840.c314 unsigned int cur_tx, dirty_tx; member
832 np->tx_q_bytes = np->dirty_tx = np->cur_tx = 0; in init_rxtx_rings()
946 np->cur_tx, np->dirty_tx, np->tx_full, np->tx_q_bytes); in tx_timeout()
1006 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1035 np->cur_tx++; in start_tx()
1044 if (np->cur_tx - np->dirty_tx > TX_QUEUE_LEN || in start_tx()
1054 np->cur_tx, entry); in start_tx()
1062 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { in netdev_tx_done()
1100 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN_RESTART && in netdev_tx_done()
1141 np->cur_tx != np->dirty_tx) { in intr_handler()
[all …]
Dinterrupt.c588 for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0; in tulip_interrupt()
644 if (tp->cur_tx - dirty_tx > TX_RING_SIZE) { in tulip_interrupt()
647 dirty_tx, tp->cur_tx); in tulip_interrupt()
652 if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2) in tulip_interrupt()
/drivers/net/ethernet/dlink/
Dsundance.c398 unsigned int cur_tx, dirty_tx; member
1001 np->cur_tx, np->cur_tx % TX_RING_SIZE, in tx_timeout()
1016 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { in tx_timeout()
1030 np->cur_rx = np->cur_tx = 0; in init_ring()
1078 &np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE]; in tx_poll()
1081 for (; np->cur_tx - np->cur_task > 0; np->cur_task++) { in tx_poll()
1106 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1120 np->cur_tx++; in start_tx()
1126 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 1 && in start_tx()
1135 dev->name, np->cur_tx, entry); in start_tx()
[all …]
Ddl2k.c597 np->cur_rx = np->cur_tx = 0; in alloc_list()
656 entry = np->cur_tx % TX_RING_SIZE; in start_xmit()
693 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; in start_xmit()
694 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in start_xmit()
746 if (np->cur_tx != np->old_tx) in rio_interrupt()
770 while (entry != np->cur_tx) { in rio_free_tx()
798 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in rio_free_tx()
/drivers/net/ethernet/3com/
D3c515.c310 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member
842 vp->cur_tx = vp->dirty_tx = 0; in corkscrew_open()
980 vp->cur_tx); in corkscrew_timeout()
1013 int entry = vp->cur_tx % TX_RING_SIZE; in corkscrew_start_xmit()
1020 if (vp->cur_tx != 0) in corkscrew_start_xmit()
1021 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; in corkscrew_start_xmit()
1026 dev->name, vp->cur_tx); in corkscrew_start_xmit()
1050 vp->cur_tx++; in corkscrew_start_xmit()
1051 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) in corkscrew_start_xmit()
1175 while (lp->cur_tx - dirty_tx > 0) { in corkscrew_interrupt()
[all …]
D3c59x.c604 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member
1694 vp->cur_tx = vp->dirty_tx = 0; in vortex_up()
1936 if (vp->cur_tx - vp->dirty_tx > 0 && ioread32(ioaddr + DownListPtr) == 0) in vortex_tx_timeout()
1939 if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) { in vortex_tx_timeout()
2132 int entry = vp->cur_tx % TX_RING_SIZE; in boomerang_start_xmit()
2134 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit()
2141 dev->name, vp->cur_tx); in boomerang_start_xmit()
2153 if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) { in boomerang_start_xmit()
2238 vp->cur_tx++; in boomerang_start_xmit()
2241 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) { in boomerang_start_xmit()
[all …]
/drivers/net/ethernet/stmicro/stmmac/
Dring_mode.c35 unsigned int entry = priv->cur_tx % txsize; in stmmac_jumbo_frm()
65 entry = (++priv->cur_tx) % txsize; in stmmac_jumbo_frm()
Dchain_mode.c35 unsigned int entry = priv->cur_tx % txsize; in stmmac_jumbo_frm()
57 entry = (++priv->cur_tx) % txsize; in stmmac_jumbo_frm()
Dstmmac_main.c204 return priv->dirty_tx + priv->dma_tx_size - priv->cur_tx - 1; in stmmac_tx_avail()
230 if ((priv->dirty_tx == priv->cur_tx) && in stmmac_enable_eee_mode()
1118 priv->cur_tx = 0; in init_dma_desc_rings()
1333 while (priv->dirty_tx != priv->cur_tx) { in stmmac_tx_clean()
1365 priv->cur_tx, priv->dirty_tx); in stmmac_tx_clean()
1450 priv->cur_tx = 0; in stmmac_tx_err()
1980 entry = priv->cur_tx % txsize; in stmmac_xmit()
2015 entry = (++priv->cur_tx) % txsize; in stmmac_xmit()
2058 priv->cur_tx++; in stmmac_xmit()
2062 __func__, (priv->cur_tx % txsize), in stmmac_xmit()
[all …]
/drivers/net/ethernet/nuvoton/
Dw90p910_ether.c166 unsigned int cur_tx; member
480 ether->cur_tx = 0x0; in w90p910_reset_mac()
607 txbd = &ether->tdesc->desclist[ether->cur_tx]; in w90p910_send_frame()
608 buffer = ether->tdesc->tran_buf[ether->cur_tx]; in w90p910_send_frame()
625 if (++ether->cur_tx >= TX_DESC_SIZE) in w90p910_send_frame()
626 ether->cur_tx = 0; in w90p910_send_frame()
628 txbd = &ether->tdesc->desclist[ether->cur_tx]; in w90p910_send_frame()
955 ether->cur_tx = 0x0; in w90p910_ether_setup()
/drivers/net/ethernet/adaptec/
Dstarfire.c560 unsigned int cur_tx, dirty_tx, reap_tx; member
1144 np->cur_rx = np->cur_tx = np->reap_tx = 0; in init_ring()
1200 if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) { in start_tx()
1212 prev_tx = np->cur_tx; in start_tx()
1213 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1256 dev->name, np->cur_tx, np->dirty_tx, in start_tx()
1260 np->cur_tx += np->tx_info[entry].used_slots; in start_tx()
1264 np->cur_tx += np->tx_info[entry].used_slots; in start_tx()
1268 if (np->cur_tx % (TX_RING_SIZE / 2) == 0) in start_tx()
1281 if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE) in start_tx()
[all …]
/drivers/net/ethernet/samsung/sxgbe/
Dsxgbe_main.c195 return queue->dirty_tx + tx_qsize - queue->cur_tx - 1; in sxgbe_tx_avail()
427 tx_ring->cur_tx = 0; in init_tx_ring()
749 while (tqueue->dirty_tx != tqueue->cur_tx) { in sxgbe_tx_queue_clean()
762 __func__, tqueue->cur_tx, tqueue->dirty_tx); in sxgbe_tx_queue_clean()
841 tx_ring->cur_tx = 0; in sxgbe_restart_tx_queue()
1325 entry = tqueue->cur_tx % tx_rsize; in sxgbe_xmit()
1350 entry = (++tqueue->cur_tx) % tx_rsize; in sxgbe_xmit()
1372 entry = (++tqueue->cur_tx) % tx_rsize; in sxgbe_xmit()
1412 tqueue->cur_tx++; in sxgbe_xmit()
1416 __func__, tqueue->cur_tx % tx_rsize, in sxgbe_xmit()
/drivers/net/ethernet/realtek/
D8139too.c592 unsigned long cur_tx; member
1460 tp->cur_tx = 0; in rtl8139_init_ring()
1655 tp->cur_tx = 0; in rtl8139_tx_clear()
1675 tp->cur_tx, tp->dirty_tx); in rtl8139_tx_timeout_task()
1727 entry = tp->cur_tx % NUM_TX_DESC; in rtl8139_start_xmit()
1751 tp->cur_tx++; in rtl8139_start_xmit()
1753 if ((tp->cur_tx - NUM_TX_DESC) == tp->dirty_tx) in rtl8139_start_xmit()
1774 tx_left = tp->cur_tx - dirty_tx; in rtl8139_tx_interrupt()
1819 if (tp->cur_tx - dirty_tx > NUM_TX_DESC) { in rtl8139_tx_interrupt()
1821 dirty_tx, tp->cur_tx); in rtl8139_tx_interrupt()
/drivers/net/ethernet/via/
Dvia-rhine.c475 unsigned int cur_tx, dirty_tx; member
1326 rp->dirty_tx = rp->cur_tx = 0; in alloc_tbufs()
1783 return (rp->cur_tx - rp->dirty_tx) >= TX_QUEUE_LEN; in rhine_tx_queue_full()
1798 entry = rp->cur_tx % TX_RING_SIZE; in rhine_start_tx()
1860 rp->cur_tx++; in rhine_start_tx()
1889 rp->cur_tx - 1, entry); in rhine_start_tx()
1936 unsigned int cur_tx; in rhine_tx() local
1946 cur_tx = rp->cur_tx; in rhine_tx()
1948 while (dirty_tx != cur_tx) { in rhine_tx()
/drivers/net/ethernet/sis/
Dsis190.c272 u32 cur_tx; member
695 pending = tp->cur_tx - dirty_tx; in sis190_tx_interrupt()
808 tp->dirty_tx = tp->dirty_rx = tp->cur_tx = tp->cur_rx = 0; in sis190_init_ring_indexes()
1122 tp->cur_tx = tp->dirty_tx = 0; in sis190_tx_clear()
1192 entry = tp->cur_tx % NUM_TX_DESC; in sis190_start_xmit()
1228 tp->cur_tx++; in sis190_start_xmit()
1235 if ((tp->cur_tx - NUM_TX_DESC) == dirty_tx) { in sis190_start_xmit()

123