Home
last modified time | relevance | path

Searched refs:tx_head (Results 1 – 25 of 44) sorted by relevance

12

/drivers/net/ethernet/moxa/
Dmoxart_ether.c137 priv->tx_head = 0; in moxart_mac_setup_desc_ring()
286 return CIRC_SPACE(priv->tx_head, priv->tx_tail, TX_DESC_NUM); in moxart_tx_queue_space()
292 unsigned tx_head = priv->tx_head; in moxart_tx_finished() local
295 while (tx_tail != tx_head) { in moxart_tx_finished()
338 unsigned int tx_head; in moxart_mac_start_xmit() local
344 tx_head = priv->tx_head; in moxart_mac_start_xmit()
345 desc = priv->tx_desc_base + (TX_REG_DESC_SIZE * tx_head); in moxart_mac_start_xmit()
359 priv->tx_mapping[tx_head] = dma_map_single(&ndev->dev, skb->data, in moxart_mac_start_xmit()
361 if (dma_mapping_error(&ndev->dev, priv->tx_mapping[tx_head])) { in moxart_mac_start_xmit()
366 priv->tx_len[tx_head] = len; in moxart_mac_start_xmit()
[all …]
/drivers/net/ethernet/amd/
Dsun3lance.c141 struct lance_tx_head tx_head[TX_RING_SIZE]; member
398 MEM->init.tdra = dvma_vtob(MEM->tx_head); in lance_probe()
400 (dvma_vtob(MEM->tx_head) >> 16); in lance_probe()
404 (dvma_vtob(MEM->tx_head)))); in lance_probe()
465 MEM->tx_head[i].base = dvma_vtob(MEM->tx_data[i]); in lance_init_ring()
466 MEM->tx_head[i].flag = 0; in lance_init_ring()
467 MEM->tx_head[i].base_hi = in lance_init_ring()
469 MEM->tx_head[i].length = 0; in lance_init_ring()
470 MEM->tx_head[i].misc = 0; in lance_init_ring()
496 MEM->init.tdra = dvma_vtob(MEM->tx_head); in lance_init_ring()
[all …]
Datarilance.c157 struct lance_tx_head tx_head[TX_RING_SIZE]; member
620 MEM->init.tx_ring.adr_lo = offsetof( struct lance_memory, tx_head ); in lance_probe1()
707 MEM->tx_head[i].base = offset; in lance_init_ring()
708 MEM->tx_head[i].flag = TMD1_OWN_HOST; in lance_init_ring()
709 MEM->tx_head[i].base_hi = 0; in lance_init_ring()
710 MEM->tx_head[i].length = 0; in lance_init_ring()
711 MEM->tx_head[i].misc = 0; in lance_init_ring()
758 i, MEM->tx_head[i].base, in lance_tx_timeout()
759 -MEM->tx_head[i].length, in lance_tx_timeout()
760 MEM->tx_head[i].misc )); in lance_tx_timeout()
[all …]
Dau1000_eth.c737 aup->tx_head = (aup->tx_dma_ring[0]->buff_stat & 0xC) >> 2; in au1000_init()
738 aup->tx_tail = aup->tx_head; in au1000_init()
1004 skb->data, aup->tx_head); in au1000_tx()
1006 ptxd = aup->tx_dma_ring[aup->tx_head]; in au1000_tx()
1023 pDB = aup->tx_db_inuse[aup->tx_head]; in au1000_tx()
1039 aup->tx_head = (aup->tx_head + 1) & (NUM_TX_DMA - 1); in au1000_tx()
Dau1000_eth.h95 u32 tx_head; member
/drivers/infiniband/hw/hfi1/
Diowait.h98 struct list_head tx_head; member
145 INIT_LIST_HEAD(&wait->tx_head); in iowait_init()
290 if (!list_empty(&wait->tx_head)) { in iowait_get_txhead()
292 &wait->tx_head, in iowait_get_txhead()
Dtrace_tx.h269 __field(u16, tx_head)
276 __entry->tx_head = sde->tx_head;
289 __entry->tx_head,
305 __field(u16, tx_head)
312 __entry->tx_head = sde->tx_head;
323 __entry->tx_head,
/drivers/staging/vt6656/
Drxtx.c587 union vnt_tx_head *tx_head, bool need_mic) in vnt_rxtx_rts() argument
590 struct vnt_rrv_time_rts *buf = &tx_head->tx_rts.rts; in vnt_rxtx_rts()
591 union vnt_tx_data_head *head = &tx_head->tx_rts.tx.head; in vnt_rxtx_rts()
610 head = &tx_head->tx_rts.tx.mic.head; in vnt_rxtx_rts()
619 union vnt_tx_head *tx_head, bool need_mic) in vnt_rxtx_cts() argument
622 struct vnt_rrv_time_cts *buf = &tx_head->tx_cts.cts; in vnt_rxtx_cts()
623 union vnt_tx_data_head *head = &tx_head->tx_cts.tx.head; in vnt_rxtx_cts()
637 head = &tx_head->tx_cts.tx.mic.head; in vnt_rxtx_cts()
647 union vnt_tx_head *tx_head, bool need_rts, bool need_mic) in vnt_rxtx_ab() argument
650 struct vnt_rrv_time_ab *buf = &tx_head->tx_ab.ab; in vnt_rxtx_ab()
[all …]
/drivers/net/ethernet/i825xx/
Dether1.c504 priv(dev)->tx_head = NOP_ADDR + NOP_SIZE; in ether1_init_for_open()
621 if (priv(dev)->tx_head + size > TX_AREA_END) { in ether1_txalloc()
622 if (tail > priv(dev)->tx_head) in ether1_txalloc()
627 priv(dev)->tx_head = start + size; in ether1_txalloc()
629 if (priv(dev)->tx_head < tail && (priv(dev)->tx_head + size) > tail) in ether1_txalloc()
631 start = priv(dev)->tx_head; in ether1_txalloc()
632 priv(dev)->tx_head += size; in ether1_txalloc()
731 tmp = priv(dev)->tx_head; in ether1_sendpacket()
733 priv(dev)->tx_head = tmp; in ether1_sendpacket()
836 caddr = priv(dev)->tx_head; in ether1_xmit_done()
[all …]
/drivers/net/ethernet/hisilicon/
Dhip04_eth.c166 unsigned int tx_head; member
376 count = tx_count(ACCESS_ONCE(priv->tx_head), tx_tail); in hip04_tx_reclaim()
429 unsigned int tx_head = priv->tx_head, count; in hip04_mac_start_xmit() local
430 struct tx_desc *desc = &priv->tx_desc[tx_head]; in hip04_mac_start_xmit()
434 count = tx_count(tx_head, ACCESS_ONCE(priv->tx_tail)); in hip04_mac_start_xmit()
446 priv->tx_skb[tx_head] = skb; in hip04_mac_start_xmit()
447 priv->tx_phys[tx_head] = phys; in hip04_mac_start_xmit()
451 phys = priv->tx_desc_dma + tx_head * sizeof(struct tx_desc); in hip04_mac_start_xmit()
456 priv->tx_head = TX_NEXT(tx_head); in hip04_mac_start_xmit()
636 priv->tx_head = 0; in hip04_mac_open()
/drivers/net/irda/
Dau1k_ir.c158 u32 tx_head; member
310 aup->tx_head = aup->tx_tail = aup->rx_head = 0; in au1k_irda_set_speed()
423 while (!(ptxd->flags & AU_OWN) && (aup->tx_tail != aup->tx_head)) { in au1k_tx_ack()
438 if (aup->tx_tail == aup->tx_head) { in au1k_tx_ack()
553 aup->tx_head = 0; in au1k_init()
672 if (aup->tx_tail == aup->tx_head) { in au1k_irda_hard_xmit()
680 ptxd = aup->tx_ring[aup->tx_head]; in au1k_irda_hard_xmit()
688 } else if (((aup->tx_head + 1) & (NUM_IR_DESC - 1)) == aup->tx_tail) { in au1k_irda_hard_xmit()
695 pDB = aup->tx_db_inuse[aup->tx_head]; in au1k_irda_hard_xmit()
724 aup->tx_head = (aup->tx_head + 1) & (NUM_IR_DESC - 1); in au1k_irda_hard_xmit()
/drivers/net/can/
Dti_hecc.c222 u32 tx_head; member
230 return priv->tx_head & HECC_TX_MB_MASK; in get_tx_head_mb()
240 return (priv->tx_head >> HECC_TX_PRIO_SHIFT) & MAX_TX_PRIO; in get_tx_head_prio()
392 priv->tx_head = priv->tx_tail = HECC_TX_MASK; in ti_hecc_start()
503 priv->tx_head, priv->tx_tail); in ti_hecc_xmit()
529 --priv->tx_head; in ti_hecc_xmit()
531 (priv->tx_head & HECC_TX_MASK) == HECC_TX_MASK) { in ti_hecc_xmit()
777 while (priv->tx_tail - priv->tx_head > 0) { in ti_hecc_interrupt()
796 if (((priv->tx_head == priv->tx_tail) && in ti_hecc_interrupt()
797 ((priv->tx_head & HECC_TX_MASK) != HECC_TX_MASK)) || in ti_hecc_interrupt()
[all …]
Dxilinx_can.c135 unsigned int tx_head; member
441 can_put_echo_skb(skb, ndev, priv->tx_head % priv->tx_max); in xcan_start_xmit()
442 priv->tx_head++; in xcan_start_xmit()
458 if ((priv->tx_head - priv->tx_tail) == priv->tx_max) in xcan_start_xmit()
747 while ((priv->tx_head - priv->tx_tail > 0) && in xcan_tx_interrupt()
/drivers/net/ethernet/seeq/
Dether3.c337 priv(dev)->tx_head = 0; in ether3_init_for_open()
468 priv(dev)->tx_head, priv(dev)->tx_tail); in ether3_timeout()
476 priv(dev)->tx_head = priv(dev)->tx_tail = 0; in ether3_timeout()
504 next_ptr = (priv(dev)->tx_head + 1) & 15; in ether3_sendpacket()
513 ptr = 0x600 * priv(dev)->tx_head; in ether3_sendpacket()
514 priv(dev)->tx_head = next_ptr; in ether3_sendpacket()
536 next_ptr = (priv(dev)->tx_head + 1) & 15; in ether3_sendpacket()
Dether3.h164 unsigned char tx_head; /* buffer nr to insert next packet */ member
/drivers/net/ethernet/cadence/
Dmacb.c565 queue->tx_tail, queue->tx_head); in macb_tx_error_task()
589 for (tail = queue->tx_tail; tail != queue->tx_head; tail++) { in macb_tx_error_task()
644 queue->tx_head = 0; in macb_tx_error_task()
675 head = queue->tx_head; in macb_tx_interrupt()
722 CIRC_CNT(queue->tx_head, queue->tx_tail, in macb_tx_interrupt()
1220 unsigned int len, entry, i, tx_head = queue->tx_head; in macb_tx_map() local
1233 entry = macb_tx_ring_wrap(tx_head); in macb_tx_map()
1251 tx_head++; in macb_tx_map()
1262 entry = macb_tx_ring_wrap(tx_head); in macb_tx_map()
1279 tx_head++; in macb_tx_map()
[all …]
/drivers/net/ethernet/dec/tulip/
Dde2104x.c102 (((CP)->tx_tail <= (CP)->tx_head) ? \
103 (CP)->tx_tail + (DE_TX_RING_SIZE - 1) - (CP)->tx_head : \
104 (CP)->tx_tail - (CP)->tx_head - 1)
298 unsigned tx_head; member
509 de->rx_tail, de->tx_head, de->tx_tail); in de_interrupt()
544 unsigned tx_head = de->tx_head; in de_tx() local
547 while (tx_tail != tx_head) { in de_tx()
623 entry = de->tx_head; in de_start_xmit()
644 de->tx_head = NEXT_TX(entry); in de_start_xmit()
753 entry = de->tx_head; in __de_set_rx_mode()
[all …]
/drivers/net/can/rcar/
Drcar_can.c93 u32 tx_head; member
378 if (priv->tx_head - priv->tx_tail <= unsent) in rcar_can_tx_done()
614 priv->tx_dlc[priv->tx_head % RCAR_CAN_FIFO_DEPTH] = cf->can_dlc; in rcar_can_start_xmit()
615 can_put_echo_skb(skb, ndev, priv->tx_head % RCAR_CAN_FIFO_DEPTH); in rcar_can_start_xmit()
616 priv->tx_head++; in rcar_can_start_xmit()
623 if (priv->tx_head - priv->tx_tail >= RCAR_CAN_FIFO_DEPTH) in rcar_can_start_xmit()
/drivers/infiniband/ulp/ipoib/
Dipoib_ib.c602 tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)]; in ipoib_send()
625 rc = post_send(priv, priv->tx_head & (ipoib_sendq_size - 1), in ipoib_send()
638 address->last_send = priv->tx_head; in ipoib_send()
639 ++priv->tx_head; in ipoib_send()
870 while (priv->tx_head != priv->tx_tail || recvs_pending(dev)) { in ipoib_ib_dev_stop()
873 priv->tx_head - priv->tx_tail, recvs_pending(dev)); in ipoib_ib_dev_stop()
879 while ((int) priv->tx_tail - (int) priv->tx_head < 0) { in ipoib_ib_dev_stop()
Dipoib_cm.c742 tx->tx_head, skb->len, tx->qp->qp_num); in ipoib_cm_send()
751 tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; in ipoib_cm_send()
763 rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), tx_req); in ipoib_cm_send()
771 ++tx->tx_head; in ipoib_cm_send()
1192 p->qp ? p->qp->qp_num : 0, p->tx_head, p->tx_tail); in ipoib_cm_tx_destroy()
1200 while ((int) p->tx_tail - (int) p->tx_head < 0) { in ipoib_cm_tx_destroy()
1203 p->tx_head - p->tx_tail); in ipoib_cm_tx_destroy()
1213 while ((int) p->tx_tail - (int) p->tx_head < 0) { in ipoib_cm_tx_destroy()
/drivers/net/ethernet/realtek/
D8139cp.c116 (((CP)->tx_tail <= (CP)->tx_head) ? \
117 (CP)->tx_tail + (CP_TX_RING_SIZE - 1) - (CP)->tx_head : \
118 (CP)->tx_tail - (CP)->tx_head - 1)
341 unsigned tx_head ____cacheline_aligned;
652 unsigned tx_head = cp->tx_head; in cp_tx() local
656 while (tx_tail != tx_head) { in cp_tx()
753 entry = cp->tx_head; in cp_start_xmit()
869 cp->tx_head = NEXT_TX(entry); in cp_start_xmit()
968 cp->tx_head = cp->tx_tail = 0; in cp_stop_hw()
1101 cp->tx_head = cp->tx_tail = 0; in cp_init_rings_index()
[all …]
/drivers/net/ethernet/silan/
Dsc92031.c280 unsigned tx_head; member
411 while (priv->tx_head - priv->tx_tail > 0) { in _sc92031_tx_clear()
415 priv->tx_head = priv->tx_tail = 0; in _sc92031_tx_clear()
658 while (priv->tx_head - priv->tx_tail > 0) { in _sc92031_tx_tasklet()
956 BUG_ON(priv->tx_head - priv->tx_tail >= NUM_TX_DESC); in sc92031_start_xmit()
958 entry = priv->tx_head++ % NUM_TX_DESC; in sc92031_start_xmit()
983 if (priv->tx_head - priv->tx_tail >= NUM_TX_DESC) in sc92031_start_xmit()
1014 priv->tx_head = priv->tx_tail = 0; in sc92031_open()
/drivers/net/ethernet/calxeda/
Dxgmac.c372 unsigned int tx_head; member
411 dma_ring_space((p)->tx_head, (p)->tx_tail, DMA_TX_RING_SZ)
780 priv->tx_head = 0; in xgmac_dma_desc_rings_init()
878 while (dma_ring_cnt(priv->tx_head, priv->tx_tail, DMA_TX_RING_SZ)) { in xgmac_tx_complete()
888 priv->tx_head, priv->tx_tail); in xgmac_tx_complete()
935 priv->tx_head = 0; in xgmac_tx_timeout_work()
1101 entry = priv->tx_head; in xgmac_xmit()
1146 priv->tx_head = dma_ring_incr(entry, DMA_TX_RING_SZ); in xgmac_xmit()
1160 entry = priv->tx_head; in xgmac_xmit()
/drivers/net/can/mscan/
Dmscan.c146 INIT_LIST_HEAD(&priv->tx_head); in mscan_start()
281 list_add_tail(&priv->tx_queue[buf_id].list, &priv->tx_head); in mscan_start_xmit()
451 list_for_each_safe(pos, tmp, &priv->tx_head) { in mscan_isr()
467 if (list_empty(&priv->tx_head)) { in mscan_isr()
/drivers/net/ethernet/ti/
Dtlan.c1083 dev->name, priv->tx_head, priv->tx_tail); in tlan_start_tx()
1379 priv->tx_head, priv->tx_tail); in tlan_handle_tx_eof()
1380 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof()
1402 CIRC_INC(priv->tx_head, TLAN_NUM_TX_LISTS); in tlan_handle_tx_eof()
1403 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof()
1413 priv->tx_head, priv->tx_tail); in tlan_handle_tx_eof()
1414 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eof()
1416 + sizeof(struct tlan_list)*priv->tx_head; in tlan_handle_tx_eof()
1657 priv->tx_head, priv->tx_tail); in tlan_handle_tx_eoc()
1658 head_list = priv->tx_list + priv->tx_head; in tlan_handle_tx_eoc()
[all …]

12