Home
last modified time | relevance | path

Searched refs:tbd (Results 1 – 12 of 12) sorted by relevance

/drivers/net/ethernet/qlogic/qlge/
Dqlge_dbg.c1854 void ql_dump_tx_desc(struct tx_buf_desc *tbd) in ql_dump_tx_desc() argument
1857 le64_to_cpu((u64) tbd->addr)); in ql_dump_tx_desc()
1859 le32_to_cpu(tbd->len & TX_DESC_LEN_MASK)); in ql_dump_tx_desc()
1861 tbd->len & TX_DESC_C ? "C" : ".", in ql_dump_tx_desc()
1862 tbd->len & TX_DESC_E ? "E" : "."); in ql_dump_tx_desc()
1863 tbd++; in ql_dump_tx_desc()
1865 le64_to_cpu((u64) tbd->addr)); in ql_dump_tx_desc()
1867 le32_to_cpu(tbd->len & TX_DESC_LEN_MASK)); in ql_dump_tx_desc()
1869 tbd->len & TX_DESC_C ? "C" : ".", in ql_dump_tx_desc()
1870 tbd->len & TX_DESC_E ? "E" : "."); in ql_dump_tx_desc()
[all …]
Dqlge_main.c1317 struct tx_buf_desc *tbd = mac_iocb_ptr->tbd; in ql_map_send() local
1337 tbd->len = cpu_to_le32(len); in ql_map_send()
1338 tbd->addr = cpu_to_le64(map); in ql_map_send()
1352 tbd++; in ql_map_send()
1384 tbd->addr = cpu_to_le64(map); in ql_map_send()
1390 tbd->len = in ql_map_send()
1397 tbd = (struct tx_buf_desc *)&tx_ring_desc->oal; in ql_map_send()
1412 tbd->addr = cpu_to_le64(map); in ql_map_send()
1413 tbd->len = cpu_to_le32(skb_frag_size(frag)); in ql_map_send()
1422 tbd->len = cpu_to_le32(le32_to_cpu(tbd->len) | TX_DESC_E); in ql_map_send()
Dqlge.h1097 struct tx_buf_desc tbd[TX_DESC_PER_IOCB]; member
1142 struct tx_buf_desc tbd[TX_DESC_PER_IOCB]; member
2314 void ql_dump_tx_desc(struct tx_buf_desc *tbd);
/drivers/net/ethernet/i825xx/
Dlib82596.c202 u32 tbd; member
972 struct i596_tbd *tbd; in i596_start_xmit() local
988 tbd = lp->dma->tbds + lp->next_tx_cmd; in i596_start_xmit()
1000 tx_cmd->tbd = SWAP32(virt_to_dma(lp, tbd)); in i596_start_xmit()
1001 tbd->next = I596_NULL; in i596_start_xmit()
1008 tbd->pad = 0; in i596_start_xmit()
1009 tbd->size = SWAP16(EOF | length); in i596_start_xmit()
1013 tbd->data = SWAP32(tx_cmd->dma_addr); in i596_start_xmit()
1017 DMA_WBACK_INV(dev, tbd, sizeof(struct i596_tbd)); in i596_start_xmit()
D82596.c235 struct i596_tbd *tbd; member
1053 struct i596_tbd *tbd; in i596_start_xmit() local
1067 tbd = lp->tbds + lp->next_tx_cmd; in i596_start_xmit()
1078 tx_cmd->tbd = WSWAPtbd(virt_to_bus(tbd)); in i596_start_xmit()
1079 tbd->next = I596_NULL; in i596_start_xmit()
1086 tbd->pad = 0; in i596_start_xmit()
1087 tbd->size = EOF | length; in i596_start_xmit()
1089 tbd->data = WSWAPchar(virt_to_bus(skb->data)); in i596_start_xmit()
Dether1.c676 tbd_t tbd; in ether1_sendpacket() local
707 tbd.tbd_opts = TBD_EOL | skb->len; in ether1_sendpacket()
708 tbd.tbd_link = I82586_NULL; in ether1_sendpacket()
709 tbd.tbd_bufl = dataddr; in ether1_sendpacket()
710 tbd.tbd_bufh = 0; in ether1_sendpacket()
717 ether1_writebuffer (dev, &tbd, tbdaddr, TBD_SIZE); in ether1_sendpacket()
/drivers/atm/
Didt77252.c706 struct scqe *tbd; in push_on_scq() local
741 tbd = &IDT77252_PRV_TBD(skb); in push_on_scq()
744 scq->next->word_1 = cpu_to_le32(tbd->word_1 | in push_on_scq()
746 scq->next->word_2 = cpu_to_le32(tbd->word_2); in push_on_scq()
747 scq->next->word_3 = cpu_to_le32(tbd->word_3); in push_on_scq()
748 scq->next->word_4 = cpu_to_le32(tbd->word_4); in push_on_scq()
833 struct scqe *tbd; in queue_skb() local
846 tbd = &IDT77252_PRV_TBD(skb); in queue_skb()
858 tbd->word_1 = SAR_TBD_OAM | ATM_CELL_PAYLOAD | SAR_TBD_EPDU; in queue_skb()
859 tbd->word_2 = IDT77252_PRV_PADDR(skb) + 4; in queue_skb()
[all …]
Didt77252.h786 struct scqe tbd; /* Transmit Buffer Descriptor */ member
792 (((struct idt77252_skb_prv *)(ATM_SKB(skb)+1))->tbd)
Dnicstar.c135 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd,
1723 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, in push_scqe() argument
1754 *scq->next = *tbd; in push_scqe()
1760 card->index, le32_to_cpu(tbd->word_1), le32_to_cpu(tbd->word_2), in push_scqe()
1761 le32_to_cpu(tbd->word_3), le32_to_cpu(tbd->word_4), in push_scqe()
/drivers/net/wireless/ipw2x00/
Dipw2100.c2821 struct ipw2100_bd *tbd; in __ipw2100_tx_process() local
2834 tbd = &txq->drv[packet->index]; in __ipw2100_tx_process()
2846 descriptors_used = tbd->num_fragments; in __ipw2100_tx_process()
2847 frag_num = tbd->num_fragments - 1; in __ipw2100_tx_process()
2931 tbd = &txq->drv[(packet->index + 1 + i) % txq->entries]; in __ipw2100_tx_process()
2935 tbd->host_addr, tbd->buf_length); in __ipw2100_tx_process()
2938 tbd->host_addr, in __ipw2100_tx_process()
2939 tbd->buf_length, PCI_DMA_TODEVICE); in __ipw2100_tx_process()
3013 struct ipw2100_bd *tbd; in ipw2100_tx_send_commands() local
3041 tbd = &txq->drv[txq->next]; in ipw2100_tx_send_commands()
[all …]
/drivers/net/ethernet/intel/
De100.c520 } tbd; member
1788 cb->u.tcb.tbd_array = cb->dma_addr + offsetof(struct cb, u.tcb.tbd); in e100_xmit_prepare()
1792 cb->u.tcb.tbd.buf_addr = cpu_to_le32(dma_addr); in e100_xmit_prepare()
1793 cb->u.tcb.tbd.size = cpu_to_le16(skb->len); in e100_xmit_prepare()
1857 le32_to_cpu(cb->u.tcb.tbd.buf_addr), in e100_tx_clean()
1858 le16_to_cpu(cb->u.tcb.tbd.size), in e100_tx_clean()
1884 le32_to_cpu(cb->u.tcb.tbd.buf_addr), in e100_clean_cbs()
1885 le16_to_cpu(cb->u.tcb.tbd.size), in e100_clean_cbs()
/drivers/net/ethernet/chelsio/cxgb3/
Dsge.c2861 unsigned int tbd[SGE_TXQ_PER_SET] = {0, 0}; in sge_timer_tx() local
2865 tbd[TXQ_ETH] = reclaim_completed_tx(adap, &qs->txq[TXQ_ETH], in sge_timer_tx()
2871 tbd[TXQ_OFLD] = reclaim_completed_tx(adap, &qs->txq[TXQ_OFLD], in sge_timer_tx()
2877 (max(tbd[TXQ_ETH], tbd[TXQ_OFLD]) / in sge_timer_tx()