/drivers/net/ethernet/qlogic/qlge/ |
D | qlge_dbg.c | 1854 void ql_dump_tx_desc(struct tx_buf_desc *tbd) in ql_dump_tx_desc() argument 1857 le64_to_cpu((u64) tbd->addr)); in ql_dump_tx_desc() 1859 le32_to_cpu(tbd->len & TX_DESC_LEN_MASK)); in ql_dump_tx_desc() 1861 tbd->len & TX_DESC_C ? "C" : ".", in ql_dump_tx_desc() 1862 tbd->len & TX_DESC_E ? "E" : "."); in ql_dump_tx_desc() 1863 tbd++; in ql_dump_tx_desc() 1865 le64_to_cpu((u64) tbd->addr)); in ql_dump_tx_desc() 1867 le32_to_cpu(tbd->len & TX_DESC_LEN_MASK)); in ql_dump_tx_desc() 1869 tbd->len & TX_DESC_C ? "C" : ".", in ql_dump_tx_desc() 1870 tbd->len & TX_DESC_E ? "E" : "."); in ql_dump_tx_desc() [all …]
|
D | qlge_main.c | 1317 struct tx_buf_desc *tbd = mac_iocb_ptr->tbd; in ql_map_send() local 1337 tbd->len = cpu_to_le32(len); in ql_map_send() 1338 tbd->addr = cpu_to_le64(map); in ql_map_send() 1352 tbd++; in ql_map_send() 1384 tbd->addr = cpu_to_le64(map); in ql_map_send() 1390 tbd->len = in ql_map_send() 1397 tbd = (struct tx_buf_desc *)&tx_ring_desc->oal; in ql_map_send() 1412 tbd->addr = cpu_to_le64(map); in ql_map_send() 1413 tbd->len = cpu_to_le32(skb_frag_size(frag)); in ql_map_send() 1422 tbd->len = cpu_to_le32(le32_to_cpu(tbd->len) | TX_DESC_E); in ql_map_send()
|
D | qlge.h | 1097 struct tx_buf_desc tbd[TX_DESC_PER_IOCB]; member 1142 struct tx_buf_desc tbd[TX_DESC_PER_IOCB]; member 2314 void ql_dump_tx_desc(struct tx_buf_desc *tbd);
|
/drivers/net/ethernet/i825xx/ |
D | lib82596.c | 202 u32 tbd; member 972 struct i596_tbd *tbd; in i596_start_xmit() local 988 tbd = lp->dma->tbds + lp->next_tx_cmd; in i596_start_xmit() 1000 tx_cmd->tbd = SWAP32(virt_to_dma(lp, tbd)); in i596_start_xmit() 1001 tbd->next = I596_NULL; in i596_start_xmit() 1008 tbd->pad = 0; in i596_start_xmit() 1009 tbd->size = SWAP16(EOF | length); in i596_start_xmit() 1013 tbd->data = SWAP32(tx_cmd->dma_addr); in i596_start_xmit() 1017 DMA_WBACK_INV(dev, tbd, sizeof(struct i596_tbd)); in i596_start_xmit()
|
D | 82596.c | 235 struct i596_tbd *tbd; member 1053 struct i596_tbd *tbd; in i596_start_xmit() local 1067 tbd = lp->tbds + lp->next_tx_cmd; in i596_start_xmit() 1078 tx_cmd->tbd = WSWAPtbd(virt_to_bus(tbd)); in i596_start_xmit() 1079 tbd->next = I596_NULL; in i596_start_xmit() 1086 tbd->pad = 0; in i596_start_xmit() 1087 tbd->size = EOF | length; in i596_start_xmit() 1089 tbd->data = WSWAPchar(virt_to_bus(skb->data)); in i596_start_xmit()
|
D | ether1.c | 676 tbd_t tbd; in ether1_sendpacket() local 707 tbd.tbd_opts = TBD_EOL | skb->len; in ether1_sendpacket() 708 tbd.tbd_link = I82586_NULL; in ether1_sendpacket() 709 tbd.tbd_bufl = dataddr; in ether1_sendpacket() 710 tbd.tbd_bufh = 0; in ether1_sendpacket() 717 ether1_writebuffer (dev, &tbd, tbdaddr, TBD_SIZE); in ether1_sendpacket()
|
/drivers/atm/ |
D | idt77252.c | 706 struct scqe *tbd; in push_on_scq() local 741 tbd = &IDT77252_PRV_TBD(skb); in push_on_scq() 744 scq->next->word_1 = cpu_to_le32(tbd->word_1 | in push_on_scq() 746 scq->next->word_2 = cpu_to_le32(tbd->word_2); in push_on_scq() 747 scq->next->word_3 = cpu_to_le32(tbd->word_3); in push_on_scq() 748 scq->next->word_4 = cpu_to_le32(tbd->word_4); in push_on_scq() 833 struct scqe *tbd; in queue_skb() local 846 tbd = &IDT77252_PRV_TBD(skb); in queue_skb() 858 tbd->word_1 = SAR_TBD_OAM | ATM_CELL_PAYLOAD | SAR_TBD_EPDU; in queue_skb() 859 tbd->word_2 = IDT77252_PRV_PADDR(skb) + 4; in queue_skb() [all …]
|
D | idt77252.h | 786 struct scqe tbd; /* Transmit Buffer Descriptor */ member 792 (((struct idt77252_skb_prv *)(ATM_SKB(skb)+1))->tbd)
|
D | nicstar.c | 135 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, 1723 static int push_scqe(ns_dev * card, vc_map * vc, scq_info * scq, ns_scqe * tbd, in push_scqe() argument 1754 *scq->next = *tbd; in push_scqe() 1760 card->index, le32_to_cpu(tbd->word_1), le32_to_cpu(tbd->word_2), in push_scqe() 1761 le32_to_cpu(tbd->word_3), le32_to_cpu(tbd->word_4), in push_scqe()
|
/drivers/net/wireless/ipw2x00/ |
D | ipw2100.c | 2821 struct ipw2100_bd *tbd; in __ipw2100_tx_process() local 2834 tbd = &txq->drv[packet->index]; in __ipw2100_tx_process() 2846 descriptors_used = tbd->num_fragments; in __ipw2100_tx_process() 2847 frag_num = tbd->num_fragments - 1; in __ipw2100_tx_process() 2931 tbd = &txq->drv[(packet->index + 1 + i) % txq->entries]; in __ipw2100_tx_process() 2935 tbd->host_addr, tbd->buf_length); in __ipw2100_tx_process() 2938 tbd->host_addr, in __ipw2100_tx_process() 2939 tbd->buf_length, PCI_DMA_TODEVICE); in __ipw2100_tx_process() 3013 struct ipw2100_bd *tbd; in ipw2100_tx_send_commands() local 3041 tbd = &txq->drv[txq->next]; in ipw2100_tx_send_commands() [all …]
|
/drivers/net/ethernet/intel/ |
D | e100.c | 520 } tbd; member 1788 cb->u.tcb.tbd_array = cb->dma_addr + offsetof(struct cb, u.tcb.tbd); in e100_xmit_prepare() 1792 cb->u.tcb.tbd.buf_addr = cpu_to_le32(dma_addr); in e100_xmit_prepare() 1793 cb->u.tcb.tbd.size = cpu_to_le16(skb->len); in e100_xmit_prepare() 1857 le32_to_cpu(cb->u.tcb.tbd.buf_addr), in e100_tx_clean() 1858 le16_to_cpu(cb->u.tcb.tbd.size), in e100_tx_clean() 1884 le32_to_cpu(cb->u.tcb.tbd.buf_addr), in e100_clean_cbs() 1885 le16_to_cpu(cb->u.tcb.tbd.size), in e100_clean_cbs()
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | sge.c | 2861 unsigned int tbd[SGE_TXQ_PER_SET] = {0, 0}; in sge_timer_tx() local 2865 tbd[TXQ_ETH] = reclaim_completed_tx(adap, &qs->txq[TXQ_ETH], in sge_timer_tx() 2871 tbd[TXQ_OFLD] = reclaim_completed_tx(adap, &qs->txq[TXQ_OFLD], in sge_timer_tx() 2877 (max(tbd[TXQ_ETH], tbd[TXQ_OFLD]) / in sge_timer_tx()
|