Lines Matching refs:txreq
940 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in __get_txreq()
969 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in get_txreq()
993 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) { in qib_put_txreq()
994 tx->txreq.flags &= ~QIB_SDMA_TXREQ_F_FREEBUF; in qib_put_txreq()
996 tx->txreq.addr, tx->hdr_dwords << 2, in qib_put_txreq()
1004 list_add(&tx->txreq.list, &dev->txreq_free); in qib_put_txreq()
1049 if (qp->s_tx->txreq.sg_count > avail) in qib_verbs_sdma_desc_avail()
1051 avail -= qp->s_tx->txreq.sg_count; in qib_verbs_sdma_desc_avail()
1078 container_of(cookie, struct qib_verbs_txreq, txreq); in sdma_complete()
1087 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_FREEBUF) in sdma_complete()
1166 tx->txreq.callback = sdma_complete; in qib_verbs_send_dma()
1168 tx->txreq.flags = QIB_SDMA_TXREQ_F_HEADTOHOST; in qib_verbs_send_dma()
1170 tx->txreq.flags = QIB_SDMA_TXREQ_F_INTREQ; in qib_verbs_send_dma()
1172 tx->txreq.flags |= QIB_SDMA_TXREQ_F_USELARGEBUF; in qib_verbs_send_dma()
1189 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEDESC; in qib_verbs_send_dma()
1190 tx->txreq.sg_count = ndesc; in qib_verbs_send_dma()
1191 tx->txreq.addr = dev->pio_hdrs_phys + in qib_verbs_send_dma()
1208 tx->txreq.addr = dma_map_single(&dd->pcidev->dev, phdr, in qib_verbs_send_dma()
1210 if (dma_mapping_error(&dd->pcidev->dev, tx->txreq.addr)) in qib_verbs_send_dma()
1213 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEBUF; in qib_verbs_send_dma()
1214 tx->txreq.sg_count = 1; in qib_verbs_send_dma()
2129 list_add(&tx->txreq.list, &dev->txreq_free); in qib_register_ib_device()
2254 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in qib_register_ib_device()
2308 tx = list_entry(l, struct qib_verbs_txreq, txreq.list); in qib_unregister_ib_device()