/drivers/ide/ |
D | au1xxx-ide.c | 81 if (!au1xxx_dbdma_put_source(ahwif->tx_chan, virt_to_phys(addr), in auide_outsw() 86 ctp = *((chan_tab_t **)ahwif->tx_chan); in auide_outsw() 252 if (!au1xxx_dbdma_put_source(ahwif->tx_chan, in auide_build_dmatable() 378 auide->tx_chan = au1xxx_dbdma_chan_alloc(auide->target_dev_id, in auide_ddma_init() 389 auide->tx_desc_head = (void*)au1xxx_dbdma_ring_alloc(auide->tx_chan, in auide_ddma_init() 397 au1xxx_dbdma_start( auide->tx_chan ); in auide_ddma_init() 425 auide->tx_chan = au1xxx_dbdma_chan_alloc(DSCR_CMD0_ALWAYS, in auide_ddma_init() 436 auide->tx_desc_head = (void*)au1xxx_dbdma_ring_alloc(auide->tx_chan, in auide_ddma_init() 441 au1xxx_dbdma_start( auide->tx_chan ); in auide_ddma_init()
|
/drivers/net/ethernet/broadcom/ |
D | bcm63xx_enet.c | 498 ENETDMAC_IR, priv->tx_chan); in bcm_enet_poll() 520 ENETDMAC_IRMASK, priv->tx_chan); in bcm_enet_poll() 564 enet_dmac_writel(priv, 0, ENETDMAC_IRMASK, priv->tx_chan); in bcm_enet_isr_dma() 644 ENETDMAC_CHANCFG, priv->tx_chan); in bcm_enet_start_xmit() 923 enet_dmac_writel(priv, 0, ENETDMAC_IRMASK, priv->tx_chan); in bcm_enet_open() 1014 ENETDMAS_RSTART_REG, priv->tx_chan); in bcm_enet_open() 1019 ENETDMAC_RSTART, priv->tx_chan); in bcm_enet_open() 1025 enet_dmas_writel(priv, 0, ENETDMAS_SRAM2_REG, priv->tx_chan); in bcm_enet_open() 1027 enet_dmas_writel(priv, 0, ENETDMAS_SRAM3_REG, priv->tx_chan); in bcm_enet_open() 1029 enet_dmas_writel(priv, 0, ENETDMAS_SRAM4_REG, priv->tx_chan); in bcm_enet_open() [all …]
|
D | bcm63xx_enet.h | 252 int tx_chan; member
|
/drivers/tty/serial/ |
D | samsung.c | 176 if (dma && dma->tx_chan && ourport->tx_in_progress == S3C24XX_TX_DMA) { in s3c24xx_serial_stop_tx() 177 dmaengine_pause(dma->tx_chan); in s3c24xx_serial_stop_tx() 178 dmaengine_tx_status(dma->tx_chan, dma->tx_cookie, &state); in s3c24xx_serial_stop_tx() 179 dmaengine_terminate_all(dma->tx_chan); in s3c24xx_serial_stop_tx() 210 dmaengine_tx_status(dma->tx_chan, dma->tx_cookie, &state); in s3c24xx_serial_tx_dma_complete() 301 dma->tx_desc = dmaengine_prep_slave_single(dma->tx_chan, in s3c24xx_serial_start_tx_dma() 315 dma_async_issue_pending(dma->tx_chan); in s3c24xx_serial_start_tx_dma() 333 if (!ourport->dma || !ourport->dma->tx_chan || in s3c24xx_serial_start_next_tx() 351 if (!ourport->dma || !ourport->dma->tx_chan) in s3c24xx_serial_start_tx() 355 if (ourport->dma && ourport->dma->tx_chan) { in s3c24xx_serial_start_tx() [all …]
|
D | samsung.h | 54 struct dma_chan *tx_chan; member
|
/drivers/atm/ |
D | zatm.c | 328 zpeekl(zatm_dev,zatm_vcc->tx_chan*VC_SIZE/4+i)); 332 qrp = (unsigned long *) zpeekl(zatm_dev,zatm_vcc->tx_chan*VC_SIZE/4+ 700 DPRINTK("QRP=0x%08lx\n",zpeekl(zatm_dev,zatm_vcc->tx_chan*VC_SIZE/4+ in do_tx() 703 zout(uPD98401_TX_READY | (zatm_vcc->tx_chan << in do_tx() 879 chan = zatm_vcc->tx_chan; in close_tx() 904 zatm_vcc->tx_chan = 0; in close_tx() 926 zatm_vcc->tx_chan = 0; in open_tx_first() 954 zatm_vcc->tx_chan = chan; in open_tx_first() 982 if (!zatm_vcc->tx_chan) return 0; in open_tx_second() 985 zpokel(zatm_dev,0,zatm_vcc->tx_chan*VC_SIZE/4); in open_tx_second() [all …]
|
D | zatm.h | 48 int tx_chan; /* TX channel, 0 if none */ member
|
/drivers/usb/renesas_usbhs/ |
D | fifo.c | 775 return fifo->tx_chan; in usbhsf_dma_chan_get() 1254 if (fifo->tx_chan) in usbhsf_dma_quit() 1255 dma_release_channel(fifo->tx_chan); in usbhsf_dma_quit() 1259 fifo->tx_chan = NULL; in usbhsf_dma_quit() 1269 fifo->tx_chan = dma_request_channel(mask, usbhsf_dma_filter, in usbhsf_dma_init_pdev() 1290 fifo->tx_chan = dma_request_slave_channel_reason(dev, name); in usbhsf_dma_init_dt() 1291 if (IS_ERR(fifo->tx_chan)) in usbhsf_dma_init_dt() 1292 fifo->tx_chan = NULL; in usbhsf_dma_init_dt() 1310 if (fifo->tx_chan || fifo->rx_chan) in usbhsf_dma_init() 1313 fifo->tx_chan ? "[TX]" : " ", in usbhsf_dma_init()
|
D | fifo.h | 34 struct dma_chan *tx_chan; member
|
/drivers/mmc/host/ |
D | au1xmmc.c | 112 u32 tx_chan; member 152 (((h)->flags & HOST_F_XMIT) ? (h)->tx_chan : (h)->rx_chan) 892 host->tx_chan = au1xxx_dbdma_chan_alloc(memid, txid, in au1xmmc_dbdma_init() 894 if (!host->tx_chan) { in au1xmmc_dbdma_init() 903 au1xxx_dbdma_chan_free(host->tx_chan); in au1xmmc_dbdma_init() 907 au1xxx_dbdma_set_devwidth(host->tx_chan, 8); in au1xmmc_dbdma_init() 910 au1xxx_dbdma_ring_alloc(host->tx_chan, AU1XMMC_DESCRIPTOR_COUNT); in au1xmmc_dbdma_init() 923 au1xxx_dbdma_chan_free(host->tx_chan); in au1xmmc_dbdma_shutdown()
|
D | omap_hsmmc.c | 201 struct dma_chan *tx_chan; member 920 return data->flags & MMC_DATA_WRITE ? host->tx_chan : host->rx_chan; in omap_hsmmc_get_dma_chan() 2138 host->tx_chan = dma_request_chan(&pdev->dev, "tx"); in omap_hsmmc_probe() 2139 if (IS_ERR(host->tx_chan)) { in omap_hsmmc_probe() 2141 ret = PTR_ERR(host->tx_chan); in omap_hsmmc_probe() 2154 dma_get_max_seg_size(host->tx_chan->device->dev)); in omap_hsmmc_probe() 2211 if (!IS_ERR_OR_NULL(host->tx_chan)) in omap_hsmmc_probe() 2212 dma_release_channel(host->tx_chan); in omap_hsmmc_probe() 2234 dma_release_channel(host->tx_chan); in omap_hsmmc_remove()
|
/drivers/spi/ |
D | spi-sirf.c | 275 struct dma_chan *tx_chan; member 560 tx_desc = dmaengine_prep_slave_single(sspi->tx_chan, in spi_sirfsoc_dma_transfer() 568 dma_async_issue_pending(sspi->tx_chan); in spi_sirfsoc_dma_transfer() 594 dmaengine_terminate_all(sspi->tx_chan); in spi_sirfsoc_dma_transfer() 1138 sspi->tx_chan = dma_request_slave_channel(&pdev->dev, "tx"); in spi_sirfsoc_probe() 1139 if (!sspi->tx_chan) { in spi_sirfsoc_probe() 1168 dma_release_channel(sspi->tx_chan); in spi_sirfsoc_probe() 1188 dma_release_channel(sspi->tx_chan); in spi_sirfsoc_remove()
|
/drivers/mtd/nand/ |
D | qcom_nandc.c | 351 struct dma_chan *tx_chan; member 694 } else if (chan == nandc->tx_chan) { in prepare_bam_async_desc() 761 ret = prepare_bam_async_desc(nandc, nandc->tx_chan, in prep_bam_dma_desc_data() 1166 r = prepare_bam_async_desc(nandc, nandc->tx_chan, in submit_descs() 1183 dma_async_issue_pending(nandc->tx_chan); in submit_descs() 2434 nandc->tx_chan = dma_request_slave_channel(nandc->dev, "tx"); in qcom_nandc_alloc() 2435 if (!nandc->tx_chan) { in qcom_nandc_alloc() 2491 if (nandc->tx_chan) in qcom_nandc_unalloc() 2492 dma_release_channel(nandc->tx_chan); in qcom_nandc_unalloc()
|
/drivers/net/ethernet/chelsio/cxgb4/ |
D | cxgb4_ethtool.c | 420 t4_get_port_stats_offset(adapter, pi->tx_chan, in get_stats() 463 t4_restart_aneg(p->adapter, p->adapter->pf, p->tx_chan); in restart_autoneg() 821 ret = t4_link_l1cfg(pi->adapter, pi->adapter->mbox, pi->tx_chan, lc); in set_link_ksettings() 920 pi->tx_chan, lc); in set_fecparam() 954 return t4_link_l1cfg(p->adapter, p->adapter->mbox, p->tx_chan, in set_pauseparam()
|
D | cxgb4_main.c | 449 ret = t4_link_l1cfg(pi->adapter, mb, pi->tx_chan, in link_start() 742 return netdev2pinfo(p->ingr_map[qid]->netdev)->tx_chan; in rxq_to_chan() 853 pi->tx_chan)); in setup_sge_queues() 892 RSSCONTROL_V(netdev2pinfo(adap->port[0])->tx_chan) | in setup_sge_queues() 1586 return netdev2pinfo(dev)->tx_chan; in cxgb4_port_chan() 2416 t4_get_port_stats_offset(adapter, p->tx_chan, &stats, in cxgb_get_stats() 2726 pi->tx_chan, class_id, 0, in cxgb_set_vf_rate() 2851 p.u.params.channel = pi->tx_chan; in cxgb_set_tx_maxrate() 4142 ret = t4_alloc_vi(adap, adap->mbox, p->tx_chan, adap->pf, 0, 1, in eeh_slot_reset() 4698 pi->tx_chan = adap->pf % adap->params.nports; in config_mgmt_dev() [all …]
|
D | cxgb4_ptp.c | 186 c.u.init.txchan = pi->tx_chan; in cxgb4_ptp_redirect_rx_packet()
|
D | sge.c | 1328 ctrl0 = TXPKT_OPCODE_V(CPL_TX_PKT_XT) | TXPKT_INTF_V(pi->tx_chan) | in t4_eth_xmit() 2745 c.iqdroprss_to_iqesize = htons(FW_IQ_CMD_IQPCIECH_V(pi->tx_chan) | in t4_sge_alloc_rxq() 2948 FW_EQ_ETH_CMD_PCIECHN_V(pi->tx_chan) | in t4_sge_alloc_eth_txq() 3004 FW_EQ_CTRL_CMD_PCIECHN_V(pi->tx_chan) | in t4_sge_alloc_ctrl_txq() 3074 FW_EQ_OFLD_CMD_PCIECHN_V(pi->tx_chan) | in t4_sge_alloc_uld_txq()
|
D | l2t.c | 431 lport = netdev2pinfo(physdev)->tx_chan + 4; in cxgb4_l2t_get()
|
/drivers/infiniband/hw/cxgb4/ |
D | cm.c | 611 flowc->mnemval[1].val = cpu_to_be32(ep->tx_chan); in send_flowc() 613 flowc->mnemval[2].val = cpu_to_be32(ep->tx_chan); in send_flowc() 756 TX_CHAN_V(ep->tx_chan) | in send_connect() 1944 TX_CHAN_V(ep->tx_chan) | in send_fw_act_open_req() 1950 TX_QUEUE_V(ep->com.dev->rdev.lldi.tx_modq[ep->tx_chan]) | in send_fw_act_open_req() 2043 ep->tx_chan = cxgb4_port_chan(pdev); in import_ep() 2063 ep->tx_chan = cxgb4_port_chan(pdev); in import_ep() 2170 __func__, ep->txq_idx, ep->tx_chan, ep->smac_idx, ep->rss_qid, in c4iw_reconnect() 2396 TX_CHAN_V(ep->tx_chan) | in accept_cr() 2583 child_ep->tx_chan, child_ep->smac_idx, child_ep->rss_qid); in pass_accept_req() [all …]
|
D | iw_cxgb4.h | 880 u32 tx_chan; member
|
/drivers/scsi/cxgbi/cxgb4i/ |
D | cxgb4i.c | 203 TX_CHAN_V(csk->tx_chan) | in send_act_open_req() 322 TX_CHAN_V(csk->tx_chan) | in send_act_open_req6() 608 flowc->mnemval[1].val = htonl(csk->tx_chan); in send_tx_flowc_wr() 610 flowc->mnemval[2].val = htonl(csk->tx_chan); in send_tx_flowc_wr() 630 csk, csk->tid, 0, csk->tx_chan, csk->rss_qid, in send_tx_flowc_wr() 1688 csk->tx_chan = cxgb4_port_chan(ndev); in init_act_open() 1721 csk->flags, csk->tx_chan, csk->txq_idx, csk->rss_qid, in init_act_open()
|
/drivers/target/iscsi/cxgbit/ |
D | cxgbit_cm.c | 878 csk->tx_chan = cxgb4_port_chan(ndev); in cxgbit_offload_init() 914 csk->tx_chan = cxgb4_port_chan(ndev); in cxgbit_offload_init() 1111 TX_CHAN_V(csk->tx_chan) | in cxgbit_pass_accept_rpl() 1382 flowc->mnemval[1].val = cpu_to_be32(csk->tx_chan); in cxgbit_send_tx_flowc_wr() 1384 flowc->mnemval[2].val = cpu_to_be32(csk->tx_chan); in cxgbit_send_tx_flowc_wr() 1422 __func__, csk, csk->tx_chan, csk->rss_qid, csk->snd_nxt, in cxgbit_send_tx_flowc_wr()
|
D | cxgbit.h | 224 u32 tx_chan; member
|
/drivers/ntb/ |
D | ntb_transport.c | 2167 struct dma_chan *rx_chan, *tx_chan; in ntb_transport_max_size() local 2173 tx_chan = qp->tx_dma_chan; in ntb_transport_max_size() 2176 tx_chan ? tx_chan->device->copy_align : 0); in ntb_transport_max_size()
|
/drivers/scsi/cxgbi/ |
D | libcxgbi.h | 115 unsigned int tx_chan; member
|