Home
last modified time | relevance | path

Searched refs:rx_chan (Results 1 – 20 of 20) sorted by relevance

/drivers/mtd/hyperbus/
Dhbmc-am654.c29 struct dma_chan *rx_chan; member
78 struct dma_chan *rx_chan = priv->rx_chan; in am654_hbmc_dma_read() local
84 if (!priv->rx_chan || !virt_addr_valid(to) || object_is_on_stack(to)) in am654_hbmc_dma_read()
87 dma_dst = dma_map_single(rx_chan->device->dev, to, len, DMA_FROM_DEVICE); in am654_hbmc_dma_read()
88 if (dma_mapping_error(rx_chan->device->dev, dma_dst)) { in am654_hbmc_dma_read()
94 tx = dmaengine_prep_dma_memcpy(rx_chan, dma_dst, dma_src, len, flags); in am654_hbmc_dma_read()
112 dma_async_issue_pending(rx_chan); in am654_hbmc_dma_read()
114 dmaengine_terminate_sync(rx_chan); in am654_hbmc_dma_read()
120 dma_unmap_single(rx_chan->device->dev, dma_dst, len, DMA_FROM_DEVICE); in am654_hbmc_dma_read()
140 struct dma_chan *rx_chan; in am654_hbmc_request_mmap_dma() local
[all …]
/drivers/soc/xilinx/
Dzynqmp_power.c32 static struct mbox_chan *rx_chan; variable
104 ret = mbox_send_message(rx_chan, NULL); in ipi_receive_callback()
205 rx_chan = mbox_request_channel_byname(client, "rx"); in zynqmp_pm_probe()
206 if (IS_ERR(rx_chan)) { in zynqmp_pm_probe()
208 return PTR_ERR(rx_chan); in zynqmp_pm_probe()
243 if (!rx_chan) in zynqmp_pm_remove()
244 mbox_free_channel(rx_chan); in zynqmp_pm_remove()
/drivers/net/ethernet/broadcom/
Dbcm63xx_enet.c265 enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); in bcm_enet_refill_rx()
267 enet_dmac_writel(priv, 1, ENETDMAC_BUFALLOC, priv->rx_chan); in bcm_enet_refill_rx()
403 ENETDMAC_CHANCFG, priv->rx_chan); in bcm_enet_receive_queue()
479 ENETDMAC_IR, priv->rx_chan); in bcm_enet_poll()
501 ENETDMAC_IRMASK, priv->rx_chan); in bcm_enet_poll()
546 enet_dmac_writel(priv, 0, ENETDMAC_IRMASK, priv->rx_chan); in bcm_enet_isr_dma()
761 val |= ENETDMA_CFG_FLOWCH_MASK(priv->rx_chan); in bcm_enet_set_flow()
763 val &= ~ENETDMA_CFG_FLOWCH_MASK(priv->rx_chan); in bcm_enet_set_flow()
896 enet_dmac_writel(priv, 0, ENETDMAC_IRMASK, priv->rx_chan); in bcm_enet_open()
972 ENETDMA_BUFALLOC_REG(priv->rx_chan)); in bcm_enet_open()
[all …]
Dbcm63xx_enet.h216 int rx_chan; member
/drivers/spi/
Dspi-ti-qspi.c58 struct dma_chan *rx_chan; member
446 struct dma_chan *chan = qspi->rx_chan; in ti_qspi_dma_xfer()
632 if (qspi->rx_chan) { in ti_qspi_exec_mem_op()
744 if (qspi->rx_chan) in ti_qspi_dma_cleanup()
745 dma_release_channel(qspi->rx_chan); in ti_qspi_dma_cleanup()
860 qspi->rx_chan = dma_request_chan_by_mask(&mask); in ti_qspi_probe()
861 if (IS_ERR(qspi->rx_chan)) { in ti_qspi_probe()
864 qspi->rx_chan = NULL; in ti_qspi_probe()
875 dma_release_channel(qspi->rx_chan); in ti_qspi_probe()
878 master->dma_rx = qspi->rx_chan; in ti_qspi_probe()
[all …]
Dspi-cadence-quadspi.c69 struct dma_chan *rx_chan; member
912 if (!cqspi->rx_chan || !virt_addr_valid(buf)) { in cqspi_direct_read_execute()
917 ddev = cqspi->rx_chan->device->dev; in cqspi_direct_read_execute()
923 tx = dmaengine_prep_dma_memcpy(cqspi->rx_chan, dma_dst, dma_src, in cqspi_direct_read_execute()
943 dma_async_issue_pending(cqspi->rx_chan); in cqspi_direct_read_execute()
946 dmaengine_terminate_sync(cqspi->rx_chan); in cqspi_direct_read_execute()
1119 cqspi->rx_chan = dma_request_chan_by_mask(&mask); in cqspi_request_mmap_dma()
1120 if (IS_ERR(cqspi->rx_chan)) { in cqspi_request_mmap_dma()
1121 int ret = PTR_ERR(cqspi->rx_chan); in cqspi_request_mmap_dma()
1122 cqspi->rx_chan = NULL; in cqspi_request_mmap_dma()
[all …]
Dspi-sirf.c273 struct dma_chan *rx_chan; member
550 rx_desc = dmaengine_prep_slave_single(sspi->rx_chan, in spi_sirfsoc_dma_transfer()
568 dma_async_issue_pending(sspi->rx_chan); in spi_sirfsoc_dma_transfer()
580 dmaengine_terminate_all(sspi->rx_chan); in spi_sirfsoc_dma_transfer()
1129 sspi->rx_chan = dma_request_chan(&pdev->dev, "rx"); in spi_sirfsoc_probe()
1130 if (IS_ERR(sspi->rx_chan)) { in spi_sirfsoc_probe()
1132 ret = PTR_ERR(sspi->rx_chan); in spi_sirfsoc_probe()
1167 dma_release_channel(sspi->rx_chan); in spi_sirfsoc_probe()
1184 dma_release_channel(sspi->rx_chan); in spi_sirfsoc_remove()
/drivers/tty/serial/
Dsamsung_tty.c92 struct dma_chan *rx_chan; member
522 if (dma && dma->rx_chan) { in s3c24xx_serial_stop_rx()
524 dma_status = dmaengine_tx_status(dma->rx_chan, in s3c24xx_serial_stop_rx()
529 dmaengine_terminate_all(dma->rx_chan); in s3c24xx_serial_stop_rx()
578 dmaengine_tx_status(dma->rx_chan, dma->rx_cookie, &state); in s3c24xx_serial_rx_dma_complete()
604 dma->rx_desc = dmaengine_prep_slave_single(dma->rx_chan, in s3c64xx_start_rx_dma()
617 dma_async_issue_pending(dma->rx_chan); in s3c64xx_start_rx_dma()
692 dmaengine_pause(dma->rx_chan); in s3c24xx_serial_rx_chars_dma()
693 dmaengine_tx_status(dma->rx_chan, dma->rx_cookie, &state); in s3c24xx_serial_rx_chars_dma()
694 dmaengine_terminate_all(dma->rx_chan); in s3c24xx_serial_rx_chars_dma()
[all …]
/drivers/usb/renesas_usbhs/
Dfifo.c772 return fifo->rx_chan; in usbhsf_dma_chan_get()
1250 if (fifo->rx_chan) in usbhsf_dma_quit()
1251 dma_release_channel(fifo->rx_chan); in usbhsf_dma_quit()
1254 fifo->rx_chan = NULL; in usbhsf_dma_quit()
1268 fifo->rx_chan = dma_request_channel(mask, usbhsf_dma_filter, in usbhsf_dma_init_pdev()
1288 fifo->rx_chan = dma_request_chan(dev, name); in usbhsf_dma_init_dt()
1289 if (IS_ERR(fifo->rx_chan)) in usbhsf_dma_init_dt()
1290 fifo->rx_chan = NULL; in usbhsf_dma_init_dt()
1304 if (fifo->tx_chan || fifo->rx_chan) in usbhsf_dma_init()
1308 fifo->rx_chan ? "[RX]" : " "); in usbhsf_dma_init()
Dfifo.h26 struct dma_chan *rx_chan; member
/drivers/mmc/host/
Dau1xmmc.c112 u32 rx_chan; member
151 (((h)->flags & HOST_F_XMIT) ? (h)->tx_chan : (h)->rx_chan)
885 host->rx_chan = au1xxx_dbdma_chan_alloc(rxid, memid, in au1xmmc_dbdma_init()
887 if (!host->rx_chan) { in au1xmmc_dbdma_init()
894 au1xxx_dbdma_set_devwidth(host->rx_chan, 8); in au1xmmc_dbdma_init()
897 au1xxx_dbdma_ring_alloc(host->rx_chan, AU1XMMC_DESCRIPTOR_COUNT); in au1xmmc_dbdma_init()
910 au1xxx_dbdma_chan_free(host->rx_chan); in au1xmmc_dbdma_shutdown()
Domap_hsmmc.c196 struct dma_chan *rx_chan; member
818 return data->flags & MMC_DATA_WRITE ? host->tx_chan : host->rx_chan; in omap_hsmmc_get_dma_chan()
1938 host->rx_chan = dma_request_chan(&pdev->dev, "rx"); in omap_hsmmc_probe()
1939 if (IS_ERR(host->rx_chan)) { in omap_hsmmc_probe()
1941 ret = PTR_ERR(host->rx_chan); in omap_hsmmc_probe()
1960 dma_get_max_seg_size(host->rx_chan->device->dev), in omap_hsmmc_probe()
2014 if (!IS_ERR_OR_NULL(host->rx_chan)) in omap_hsmmc_probe()
2015 dma_release_channel(host->rx_chan); in omap_hsmmc_probe()
2034 dma_release_channel(host->rx_chan); in omap_hsmmc_remove()
Dsdhci.c1191 host->rx_chan = dma_request_chan(mmc->parent, "rx"); in sdhci_external_dma_init()
1192 if (IS_ERR(host->rx_chan)) { in sdhci_external_dma_init()
1198 ret = PTR_ERR(host->rx_chan); in sdhci_external_dma_init()
1201 host->rx_chan = NULL; in sdhci_external_dma_init()
1210 return data->flags & MMC_DATA_WRITE ? host->tx_chan : host->rx_chan; in sdhci_external_dma_channel()
1275 if (host->rx_chan) { in sdhci_external_dma_release()
1276 dma_release_channel(host->rx_chan); in sdhci_external_dma_release()
1277 host->rx_chan = NULL; in sdhci_external_dma_release()
Dsdhci.h564 struct dma_chan *rx_chan; member
/drivers/atm/
Dzatm.c495 zatm_vcc->rx_chan = 0; in open_rx_first()
529 zatm_vcc->rx_chan = chan; in open_rx_first()
546 if (!zatm_vcc->rx_chan) return 0; in open_rx_second()
552 ((zatm_vcc->rx_chan | uPD98401_RXLT_ENBL) << shift),pos); in open_rx_second()
567 if (!zatm_vcc->rx_chan) return; in close_rx()
583 zout(uPD98401_DEACT_CHAN | uPD98401_CHAN_RT | (zatm_vcc->rx_chan << in close_rx()
587 zout(uPD98401_CLOSE_CHAN | uPD98401_CHAN_RT | (zatm_vcc->rx_chan << in close_rx()
592 "%d\n",vcc->dev->number,zatm_vcc->rx_chan); in close_rx()
594 zatm_dev->rx_map[zatm_vcc->rx_chan] = NULL; in close_rx()
595 zatm_vcc->rx_chan = 0; in close_rx()
Dzatm.h45 int rx_chan; /* RX channel, 0 if none */ member
/drivers/mtd/nand/raw/
Dqcom_nandc.c375 struct dma_chan *rx_chan; member
1293 r = prepare_bam_async_desc(nandc, nandc->rx_chan, 0); in submit_descs()
1326 dma_async_issue_pending(nandc->rx_chan); in submit_descs()
2646 if (nandc->rx_chan) in qcom_nandc_unalloc()
2647 dma_release_channel(nandc->rx_chan); in qcom_nandc_unalloc()
2711 nandc->rx_chan = dma_request_chan(nandc->dev, "rx"); in qcom_nandc_alloc()
2712 if (IS_ERR(nandc->rx_chan)) { in qcom_nandc_alloc()
2713 ret = PTR_ERR(nandc->rx_chan); in qcom_nandc_alloc()
2714 nandc->rx_chan = NULL; in qcom_nandc_alloc()
/drivers/ntb/
Dntb_transport.c2410 struct dma_chan *rx_chan, *tx_chan; in ntb_transport_max_size() local
2415 rx_chan = qp->rx_dma_chan; in ntb_transport_max_size()
2418 copy_align = max(rx_chan ? rx_chan->device->copy_align : 0, in ntb_transport_max_size()
/drivers/scsi/cxgbi/
Dlibcxgbi.h124 unsigned int rx_chan; member
/drivers/net/ethernet/chelsio/inline_crypto/chtls/
Dchtls.h302 u32 rx_chan; member