/drivers/spi/ |
D | spi-pxa2xx-dma.c | 102 chan = drv_data->controller->dma_rx; in pxa2xx_spi_dma_prepare_one() 123 dmaengine_terminate_async(drv_data->controller->dma_rx); in pxa2xx_spi_dma_transfer() 169 dma_async_issue_pending(drv_data->controller->dma_rx); in pxa2xx_spi_dma_start() 178 dmaengine_terminate_sync(drv_data->controller->dma_rx); in pxa2xx_spi_dma_stop() 197 controller->dma_rx = dma_request_slave_channel_compat(mask, in pxa2xx_spi_dma_setup() 199 if (!controller->dma_rx) { in pxa2xx_spi_dma_setup() 212 if (controller->dma_rx) { in pxa2xx_spi_dma_release() 213 dmaengine_terminate_sync(controller->dma_rx); in pxa2xx_spi_dma_release() 214 dma_release_channel(controller->dma_rx); in pxa2xx_spi_dma_release() 215 controller->dma_rx = NULL; in pxa2xx_spi_dma_release()
|
D | spi-bcm2835.c | 522 dma_sync_single_for_device(ctlr->dma_rx->device->dev, in bcm2835_spi_transfer_prologue() 645 dmaengine_terminate_async(ctlr->dma_rx); in bcm2835_spi_dma_tx_done() 686 chan = ctlr->dma_rx; in bcm2835_spi_prepare_sg() 821 dma_async_issue_pending(ctlr->dma_rx); in bcm2835_spi_transfer_one_dma() 831 dmaengine_terminate_async(ctlr->dma_rx); in bcm2835_spi_transfer_one_dma() 875 if (ctlr->dma_rx) { in bcm2835_dma_release() 876 dmaengine_terminate_sync(ctlr->dma_rx); in bcm2835_dma_release() 877 dma_release_channel(ctlr->dma_rx); in bcm2835_dma_release() 878 ctlr->dma_rx = NULL; in bcm2835_dma_release() 907 ctlr->dma_rx = dma_request_chan(dev, "rx"); in bcm2835_dma_init() [all …]
|
D | spi-uniphier.c | 356 if ((!master->dma_tx && !master->dma_rx) in uniphier_spi_can_dma() 358 || (!master->dma_rx && t->rx_buf)) in uniphier_spi_can_dma() 415 dmaengine_slave_config(master->dma_rx, &rxconf); in uniphier_spi_transfer_one_dma() 418 master->dma_rx, in uniphier_spi_transfer_one_dma() 431 dma_async_issue_pending(master->dma_rx); in uniphier_spi_transfer_one_dma() 466 dmaengine_terminate_sync(master->dma_rx); in uniphier_spi_transfer_one_dma() 595 dmaengine_terminate_async(master->dma_rx); in uniphier_spi_handle_err() 734 master->dma_rx = dma_request_chan(&pdev->dev, "rx"); in uniphier_spi_probe() 735 if (IS_ERR_OR_NULL(master->dma_rx)) { in uniphier_spi_probe() 736 if (PTR_ERR(master->dma_rx) == -EPROBE_DEFER) { in uniphier_spi_probe() [all …]
|
D | spi-sun6i.c | 215 dmaengine_slave_config(master->dma_rx, &rxconf); in sun6i_spi_prepare_dma() 217 rxdesc = dmaengine_prep_slave_sg(master->dma_rx, in sun6i_spi_prepare_dma() 246 dmaengine_terminate_sync(master->dma_rx); in sun6i_spi_prepare_dma() 253 dma_async_issue_pending(master->dma_rx); in sun6i_spi_prepare_dma() 468 dmaengine_terminate_sync(master->dma_rx); in sun6i_spi_transfer_one() 660 master->dma_rx = dma_request_chan(&pdev->dev, "rx"); in sun6i_spi_probe() 661 if (IS_ERR(master->dma_rx)) { in sun6i_spi_probe() 662 if (PTR_ERR(master->dma_rx) == -EPROBE_DEFER) { in sun6i_spi_probe() 667 master->dma_rx = NULL; in sun6i_spi_probe() 670 if (master->dma_tx && master->dma_rx) { in sun6i_spi_probe() [all …]
|
D | spi-at91-usart.c | 148 ctlr->dma_rx = dma_request_chan(dev, "rx"); in at91_usart_spi_configure_dma() 149 if (IS_ERR_OR_NULL(ctlr->dma_rx)) { in at91_usart_spi_configure_dma() 150 if (IS_ERR(ctlr->dma_rx)) { in at91_usart_spi_configure_dma() 151 err = PTR_ERR(ctlr->dma_rx); in at91_usart_spi_configure_dma() 170 if (dmaengine_slave_config(ctlr->dma_rx, &slave_config)) { in at91_usart_spi_configure_dma() 191 if (!IS_ERR_OR_NULL(ctlr->dma_rx)) in at91_usart_spi_configure_dma() 192 dma_release_channel(ctlr->dma_rx); in at91_usart_spi_configure_dma() 194 ctlr->dma_rx = NULL; in at91_usart_spi_configure_dma() 202 if (ctlr->dma_rx) in at91_usart_spi_release_dma() 203 dma_release_channel(ctlr->dma_rx); in at91_usart_spi_release_dma() [all …]
|
D | spi-rockchip.c | 290 dmaengine_terminate_async(ctlr->dma_rx); in rockchip_spi_handle_err() 460 dmaengine_slave_config(ctlr->dma_rx, &rxconf); in rockchip_spi_prepare_dma() 463 ctlr->dma_rx, in rockchip_spi_prepare_dma() 490 dmaengine_terminate_sync(ctlr->dma_rx); in rockchip_spi_prepare_dma() 501 ctlr->dma_rx->cookie = dmaengine_submit(rxdesc); in rockchip_spi_prepare_dma() 502 dma_async_issue_pending(ctlr->dma_rx); in rockchip_spi_prepare_dma() 620 dmaengine_pause(ctlr->dma_rx); in rockchip_spi_slave_abort() 621 status = dmaengine_tx_status(ctlr->dma_rx, ctlr->dma_rx->cookie, &state); in rockchip_spi_slave_abort() 651 dmaengine_terminate_sync(ctlr->dma_rx); in rockchip_spi_slave_abort() 890 ctlr->dma_rx = dma_request_chan(rs->dev, "rx"); in rockchip_spi_probe() [all …]
|
D | spi-pic32.c | 309 if (!master->dma_rx || !master->dma_tx) in pic32_spi_dma_transfer() 312 desc_rx = dmaengine_prep_slave_sg(master->dma_rx, in pic32_spi_dma_transfer() 346 dma_async_issue_pending(master->dma_rx); in pic32_spi_dma_transfer() 352 dmaengine_terminate_all(master->dma_rx); in pic32_spi_dma_transfer() 383 ret = dmaengine_slave_config(master->dma_rx, &cfg); in pic32_spi_dma_config() 554 dmaengine_terminate_all(master->dma_rx); in pic32_spi_one_transfer() 615 master->dma_rx = dma_request_chan(dev, "spi-rx"); in pic32_spi_dma_prep() 616 if (IS_ERR(master->dma_rx)) { in pic32_spi_dma_prep() 617 if (PTR_ERR(master->dma_rx) == -EPROBE_DEFER) in pic32_spi_dma_prep() 622 master->dma_rx = NULL; in pic32_spi_dma_prep() [all …]
|
D | spi-fsl-lpspi.c | 180 if (!controller->dma_rx) in fsl_lpspi_can_dma() 368 ret = dmaengine_slave_config(controller->dma_rx, &rx); in fsl_lpspi_dma_configure() 559 desc_rx = dmaengine_prep_slave_sg(controller->dma_rx, in fsl_lpspi_dma_transfer() 569 dma_async_issue_pending(controller->dma_rx); in fsl_lpspi_dma_transfer() 597 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer() 607 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer() 617 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer() 627 dmaengine_terminate_all(controller->dma_rx); in fsl_lpspi_dma_transfer() 640 if (controller->dma_rx) { in fsl_lpspi_dma_exit() 641 dma_release_channel(controller->dma_rx); in fsl_lpspi_dma_exit() [all …]
|
D | spi-ep93xx.c | 93 struct dma_chan *dma_rx; member 296 chan = espi->dma_rx; in ep93xx_spi_dma_prepare() 384 chan = espi->dma_rx; in ep93xx_spi_dma_finish() 430 dma_async_issue_pending(espi->dma_rx); in ep93xx_spi_dma_transfer() 505 if (espi->dma_rx && xfer->len > SPI_FIFO_SIZE) in ep93xx_spi_transfer_one() 603 espi->dma_rx = dma_request_channel(mask, ep93xx_spi_dma_filter, in ep93xx_spi_setup_dma() 605 if (!espi->dma_rx) { in ep93xx_spi_setup_dma() 624 dma_release_channel(espi->dma_rx); in ep93xx_spi_setup_dma() 625 espi->dma_rx = NULL; in ep93xx_spi_setup_dma() 634 if (espi->dma_rx) { in ep93xx_spi_release_dma() [all …]
|
D | spi-davinci.c | 120 struct dma_chan *dma_rx; member 396 if (dspi->dma_rx && dspi->dma_tx) in davinci_spi_of_setup() 621 dmaengine_slave_config(dspi->dma_rx, &dma_rx_conf); in davinci_spi_bufs() 624 rxdesc = dmaengine_prep_slave_sg(dspi->dma_rx, in davinci_spi_bufs() 657 dma_async_issue_pending(dspi->dma_rx); in davinci_spi_bufs() 748 dspi->dma_rx = dma_request_chan(sdev, "rx"); in davinci_spi_request_dma() 749 if (IS_ERR(dspi->dma_rx)) in davinci_spi_request_dma() 750 return PTR_ERR(dspi->dma_rx); in davinci_spi_request_dma() 754 dma_release_channel(dspi->dma_rx); in davinci_spi_request_dma() 963 dspi->dma_rx = NULL; in davinci_spi_probe() [all …]
|
D | spi-omap2-mcspi.c | 93 struct dma_chan *dma_rx; member 472 dmaengine_slave_config(mcspi_dma->dma_rx, &cfg); in omap2_mcspi_rx_dma() 503 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_rx, sg_out[0], in omap2_mcspi_rx_dma() 514 dma_async_issue_pending(mcspi_dma->dma_rx); in omap2_mcspi_rx_dma() 519 dmaengine_terminate_sync(mcspi_dma->dma_rx); in omap2_mcspi_rx_dma() 991 mcspi_dma->dma_rx = dma_request_chan(mcspi->dev, in omap2_mcspi_request_dma() 993 if (IS_ERR(mcspi_dma->dma_rx)) { in omap2_mcspi_request_dma() 994 ret = PTR_ERR(mcspi_dma->dma_rx); in omap2_mcspi_request_dma() 995 mcspi_dma->dma_rx = NULL; in omap2_mcspi_request_dma() 1004 dma_release_channel(mcspi_dma->dma_rx); in omap2_mcspi_request_dma() [all …]
|
D | spi-stm32.c | 310 struct dma_chan *dma_rx; member 665 if (spi->cur_usedma && spi->dma_rx) in stm32f4_spi_disable() 666 dmaengine_terminate_all(spi->dma_rx); in stm32f4_spi_disable() 708 if (spi->cur_usedma && spi->dma_rx) in stm32h7_spi_disable() 709 dmaengine_terminate_all(spi->dma_rx); in stm32h7_spi_disable() 1240 if (spi->rx_buf && spi->dma_rx) { in stm32_spi_transfer_one_dma() 1242 dmaengine_slave_config(spi->dma_rx, &rx_dma_conf); in stm32_spi_transfer_one_dma() 1249 spi->dma_rx, xfer->rx_sg.sgl, in stm32_spi_transfer_one_dma() 1268 (spi->rx_buf && spi->dma_rx && !rx_dma_desc)) in stm32_spi_transfer_one_dma() 1283 dma_async_issue_pending(spi->dma_rx); in stm32_spi_transfer_one_dma() [all …]
|
D | spi-atmel.c | 487 if (dmaengine_slave_config(master->dma_rx, slave_config)) { in atmel_spi_dma_slave_config() 510 master->dma_rx = dma_request_chan(dev, "rx"); in atmel_spi_configure_dma() 511 if (IS_ERR(master->dma_rx)) { in atmel_spi_configure_dma() 512 err = PTR_ERR(master->dma_rx); in atmel_spi_configure_dma() 528 dma_chan_name(master->dma_rx)); in atmel_spi_configure_dma() 532 if (!IS_ERR(master->dma_rx)) in atmel_spi_configure_dma() 533 dma_release_channel(master->dma_rx); in atmel_spi_configure_dma() 537 master->dma_tx = master->dma_rx = NULL; in atmel_spi_configure_dma() 543 if (master->dma_rx) in atmel_spi_stop_dma() 544 dmaengine_terminate_all(master->dma_rx); in atmel_spi_stop_dma() [all …]
|
D | spi-imx.c | 236 if (!master->dma_rx) in spi_imx_can_dma() 1202 ret = dmaengine_slave_config(master->dma_rx, &rx); in spi_imx_dma_configure() 1280 if (master->dma_rx) { in spi_imx_sdma_exit() 1281 dma_release_channel(master->dma_rx); in spi_imx_sdma_exit() 1282 master->dma_rx = NULL; in spi_imx_sdma_exit() 1308 master->dma_rx = dma_request_chan(dev, "rx"); in spi_imx_sdma_init() 1309 if (IS_ERR(master->dma_rx)) { in spi_imx_sdma_init() 1310 ret = PTR_ERR(master->dma_rx); in spi_imx_sdma_init() 1312 master->dma_rx = NULL; in spi_imx_sdma_init() 1396 desc_rx = dmaengine_prep_slave_sg(master->dma_rx, in spi_imx_dma_transfer() [all …]
|
D | spi-qup.c | 402 chan = master->dma_rx; in spi_qup_prep_sg() 422 dmaengine_terminate_all(master->dma_rx); in spi_qup_dma_terminate() 489 dma_async_issue_pending(master->dma_rx); in spi_qup_do_dma() 897 IS_ERR_OR_NULL(master->dma_rx)) in spi_qup_can_dma() 920 if (!IS_ERR_OR_NULL(master->dma_rx)) in spi_qup_release_dma() 921 dma_release_channel(master->dma_rx); in spi_qup_release_dma() 935 master->dma_rx = dma_request_chan(dev, "rx"); in spi_qup_init_dma() 936 if (IS_ERR(master->dma_rx)) in spi_qup_init_dma() 937 return PTR_ERR(master->dma_rx); in spi_qup_init_dma() 956 ret = dmaengine_slave_config(master->dma_rx, rx_conf); in spi_qup_init_dma() [all …]
|
D | spi-sh-msiof.c | 758 desc_rx = dmaengine_prep_slave_single(p->ctlr->dma_rx, in sh_msiof_dma_once() 807 dma_async_issue_pending(p->ctlr->dma_rx); in sh_msiof_dma_once() 849 dma_sync_single_for_cpu(p->ctlr->dma_rx->device->dev, in sh_msiof_dma_once() 862 dmaengine_terminate_sync(p->ctlr->dma_rx); in sh_msiof_dma_once() 1214 ctlr->dma_rx = sh_msiof_request_dma_chan(dev, DMA_DEV_TO_MEM, in sh_msiof_request_dma() 1216 if (!ctlr->dma_rx) in sh_msiof_request_dma() 1233 rx_dev = ctlr->dma_rx->device->dev; in sh_msiof_request_dma() 1249 dma_release_channel(ctlr->dma_rx); in sh_msiof_request_dma() 1263 dma_unmap_single(ctlr->dma_rx->device->dev, p->rx_dma_addr, PAGE_SIZE, in sh_msiof_release_dma() 1269 dma_release_channel(ctlr->dma_rx); in sh_msiof_release_dma()
|
D | spi-rspi.c | 549 desc_rx = dmaengine_prep_slave_sg(rspi->ctlr->dma_rx, rx->sgl, in rspi_dma_transfer() 607 dma_async_issue_pending(rspi->ctlr->dma_rx); in rspi_dma_transfer() 618 dmaengine_synchronize(rspi->ctlr->dma_rx); in rspi_dma_transfer() 627 dmaengine_terminate_sync(rspi->ctlr->dma_rx); in rspi_dma_transfer() 641 dmaengine_terminate_sync(rspi->ctlr->dma_rx); in rspi_dma_transfer() 1153 ctlr->dma_rx = rspi_request_dma_chan(dev, DMA_DEV_TO_MEM, dma_rx_id, in rspi_request_dma() 1155 if (!ctlr->dma_rx) { in rspi_request_dma() 1170 if (ctlr->dma_rx) in rspi_release_dma() 1171 dma_release_channel(ctlr->dma_rx); in rspi_release_dma()
|
/drivers/net/ethernet/micrel/ |
D | ks8842.c | 148 ((adapter)->dma_rx.channel != -1)) 160 struct ks8842_rx_dma_ctl dma_rx; member 543 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; in __ks8842_start_new_rx_dma() 592 struct ks8842_adapter *adapter = from_tasklet(adapter, t, dma_rx.tasklet); in ks8842_rx_frame_dma_tasklet() 594 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; in ks8842_rx_frame_dma_tasklet() 837 if (adapter->dma_rx.adesc) in ks8842_dma_rx_cb() 838 tasklet_schedule(&adapter->dma_rx.tasklet); in ks8842_dma_rx_cb() 862 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_stop_dma() 884 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_dealloc_dma_bufs() 916 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; in ks8842_alloc_dma_bufs() [all …]
|
/drivers/mtd/nand/raw/ |
D | intel-nand-controller.c | 122 struct dma_chan *dma_rx; member 277 dmaengine_terminate_async(ebu_host->dma_rx); in ebu_dma_rx_callback() 305 chan = ebu_host->dma_rx; in ebu_dma_start() 573 if (ebu_host->dma_rx) in ebu_dma_cleanup() 574 dma_release_channel(ebu_host->dma_rx); in ebu_dma_cleanup() 655 ebu_host->dma_rx = dma_request_chan(dev, "rx"); in ebu_nand_probe() 656 if (IS_ERR(ebu_host->dma_rx)) { in ebu_nand_probe() 657 ret = dev_err_probe(dev, PTR_ERR(ebu_host->dma_rx), in ebu_nand_probe() 659 ebu_host->dma_rx = NULL; in ebu_nand_probe()
|
/drivers/i2c/busses/ |
D | i2c-sh_mobile.c | 136 struct dma_chan *dma_rx; member 448 ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_dma_unmap() 461 dmaengine_terminate_sync(pd->dma_rx); in sh_mobile_i2c_cleanup_dma() 519 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_xfer_dma() 526 chan = pd->dma_rx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_DEV_TO_MEM, in sh_mobile_i2c_xfer_dma() 833 if (!IS_ERR(pd->dma_rx)) { in sh_mobile_i2c_release_dma() 834 dma_release_channel(pd->dma_rx); in sh_mobile_i2c_release_dma() 835 pd->dma_rx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_release_dma() 919 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_probe()
|
D | i2c-rcar.c | 139 struct dma_chan *dma_rx; member 373 ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma_unmap() 394 dmaengine_terminate_all(priv->dma_rx); in rcar_i2c_cleanup_dma() 416 struct dma_chan *chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma() 766 chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_request_dma() 774 priv->dma_rx = chan; in rcar_i2c_request_dma() 786 if (!IS_ERR(priv->dma_rx)) { in rcar_i2c_release_dma() 787 dma_release_channel(priv->dma_rx); in rcar_i2c_release_dma() 788 priv->dma_rx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_release_dma() 1067 priv->dma_rx = priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_probe()
|
/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_dma.c | 45 dma_addr_t dma_rx, int t_rsize, int r_rsize) in sxgbe_dma_channel_init() argument 71 writel(upper_32_bits(dma_rx), in sxgbe_dma_channel_init() 73 writel(lower_32_bits(dma_rx), in sxgbe_dma_channel_init() 84 dma_addr = dma_rx + ((r_rsize - 1) * SXGBE_DESC_SIZE_BYTES); in sxgbe_dma_channel_init()
|
/drivers/mmc/host/ |
D | omap.c | 127 struct dma_chan *dma_rx; member 414 c = host->dma_rx; in mmc_omap_release_dma() 1006 c = host->dma_rx; in mmc_omap_prepare_data() 1077 host->dma_tx : host->dma_rx; in mmc_omap_start_request() 1400 host->dma_rx = dma_request_chan(&pdev->dev, "rx"); in mmc_omap_probe() 1401 if (IS_ERR(host->dma_rx)) { in mmc_omap_probe() 1402 ret = PTR_ERR(host->dma_rx); in mmc_omap_probe() 1410 host->dma_rx = NULL; in mmc_omap_probe() 1455 if (host->dma_rx) in mmc_omap_probe() 1456 dma_release_channel(host->dma_rx); in mmc_omap_probe() [all …]
|
D | davinci_mmc.c | 193 struct dma_chan *dma_rx; member 393 sync_dev = host->dma_rx; in davinci_abort_dma() 437 chan = host->dma_rx; in mmc_davinci_send_dma_request() 438 dmaengine_slave_config(host->dma_rx, &dma_rx_conf); in mmc_davinci_send_dma_request() 440 desc = dmaengine_prep_slave_sg(host->dma_rx, in mmc_davinci_send_dma_request() 492 dma_release_channel(host->dma_rx); in davinci_release_dma_channels() 503 host->dma_rx = dma_request_chan(mmc_dev(host->mmc), "rx"); in davinci_acquire_dma_channels() 504 if (IS_ERR(host->dma_rx)) { in davinci_acquire_dma_channels() 507 return PTR_ERR(host->dma_rx); in davinci_acquire_dma_channels()
|
/drivers/net/ethernet/calxeda/ |
D | xgmac.c | 355 struct xgmac_dma_desc *dma_rx; member 687 p = priv->dma_rx + entry; in xgmac_rx_refill() 737 priv->dma_rx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init() 742 if (!priv->dma_rx) in xgmac_dma_desc_rings_init() 760 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init() 766 desc_init_rx_desc(priv->dma_rx, DMA_RX_RING_SZ, priv->dma_buf_sz); in xgmac_dma_desc_rings_init() 783 priv->dma_rx, priv->dma_rx_phy); in xgmac_dma_desc_rings_init() 802 p = priv->dma_rx + i; in xgmac_free_rx_skbufs() 849 if (priv->dma_rx) { in xgmac_free_dma_desc_rings() 852 priv->dma_rx, priv->dma_rx_phy); in xgmac_free_dma_desc_rings() [all …]
|