/drivers/net/irda/ |
D | sa1100_ir.c | 64 struct sa1100_buf dma_tx; member 230 dma_unmap_sg(si->dma_tx.dev, &si->dma_tx.sg, 1, DMA_TO_DEVICE); in sa1100_irda_sirtxdma_irq() 231 dev_kfree_skb(si->dma_tx.skb); in sa1100_irda_sirtxdma_irq() 232 si->dma_tx.skb = NULL; in sa1100_irda_sirtxdma_irq() 235 dev->stats.tx_bytes += sg_dma_len(&si->dma_tx.sg); in sa1100_irda_sirtxdma_irq() 262 si->dma_tx.skb = skb; in sa1100_irda_sir_tx_start() 263 sg_set_buf(&si->dma_tx.sg, si->tx_buff.data, si->tx_buff.len); in sa1100_irda_sir_tx_start() 264 if (dma_map_sg(si->dma_tx.dev, &si->dma_tx.sg, 1, DMA_TO_DEVICE) == 0) { in sa1100_irda_sir_tx_start() 265 si->dma_tx.skb = NULL; in sa1100_irda_sir_tx_start() 271 sa1100_irda_dma_start(&si->dma_tx, DMA_MEM_TO_DEV, sa1100_irda_sirtxdma_irq, dev); in sa1100_irda_sir_tx_start() [all …]
|
/drivers/spi/ |
D | spi-pxa2xx-dma.c | 109 chan = drv_data->master->dma_tx; in pxa2xx_spi_dma_prepare_one() 138 dmaengine_terminate_async(drv_data->master->dma_tx); in pxa2xx_spi_dma_transfer() 177 dmaengine_terminate_async(drv_data->master->dma_tx); in pxa2xx_spi_dma_prepare() 185 dma_async_issue_pending(drv_data->master->dma_tx); in pxa2xx_spi_dma_start() 200 master->dma_tx = dma_request_slave_channel_compat(mask, in pxa2xx_spi_dma_setup() 202 if (!master->dma_tx) in pxa2xx_spi_dma_setup() 208 dma_release_channel(master->dma_tx); in pxa2xx_spi_dma_setup() 209 master->dma_tx = NULL; in pxa2xx_spi_dma_setup() 225 if (master->dma_tx) { in pxa2xx_spi_dma_release() 226 dmaengine_terminate_sync(master->dma_tx); in pxa2xx_spi_dma_release() [all …]
|
D | spi-rockchip.c | 199 struct rockchip_spi_dma_data dma_tx; member 331 dmaengine_terminate_async(rs->dma_tx.ch); in rockchip_spi_handle_err() 475 txconf.direction = rs->dma_tx.direction; in rockchip_spi_prepare_dma() 476 txconf.dst_addr = rs->dma_tx.addr; in rockchip_spi_prepare_dma() 482 dmaengine_slave_config(rs->dma_tx.ch, &txconf); in rockchip_spi_prepare_dma() 485 rs->dma_tx.ch, in rockchip_spi_prepare_dma() 487 rs->dma_tx.direction, DMA_PREP_INTERRUPT); in rockchip_spi_prepare_dma() 512 dma_async_issue_pending(rs->dma_tx.ch); in rockchip_spi_prepare_dma() 752 rs->dma_tx.ch = dma_request_chan(rs->dev, "tx"); in rockchip_spi_probe() 753 if (IS_ERR(rs->dma_tx.ch)) { in rockchip_spi_probe() [all …]
|
D | spi-bcm2835.c | 236 dmaengine_terminate_all(master->dma_tx); in bcm2835_spi_dma_done() 260 chan = master->dma_tx; in bcm2835_spi_prepare_sg() 325 dma_async_issue_pending(master->dma_tx); in bcm2835_spi_transfer_one_dma() 344 dmaengine_terminate_all(master->dma_tx); in bcm2835_spi_transfer_one_dma() 406 if (master->dma_tx) { in bcm2835_dma_release() 407 dmaengine_terminate_all(master->dma_tx); in bcm2835_dma_release() 408 dma_release_channel(master->dma_tx); in bcm2835_dma_release() 409 master->dma_tx = NULL; in bcm2835_dma_release() 434 master->dma_tx = dma_request_slave_channel(dev, "tx"); in bcm2835_dma_init() 435 if (!master->dma_tx) { in bcm2835_dma_init() [all …]
|
D | spi-pic32.c | 317 if (!master->dma_rx || !master->dma_tx) in pic32_spi_dma_transfer() 330 desc_tx = dmaengine_prep_slave_sg(master->dma_tx, in pic32_spi_dma_transfer() 355 dma_async_issue_pending(master->dma_tx); in pic32_spi_dma_transfer() 382 ret = dmaengine_slave_config(master->dma_tx, &cfg); in pic32_spi_dma_config() 632 master->dma_tx = dma_request_slave_channel_compat(mask, NULL, NULL, in pic32_spi_dma_prep() 634 if (!master->dma_tx) { in pic32_spi_dma_prep() 651 if (master->dma_tx) in pic32_spi_dma_prep() 652 dma_release_channel(master->dma_tx); in pic32_spi_dma_prep() 664 if (pic32s->master->dma_tx) in pic32_spi_dma_unprep() 665 dma_release_channel(pic32s->master->dma_tx); in pic32_spi_dma_unprep()
|
D | spi-sh-msiof.c | 702 dma_sync_single_for_device(p->master->dma_tx->device->dev, in sh_msiof_dma_once() 704 desc_tx = dmaengine_prep_slave_single(p->master->dma_tx, in sh_msiof_dma_once() 740 dma_async_issue_pending(p->master->dma_tx); in sh_msiof_dma_once() 776 dmaengine_terminate_all(p->master->dma_tx); in sh_msiof_dma_once() 848 while (master->dma_tx && len > 15) { in sh_msiof_transfer_one() 1085 master->dma_tx = sh_msiof_request_dma_chan(dev, DMA_MEM_TO_DEV, in sh_msiof_request_dma() 1088 if (!master->dma_tx) in sh_msiof_request_dma() 1105 tx_dev = master->dma_tx->device->dev; in sh_msiof_request_dma() 1129 dma_release_channel(master->dma_tx); in sh_msiof_request_dma() 1130 master->dma_tx = NULL; in sh_msiof_request_dma() [all …]
|
D | spi-omap2-mcspi.c | 102 struct dma_chan *dma_tx; member 420 if (mcspi_dma->dma_tx) { in omap2_mcspi_tx_dma() 423 dmaengine_slave_config(mcspi_dma->dma_tx, &cfg); in omap2_mcspi_tx_dma() 425 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl, in omap2_mcspi_tx_dma() 437 dma_async_issue_pending(mcspi_dma->dma_tx); in omap2_mcspi_tx_dma() 1013 mcspi_dma->dma_tx = dma_request_chan(&master->dev, in omap2_mcspi_request_dma() 1015 if (IS_ERR(mcspi_dma->dma_tx)) { in omap2_mcspi_request_dma() 1016 ret = PTR_ERR(mcspi_dma->dma_tx); in omap2_mcspi_request_dma() 1017 mcspi_dma->dma_tx = NULL; in omap2_mcspi_request_dma() 1060 if (!mcspi_dma->dma_rx || !mcspi_dma->dma_tx) { in omap2_mcspi_setup() [all …]
|
D | spi-qup.c | 297 chan = master->dma_tx; in spi_qup_prep_sg() 322 dmaengine_terminate_all(master->dma_tx); in spi_qup_dma_terminate() 350 dma_async_issue_pending(master->dma_tx); in spi_qup_do_dma() 681 IS_ERR_OR_NULL(master->dma_tx) || in spi_qup_can_dma() 698 if (!IS_ERR_OR_NULL(master->dma_tx)) in spi_qup_release_dma() 699 dma_release_channel(master->dma_tx); in spi_qup_release_dma() 715 master->dma_tx = dma_request_slave_channel_reason(dev, "tx"); in spi_qup_init_dma() 716 if (IS_ERR(master->dma_tx)) { in spi_qup_init_dma() 717 ret = PTR_ERR(master->dma_tx); in spi_qup_init_dma() 738 ret = dmaengine_slave_config(master->dma_tx, tx_conf); in spi_qup_init_dma() [all …]
|
D | spi-davinci.c | 129 struct dma_chan *dma_tx; member 640 dmaengine_slave_config(dspi->dma_tx, &dma_tx_conf); in davinci_spi_bufs() 676 txdesc = dmaengine_prep_slave_sg(dspi->dma_tx, in davinci_spi_bufs() 694 dma_async_issue_pending(dspi->dma_tx); in davinci_spi_bufs() 801 dspi->dma_tx = dma_request_chan(sdev, "tx"); in davinci_spi_request_dma() 802 if (IS_ERR(dspi->dma_tx)) { in davinci_spi_request_dma() 804 return PTR_ERR(dspi->dma_tx); in davinci_spi_request_dma() 1032 dspi->dma_tx = NULL; in davinci_spi_probe() 1072 dma_release_channel(dspi->dma_tx); in davinci_spi_probe() 1106 dma_release_channel(dspi->dma_tx); in davinci_spi_remove()
|
D | spi-imx.c | 843 ret = dmaengine_slave_config(master->dma_tx, &tx); in spi_imx_dma_configure() 917 if (master->dma_tx) { in spi_imx_sdma_exit() 918 dma_release_channel(master->dma_tx); in spi_imx_sdma_exit() 919 master->dma_tx = NULL; in spi_imx_sdma_exit() 935 master->dma_tx = dma_request_slave_channel_reason(dev, "tx"); in spi_imx_sdma_init() 936 if (IS_ERR(master->dma_tx)) { in spi_imx_sdma_init() 937 ret = PTR_ERR(master->dma_tx); in spi_imx_sdma_init() 939 master->dma_tx = NULL; in spi_imx_sdma_init() 1020 desc_tx = dmaengine_prep_slave_sg(master->dma_tx, in spi_imx_dma_transfer() 1024 dmaengine_terminate_all(master->dma_tx); in spi_imx_dma_transfer() [all …]
|
D | spi-ep93xx.c | 102 struct dma_chan *dma_tx; member 461 chan = espi->dma_tx; in ep93xx_spi_dma_prepare() 543 chan = espi->dma_tx; in ep93xx_spi_dma_finish() 584 dma_async_issue_pending(espi->dma_tx); in ep93xx_spi_dma_transfer() 825 espi->dma_tx = dma_request_channel(mask, ep93xx_spi_dma_filter, in ep93xx_spi_setup_dma() 827 if (!espi->dma_tx) { in ep93xx_spi_setup_dma() 849 if (espi->dma_tx) { in ep93xx_spi_release_dma() 850 dma_release_channel(espi->dma_tx); in ep93xx_spi_release_dma()
|
D | spi-dw-mid.c | 49 struct dw_dma_slave *tx = dws->dma_tx; in mid_spi_dma_init() 76 dws->master->dma_tx = dws->txchan; in mid_spi_dma_init() 325 dws->dma_tx = &mid_dma_tx; in dw_spi_mid_init()
|
D | spi-rspi.c | 556 desc_tx = dmaengine_prep_slave_sg(rspi->master->dma_tx, in rspi_dma_transfer() 596 dma_async_issue_pending(rspi->master->dma_tx); in rspi_dma_transfer() 606 dmaengine_terminate_all(rspi->master->dma_tx); in rspi_dma_transfer() 1067 master->dma_tx = rspi_request_dma_chan(dev, DMA_MEM_TO_DEV, dma_tx_id, in rspi_request_dma() 1069 if (!master->dma_tx) in rspi_request_dma() 1075 dma_release_channel(master->dma_tx); in rspi_request_dma() 1076 master->dma_tx = NULL; in rspi_request_dma() 1087 if (master->dma_tx) in rspi_release_dma() 1088 dma_release_channel(master->dma_tx); in rspi_release_dma()
|
/drivers/i2c/busses/ |
D | i2c-rcar.c | 139 struct dma_chan *dma_tx; member 313 ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma_unmap() 334 dmaengine_terminate_all(priv->dma_tx); in rcar_i2c_cleanup_dma() 354 struct dma_chan *chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma() 666 chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_request_dma() 676 priv->dma_tx = chan; in rcar_i2c_request_dma() 681 if (!IS_ERR(priv->dma_tx)) { in rcar_i2c_release_dma() 682 dma_release_channel(priv->dma_tx); in rcar_i2c_release_dma() 683 priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_release_dma() 852 priv->dma_rx = priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_probe()
|
D | i2c-sh_mobile.c | 145 struct dma_chan *dma_tx; member 524 ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_dma_unmap() 539 dmaengine_terminate_all(pd->dma_tx); in sh_mobile_i2c_cleanup_dma() 595 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_xfer_dma() 605 chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV, in sh_mobile_i2c_xfer_dma() 846 if (!IS_ERR(pd->dma_tx)) { in sh_mobile_i2c_release_dma() 847 dma_release_channel(pd->dma_tx); in sh_mobile_i2c_release_dma() 848 pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_release_dma() 950 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_probe()
|
/drivers/mmc/host/ |
D | omap.c | 131 struct dma_chan *dma_tx; member 413 c = host->dma_tx; in mmc_omap_release_dma() 1003 c = host->dma_tx; in mmc_omap_prepare_data() 1079 host->dma_tx : host->dma_rx; in mmc_omap_start_request() 1387 host->dma_tx = dma_request_chan(&pdev->dev, "tx"); in mmc_omap_probe() 1388 if (IS_ERR(host->dma_tx)) { in mmc_omap_probe() 1389 ret = PTR_ERR(host->dma_tx); in mmc_omap_probe() 1395 host->dma_tx = NULL; in mmc_omap_probe() 1403 if (host->dma_tx) in mmc_omap_probe() 1404 dma_release_channel(host->dma_tx); in mmc_omap_probe() [all …]
|
D | davinci_mmc.c | 203 struct dma_chan *dma_tx; member 406 sync_dev = host->dma_tx; in davinci_abort_dma() 426 chan = host->dma_tx; in mmc_davinci_send_dma_request() 427 dmaengine_slave_config(host->dma_tx, &dma_tx_conf); in mmc_davinci_send_dma_request() 429 desc = dmaengine_prep_slave_sg(host->dma_tx, in mmc_davinci_send_dma_request() 507 dma_release_channel(host->dma_tx); in davinci_release_dma_channels() 513 host->dma_tx = dma_request_chan(mmc_dev(host->mmc), "tx"); in davinci_acquire_dma_channels() 514 if (IS_ERR(host->dma_tx)) { in davinci_acquire_dma_channels() 516 return PTR_ERR(host->dma_tx); in davinci_acquire_dma_channels() 522 dma_release_channel(host->dma_tx); in davinci_acquire_dma_channels()
|
D | jz4740_mmc.c | 150 struct dma_chan *dma_tx; member 170 dma_release_channel(host->dma_tx); in jz4740_mmc_release_dma_channels() 181 host->dma_tx = dma_request_channel(mask, NULL, host); in jz4740_mmc_acquire_dma_channels() 182 if (!host->dma_tx) { in jz4740_mmc_acquire_dma_channels() 199 dma_release_channel(host->dma_tx); in jz4740_mmc_acquire_dma_channels() 211 return (data->flags & MMC_DATA_READ) ? host->dma_rx : host->dma_tx; in jz4740_mmc_get_dma_chan() 287 chan = host->dma_tx; in jz4740_mmc_start_dma_transfer()
|
/drivers/net/ethernet/micrel/ |
D | ks8842.c | 159 #define KS8842_USE_DMA(adapter) (((adapter)->dma_tx.channel != -1) && \ 171 struct ks8842_tx_dma_ctl dma_tx; member 434 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; in ks8842_tx_frame_dma() 859 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; in ks8842_dma_tx_cb() 875 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; in ks8842_stop_dma() 897 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; in ks8842_dealloc_dma_bufs() 929 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; in ks8842_alloc_dma_bufs() 1003 adapter->dma_tx.channel = -1; in ks8842_open() 1057 if (adapter->dma_tx.adesc) in ks8842_xmit_frame() 1184 adapter->dma_tx.channel = pdata->tx_dma_channel; in ks8842_probe() [all …]
|
/drivers/net/ethernet/samsung/sxgbe/ |
D | sxgbe_dma.c | 47 int fix_burst, int pbl, dma_addr_t dma_tx, in sxgbe_dma_channel_init() argument 69 writel(upper_32_bits(dma_tx), in sxgbe_dma_channel_init() 71 writel(lower_32_bits(dma_tx), in sxgbe_dma_channel_init() 83 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES); in sxgbe_dma_channel_init()
|
/drivers/net/ethernet/stmicro/stmmac/ |
D | dwmac100_dma.c | 36 int aal, u32 dma_tx, u32 dma_rx, int atds) in dwmac100_dma_init() argument 48 writel(dma_tx, ioaddr + DMA_TX_BASE_ADDR); in dwmac100_dma_init()
|
D | chain_mode.c | 35 struct dma_desc *desc = priv->dma_tx + entry; in stmmac_jumbo_frm() 60 desc = priv->dma_tx + entry; in stmmac_jumbo_frm()
|
D | ring_mode.c | 42 desc = priv->dma_tx + entry; in stmmac_jumbo_frm() 71 desc = priv->dma_tx + entry; in stmmac_jumbo_frm()
|
/drivers/net/ethernet/calxeda/ |
D | xgmac.c | 370 struct xgmac_dma_desc *dma_tx; member 760 priv->dma_tx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init() 765 if (!priv->dma_tx) in xgmac_dma_desc_rings_init() 770 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init() 781 desc_init_tx_desc(priv->dma_tx, DMA_TX_RING_SZ); in xgmac_dma_desc_rings_init() 832 p = priv->dma_tx + i; in xgmac_free_tx_skbufs() 853 if (priv->dma_tx) { in xgmac_free_dma_desc_rings() 856 priv->dma_tx, priv->dma_tx_phy); in xgmac_free_dma_desc_rings() 857 priv->dma_tx = NULL; in xgmac_free_dma_desc_rings() 881 struct xgmac_dma_desc *p = priv->dma_tx + entry; in xgmac_tx_complete() [all …]
|
/drivers/net/wan/ |
D | z85230.c | 541 if(!chan->dma_tx) in z8530_dma_tx() 572 if(chan->dma_tx) in z8530_dma_status() 807 c->dma_tx = 0; in z8530_sync_open() 897 c->dma_tx = 1; in z8530_sync_dma_open() 1098 c->dma_tx = 1; in z8530_sync_txdma_open() 1444 if(c->dma_tx) in z8530_tx_begin() 1465 if(c->dma_tx) in z8530_tx_begin() 1747 …if(c->dma_tx && ((unsigned long)(virt_to_bus(skb->data+skb->len))>=16*1024*1024 || spans_boundary(… in z8530_queue_xmit()
|