/drivers/spi/ |
D | spi-s3c64xx.c | 200 struct s3c64xx_spi_dma_data rx_dma; member 266 struct s3c64xx_spi_driver_data, rx_dma); in s3c64xx_spi_dmacb() 302 struct s3c64xx_spi_driver_data, rx_dma); in prepare_dma() 303 config.direction = sdd->rx_dma.direction; in prepare_dma() 306 sdd->ops->config((enum dma_ch)sdd->rx_dma.ch, &config); in prepare_dma() 337 sdd->rx_dma.ch = (void *)sdd->ops->request(sdd->rx_dma.dmach, &req, dev, "rx"); in acquire_dma() 361 sdd->ops->release((enum dma_ch)sdd->rx_dma.ch, &s3c64xx_spi_dma_client); in s3c64xx_spi_unprepare_transfer() 386 struct s3c64xx_spi_driver_data, rx_dma); in prepare_dma() 430 sdd->rx_dma.ch = dma_request_slave_channel_compat(mask, filter, in s3c64xx_spi_prepare_transfer() 431 (void*)sdd->rx_dma.dmach, dev, "rx"); in s3c64xx_spi_prepare_transfer() [all …]
|
D | spi-fsl-cpm.c | 78 if (mspi->rx_dma == mspi->dma_dummy_rx) in fsl_spi_cpm_bufs_start() 79 out_be32(&rx_bd->cbd_bufaddr, mspi->rx_dma); in fsl_spi_cpm_bufs_start() 81 out_be32(&rx_bd->cbd_bufaddr, mspi->rx_dma + xfer_ofs); in fsl_spi_cpm_bufs_start() 117 mspi->rx_dma = mspi->dma_dummy_rx; in fsl_spi_cpm_bufs() 135 mspi->rx_dma = dma_map_single(dev, mspi->rx, t->len, in fsl_spi_cpm_bufs() 137 if (dma_mapping_error(dev, mspi->rx_dma)) { in fsl_spi_cpm_bufs() 142 mspi->rx_dma = t->rx_dma; in fsl_spi_cpm_bufs() 170 dma_unmap_single(dev, mspi->rx_dma, t->len, DMA_FROM_DEVICE); in fsl_spi_cpm_bufs_complete()
|
D | spi-atmel.c | 574 as->dma.sgrx.dma_address = xfer->rx_dma + xfer->len - *plen; in atmel_spi_next_xfer_dma_submit() 622 xfer->rx_buf, xfer->rx_dma); in atmel_spi_next_xfer_dma_submit() 656 dma_addr_t *rx_dma, in atmel_spi_next_xfer_data() argument 664 *rx_dma = xfer->rx_dma + xfer->len - *plen; in atmel_spi_next_xfer_data() 666 *rx_dma = as->buffer_dma; in atmel_spi_next_xfer_data() 696 dma_addr_t tx_dma, rx_dma; in atmel_spi_pdc_next_xfer() local 711 atmel_spi_next_xfer_data(master, xfer, &tx_dma, &rx_dma, &len); in atmel_spi_pdc_next_xfer() 714 spi_writel(as, RPR, rx_dma); in atmel_spi_pdc_next_xfer() 725 xfer->rx_buf, xfer->rx_dma); in atmel_spi_pdc_next_xfer() 750 atmel_spi_next_xfer_data(master, xfer, &tx_dma, &rx_dma, &len); in atmel_spi_pdc_next_xfer() [all …]
|
D | spi-pxa2xx-pxadma.c | 61 return drv_data->rx_dma && drv_data->tx_dma; in pxa2xx_spi_map_dma_buffers() 93 drv_data->rx_dma = dma_map_single(dev, drv_data->rx, in pxa2xx_spi_map_dma_buffers() 95 if (dma_mapping_error(dev, drv_data->rx_dma)) { in pxa2xx_spi_map_dma_buffers() 113 dma_unmap_single(dev, drv_data->rx_dma, in pxa2xx_spi_unmap_dma_buffers() 296 DTADR(drv_data->rx_channel) = drv_data->rx_dma; in pxa2xx_spi_dma_prepare()
|
D | spi-dw-mid.c | 124 if (dws->rx_dma) in mid_spi_dma_transfer() 169 dws->rx_sgl.dma_address = dws->rx_dma; in mid_spi_dma_transfer()
|
D | spi-davinci.c | 598 t->rx_dma = dma_map_single(&spi->dev, buf, in davinci_spi_bufs() 600 if (!t->rx_dma) { in davinci_spi_bufs() 604 sg_dma_address(&sg_rx) = t->rx_dma; in davinci_spi_bufs() 666 dma_unmap_single(&spi->dev, t->rx_dma, in davinci_spi_bufs() 697 dma_unmap_single(&spi->dev, t->rx_dma, t->len, DMA_FROM_DEVICE); in davinci_spi_bufs()
|
D | spi-dw.c | 256 if (dws->cur_transfer->rx_dma) in map_dma_buffers() 257 dws->rx_dma = dws->cur_transfer->rx_dma; in map_dma_buffers() 424 dws->rx_dma = transfer->rx_dma; in pump_transfers()
|
D | spi-fsl-lib.h | 46 dma_addr_t rx_dma; member
|
D | spi-pxa2xx.h | 80 dma_addr_t rx_dma; member
|
D | spi-dw.h | 131 dma_addr_t rx_dma; member
|
D | spi-omap2-mcspi.c | 396 sg_dma_address(&sg) = xfer->rx_dma; in omap2_mcspi_rx_dma() 415 dma_unmap_single(mcspi->dev, xfer->rx_dma, count, in omap2_mcspi_rx_dma() 1107 t->rx_dma = dma_map_single(mcspi->dev, rx_buf, t->len, in omap2_mcspi_transfer_one_message() 1109 if (dma_mapping_error(mcspi->dev, t->rx_dma)) { in omap2_mcspi_transfer_one_message()
|
D | spi-au1550.c | 366 dma_rx_addr = t->rx_dma; in au1550_spi_dma_txrxb() 387 if (t->rx_dma == 0) { /* if DMA_ADDR_INVALID, map it */ in au1550_spi_dma_txrxb() 449 if (t->rx_buf && t->rx_dma == 0 ) in au1550_spi_dma_txrxb()
|
/drivers/net/ethernet/amd/ |
D | au1000_eth.h | 64 struct rx_dma { struct 91 struct rx_dma *rx_dma_ring[NUM_RX_DMA];
|
D | au1000_eth.c | 549 aup->rx_dma_ring[i] = (struct rx_dma *) in au1000_setup_hw_rings() 550 (tx_base + 0x100 + sizeof(struct rx_dma) * i); in au1000_setup_hw_rings() 710 struct rx_dma *prxd; in au1000_rx()
|
/drivers/infiniband/ulp/iser/ |
D | iser_initiator.c | 454 u64 rx_dma; in iser_rcv_completion() local 459 rx_dma = ib_conn->login_resp_dma; in iser_rcv_completion() 462 rx_dma = rx_desc->dma_addr; in iser_rcv_completion() 466 ib_dma_sync_single_for_cpu(ib_conn->device->ib_device, rx_dma, in iser_rcv_completion() 477 ib_dma_sync_single_for_device(ib_conn->device->ib_device, rx_dma, in iser_rcv_completion() 486 if (rx_dma == ib_conn->login_resp_dma) in iser_rcv_completion()
|
/drivers/net/ethernet/ |
D | s6gmac.c | 349 u32 rx_dma; member 378 (!s6dmac_fifo_full(pd->rx_dma, pd->rx_chan)) && in s6gmac_rx_fillfifo() 381 s6dmac_put_fifo_cache(pd->rx_dma, pd->rx_chan, in s6gmac_rx_fillfifo() 392 s6dmac_pending_count(pd->rx_dma, pd->rx_chan)) { in s6gmac_rx_interrupt() 516 if (s6dmac_termcnt_irq(pd->rx_dma, pd->rx_chan)) in s6gmac_interrupt() 847 s6dmac_disable_chan(pd->rx_dma, pd->rx_chan); in s6gmac_init_dmac() 849 s6dmac_disable_error_irqs(pd->rx_dma, 1 << S6_HIFDMA_GMACRX); in s6gmac_init_dmac() 899 s6dmac_enable_chan(pd->rx_dma, pd->rx_chan, in s6gmac_open() 984 pd->rx_dma = DMA_MASK_DMAC(i); in s6gmac_probe()
|
/drivers/net/ethernet/apple/ |
D | bmac.c | 67 volatile struct dbdma_regs __iomem *rx_dma; member 229 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_enable_and_reset_chip() 406 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_start_chip() 477 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_suspend() 613 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_init_rx_ring() 683 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_rxdma_intr() 1318 bp->rx_dma = ioremap(macio_resource_start(mdev, 2), macio_resource_len(mdev, 2)); in bmac_probe() 1319 if (!bp->rx_dma) in bmac_probe() 1372 iounmap(bp->rx_dma); in bmac_probe() 1400 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_close() [all …]
|
D | mace.c | 48 volatile struct dbdma_regs __iomem *rx_dma; member 187 mp->rx_dma = ioremap(macio_resource_start(mdev, 2), 0x1000); in mace_probe() 188 if (mp->rx_dma == NULL) { in mace_probe() 261 iounmap(mp->rx_dma); in mace_probe() 291 iounmap(mp->rx_dma); in mace_remove() 432 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_open() 502 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_close() 811 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_tx_timeout() 880 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_rxdma_intr()
|
/drivers/mmc/host/ |
D | mmc_spi.c | 534 t->tx_dma = t->rx_dma = host->data_dma; in mmc_spi_command_send() 626 t->rx_dma = dma + offsetof(struct scratch, crc_val); in mmc_spi_setup_data_message() 654 t->rx_dma = dma + offsetof(struct scratch, status); in mmc_spi_setup_data_message() 822 t->rx_dma, t->len, in mmc_spi_readblock() 833 t->rx_dma, t->len, in mmc_spi_readblock() 872 t->rx_dma += t->len; in mmc_spi_readblock() 936 t->rx_dma = dma_addr + sg->offset; in mmc_spi_data_do() 1014 host->early_status.tx_dma = host->early_status.rx_dma; in mmc_spi_data_do() 1440 host->status.rx_dma = host->data_dma + offsetof(struct scratch, status); in mmc_spi_probe()
|
/drivers/staging/et131x/ |
D | et131x.c | 1787 struct rxdma_regs __iomem *rx_dma = &adapter->regs->rxdma; in et131x_config_rx_dma_regs() local 1799 writel(upper_32_bits(rx_local->rx_status_bus), &rx_dma->dma_wb_base_hi); in et131x_config_rx_dma_regs() 1800 writel(lower_32_bits(rx_local->rx_status_bus), &rx_dma->dma_wb_base_lo); in et131x_config_rx_dma_regs() 1807 writel(upper_32_bits(rx_local->ps_ring_physaddr), &rx_dma->psr_base_hi); in et131x_config_rx_dma_regs() 1808 writel(lower_32_bits(rx_local->ps_ring_physaddr), &rx_dma->psr_base_lo); in et131x_config_rx_dma_regs() 1809 writel(rx_local->psr_num_entries - 1, &rx_dma->psr_num_des); in et131x_config_rx_dma_regs() 1810 writel(0, &rx_dma->psr_full_offset); in et131x_config_rx_dma_regs() 1812 psr_num_des = readl(&rx_dma->psr_num_des) & ET_RXDMA_PSR_NUM_DES_MASK; in et131x_config_rx_dma_regs() 1814 &rx_dma->psr_min_des); in et131x_config_rx_dma_regs() 1829 num_des = &rx_dma->fbr0_num_des; in et131x_config_rx_dma_regs() [all …]
|
/drivers/tty/serial/ |
D | ifx6x60.h | 87 dma_addr_t rx_dma; member
|
D | ifx6x60.c | 775 ifx_dev->rx_dma = ifx_dev->rx_bus; in ifx_spi_io() 777 ifx_dev->spi_xfer.rx_dma = ifx_dev->rx_dma; in ifx_spi_io() 781 ifx_dev->rx_dma = (dma_addr_t)0; in ifx_spi_io() 783 ifx_dev->spi_xfer.rx_dma = (dma_addr_t)0; in ifx_spi_io()
|
/drivers/atm/ |
D | eni.h | 80 void __iomem *rx_dma; /* RX DMA queue */ member
|
/drivers/infiniband/ulp/isert/ |
D | ib_isert.c | 1112 u64 rx_dma; in isert_rx_completion() local 1116 rx_dma = isert_conn->login_req_dma; in isert_rx_completion() 1119 rx_dma, rx_buflen); in isert_rx_completion() 1121 rx_dma = desc->dma_addr; in isert_rx_completion() 1124 rx_dma, rx_buflen); in isert_rx_completion() 1127 ib_dma_sync_single_for_cpu(ib_dev, rx_dma, rx_buflen, DMA_FROM_DEVICE); in isert_rx_completion() 1140 ib_dma_sync_single_for_device(ib_dev, rx_dma, rx_buflen, in isert_rx_completion()
|
/drivers/net/ethernet/sis/ |
D | sis190.c | 275 dma_addr_t rx_dma; member 888 SIS_W32(RxDescStartAddr, tp->rx_dma); in sis190_hw_start() 1078 tp->RxDescRing = pci_alloc_consistent(pdev, RX_RING_BYTES, &tp->rx_dma); in sis190_open() 1100 pci_free_consistent(pdev, RX_RING_BYTES, tp->RxDescRing, tp->rx_dma); in sis190_open() 1165 pci_free_consistent(pdev, RX_RING_BYTES, tp->RxDescRing, tp->rx_dma); in sis190_close()
|