Home
last modified time | relevance | path

Searched refs:rx_dma (Results 1 – 25 of 36) sorted by relevance

12

/drivers/spi/
Dspi-s3c64xx.c200 struct s3c64xx_spi_dma_data rx_dma; member
266 struct s3c64xx_spi_driver_data, rx_dma); in s3c64xx_spi_dmacb()
302 struct s3c64xx_spi_driver_data, rx_dma); in prepare_dma()
303 config.direction = sdd->rx_dma.direction; in prepare_dma()
306 sdd->ops->config((enum dma_ch)sdd->rx_dma.ch, &config); in prepare_dma()
337 sdd->rx_dma.ch = (void *)sdd->ops->request(sdd->rx_dma.dmach, &req, dev, "rx"); in acquire_dma()
361 sdd->ops->release((enum dma_ch)sdd->rx_dma.ch, &s3c64xx_spi_dma_client); in s3c64xx_spi_unprepare_transfer()
386 struct s3c64xx_spi_driver_data, rx_dma); in prepare_dma()
430 sdd->rx_dma.ch = dma_request_slave_channel_compat(mask, filter, in s3c64xx_spi_prepare_transfer()
431 (void*)sdd->rx_dma.dmach, dev, "rx"); in s3c64xx_spi_prepare_transfer()
[all …]
Dspi-fsl-cpm.c78 if (mspi->rx_dma == mspi->dma_dummy_rx) in fsl_spi_cpm_bufs_start()
79 out_be32(&rx_bd->cbd_bufaddr, mspi->rx_dma); in fsl_spi_cpm_bufs_start()
81 out_be32(&rx_bd->cbd_bufaddr, mspi->rx_dma + xfer_ofs); in fsl_spi_cpm_bufs_start()
117 mspi->rx_dma = mspi->dma_dummy_rx; in fsl_spi_cpm_bufs()
135 mspi->rx_dma = dma_map_single(dev, mspi->rx, t->len, in fsl_spi_cpm_bufs()
137 if (dma_mapping_error(dev, mspi->rx_dma)) { in fsl_spi_cpm_bufs()
142 mspi->rx_dma = t->rx_dma; in fsl_spi_cpm_bufs()
170 dma_unmap_single(dev, mspi->rx_dma, t->len, DMA_FROM_DEVICE); in fsl_spi_cpm_bufs_complete()
Dspi-atmel.c574 as->dma.sgrx.dma_address = xfer->rx_dma + xfer->len - *plen; in atmel_spi_next_xfer_dma_submit()
622 xfer->rx_buf, xfer->rx_dma); in atmel_spi_next_xfer_dma_submit()
656 dma_addr_t *rx_dma, in atmel_spi_next_xfer_data() argument
664 *rx_dma = xfer->rx_dma + xfer->len - *plen; in atmel_spi_next_xfer_data()
666 *rx_dma = as->buffer_dma; in atmel_spi_next_xfer_data()
696 dma_addr_t tx_dma, rx_dma; in atmel_spi_pdc_next_xfer() local
711 atmel_spi_next_xfer_data(master, xfer, &tx_dma, &rx_dma, &len); in atmel_spi_pdc_next_xfer()
714 spi_writel(as, RPR, rx_dma); in atmel_spi_pdc_next_xfer()
725 xfer->rx_buf, xfer->rx_dma); in atmel_spi_pdc_next_xfer()
750 atmel_spi_next_xfer_data(master, xfer, &tx_dma, &rx_dma, &len); in atmel_spi_pdc_next_xfer()
[all …]
Dspi-pxa2xx-pxadma.c61 return drv_data->rx_dma && drv_data->tx_dma; in pxa2xx_spi_map_dma_buffers()
93 drv_data->rx_dma = dma_map_single(dev, drv_data->rx, in pxa2xx_spi_map_dma_buffers()
95 if (dma_mapping_error(dev, drv_data->rx_dma)) { in pxa2xx_spi_map_dma_buffers()
113 dma_unmap_single(dev, drv_data->rx_dma, in pxa2xx_spi_unmap_dma_buffers()
296 DTADR(drv_data->rx_channel) = drv_data->rx_dma; in pxa2xx_spi_dma_prepare()
Dspi-dw-mid.c124 if (dws->rx_dma) in mid_spi_dma_transfer()
169 dws->rx_sgl.dma_address = dws->rx_dma; in mid_spi_dma_transfer()
Dspi-davinci.c598 t->rx_dma = dma_map_single(&spi->dev, buf, in davinci_spi_bufs()
600 if (!t->rx_dma) { in davinci_spi_bufs()
604 sg_dma_address(&sg_rx) = t->rx_dma; in davinci_spi_bufs()
666 dma_unmap_single(&spi->dev, t->rx_dma, in davinci_spi_bufs()
697 dma_unmap_single(&spi->dev, t->rx_dma, t->len, DMA_FROM_DEVICE); in davinci_spi_bufs()
Dspi-dw.c256 if (dws->cur_transfer->rx_dma) in map_dma_buffers()
257 dws->rx_dma = dws->cur_transfer->rx_dma; in map_dma_buffers()
424 dws->rx_dma = transfer->rx_dma; in pump_transfers()
Dspi-fsl-lib.h46 dma_addr_t rx_dma; member
Dspi-pxa2xx.h80 dma_addr_t rx_dma; member
Dspi-dw.h131 dma_addr_t rx_dma; member
Dspi-omap2-mcspi.c396 sg_dma_address(&sg) = xfer->rx_dma; in omap2_mcspi_rx_dma()
415 dma_unmap_single(mcspi->dev, xfer->rx_dma, count, in omap2_mcspi_rx_dma()
1107 t->rx_dma = dma_map_single(mcspi->dev, rx_buf, t->len, in omap2_mcspi_transfer_one_message()
1109 if (dma_mapping_error(mcspi->dev, t->rx_dma)) { in omap2_mcspi_transfer_one_message()
Dspi-au1550.c366 dma_rx_addr = t->rx_dma; in au1550_spi_dma_txrxb()
387 if (t->rx_dma == 0) { /* if DMA_ADDR_INVALID, map it */ in au1550_spi_dma_txrxb()
449 if (t->rx_buf && t->rx_dma == 0 ) in au1550_spi_dma_txrxb()
/drivers/net/ethernet/amd/
Dau1000_eth.h64 struct rx_dma { struct
91 struct rx_dma *rx_dma_ring[NUM_RX_DMA];
Dau1000_eth.c549 aup->rx_dma_ring[i] = (struct rx_dma *) in au1000_setup_hw_rings()
550 (tx_base + 0x100 + sizeof(struct rx_dma) * i); in au1000_setup_hw_rings()
710 struct rx_dma *prxd; in au1000_rx()
/drivers/infiniband/ulp/iser/
Diser_initiator.c454 u64 rx_dma; in iser_rcv_completion() local
459 rx_dma = ib_conn->login_resp_dma; in iser_rcv_completion()
462 rx_dma = rx_desc->dma_addr; in iser_rcv_completion()
466 ib_dma_sync_single_for_cpu(ib_conn->device->ib_device, rx_dma, in iser_rcv_completion()
477 ib_dma_sync_single_for_device(ib_conn->device->ib_device, rx_dma, in iser_rcv_completion()
486 if (rx_dma == ib_conn->login_resp_dma) in iser_rcv_completion()
/drivers/net/ethernet/
Ds6gmac.c349 u32 rx_dma; member
378 (!s6dmac_fifo_full(pd->rx_dma, pd->rx_chan)) && in s6gmac_rx_fillfifo()
381 s6dmac_put_fifo_cache(pd->rx_dma, pd->rx_chan, in s6gmac_rx_fillfifo()
392 s6dmac_pending_count(pd->rx_dma, pd->rx_chan)) { in s6gmac_rx_interrupt()
516 if (s6dmac_termcnt_irq(pd->rx_dma, pd->rx_chan)) in s6gmac_interrupt()
847 s6dmac_disable_chan(pd->rx_dma, pd->rx_chan); in s6gmac_init_dmac()
849 s6dmac_disable_error_irqs(pd->rx_dma, 1 << S6_HIFDMA_GMACRX); in s6gmac_init_dmac()
899 s6dmac_enable_chan(pd->rx_dma, pd->rx_chan, in s6gmac_open()
984 pd->rx_dma = DMA_MASK_DMAC(i); in s6gmac_probe()
/drivers/net/ethernet/apple/
Dbmac.c67 volatile struct dbdma_regs __iomem *rx_dma; member
229 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_enable_and_reset_chip()
406 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_start_chip()
477 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_suspend()
613 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_init_rx_ring()
683 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_rxdma_intr()
1318 bp->rx_dma = ioremap(macio_resource_start(mdev, 2), macio_resource_len(mdev, 2)); in bmac_probe()
1319 if (!bp->rx_dma) in bmac_probe()
1372 iounmap(bp->rx_dma); in bmac_probe()
1400 volatile struct dbdma_regs __iomem *rd = bp->rx_dma; in bmac_close()
[all …]
Dmace.c48 volatile struct dbdma_regs __iomem *rx_dma; member
187 mp->rx_dma = ioremap(macio_resource_start(mdev, 2), 0x1000); in mace_probe()
188 if (mp->rx_dma == NULL) { in mace_probe()
261 iounmap(mp->rx_dma); in mace_probe()
291 iounmap(mp->rx_dma); in mace_remove()
432 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_open()
502 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_close()
811 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_tx_timeout()
880 volatile struct dbdma_regs __iomem *rd = mp->rx_dma; in mace_rxdma_intr()
/drivers/mmc/host/
Dmmc_spi.c534 t->tx_dma = t->rx_dma = host->data_dma; in mmc_spi_command_send()
626 t->rx_dma = dma + offsetof(struct scratch, crc_val); in mmc_spi_setup_data_message()
654 t->rx_dma = dma + offsetof(struct scratch, status); in mmc_spi_setup_data_message()
822 t->rx_dma, t->len, in mmc_spi_readblock()
833 t->rx_dma, t->len, in mmc_spi_readblock()
872 t->rx_dma += t->len; in mmc_spi_readblock()
936 t->rx_dma = dma_addr + sg->offset; in mmc_spi_data_do()
1014 host->early_status.tx_dma = host->early_status.rx_dma; in mmc_spi_data_do()
1440 host->status.rx_dma = host->data_dma + offsetof(struct scratch, status); in mmc_spi_probe()
/drivers/staging/et131x/
Det131x.c1787 struct rxdma_regs __iomem *rx_dma = &adapter->regs->rxdma; in et131x_config_rx_dma_regs() local
1799 writel(upper_32_bits(rx_local->rx_status_bus), &rx_dma->dma_wb_base_hi); in et131x_config_rx_dma_regs()
1800 writel(lower_32_bits(rx_local->rx_status_bus), &rx_dma->dma_wb_base_lo); in et131x_config_rx_dma_regs()
1807 writel(upper_32_bits(rx_local->ps_ring_physaddr), &rx_dma->psr_base_hi); in et131x_config_rx_dma_regs()
1808 writel(lower_32_bits(rx_local->ps_ring_physaddr), &rx_dma->psr_base_lo); in et131x_config_rx_dma_regs()
1809 writel(rx_local->psr_num_entries - 1, &rx_dma->psr_num_des); in et131x_config_rx_dma_regs()
1810 writel(0, &rx_dma->psr_full_offset); in et131x_config_rx_dma_regs()
1812 psr_num_des = readl(&rx_dma->psr_num_des) & ET_RXDMA_PSR_NUM_DES_MASK; in et131x_config_rx_dma_regs()
1814 &rx_dma->psr_min_des); in et131x_config_rx_dma_regs()
1829 num_des = &rx_dma->fbr0_num_des; in et131x_config_rx_dma_regs()
[all …]
/drivers/tty/serial/
Difx6x60.h87 dma_addr_t rx_dma; member
Difx6x60.c775 ifx_dev->rx_dma = ifx_dev->rx_bus; in ifx_spi_io()
777 ifx_dev->spi_xfer.rx_dma = ifx_dev->rx_dma; in ifx_spi_io()
781 ifx_dev->rx_dma = (dma_addr_t)0; in ifx_spi_io()
783 ifx_dev->spi_xfer.rx_dma = (dma_addr_t)0; in ifx_spi_io()
/drivers/atm/
Deni.h80 void __iomem *rx_dma; /* RX DMA queue */ member
/drivers/infiniband/ulp/isert/
Dib_isert.c1112 u64 rx_dma; in isert_rx_completion() local
1116 rx_dma = isert_conn->login_req_dma; in isert_rx_completion()
1119 rx_dma, rx_buflen); in isert_rx_completion()
1121 rx_dma = desc->dma_addr; in isert_rx_completion()
1124 rx_dma, rx_buflen); in isert_rx_completion()
1127 ib_dma_sync_single_for_cpu(ib_dev, rx_dma, rx_buflen, DMA_FROM_DEVICE); in isert_rx_completion()
1140 ib_dma_sync_single_for_device(ib_dev, rx_dma, rx_buflen, in isert_rx_completion()
/drivers/net/ethernet/sis/
Dsis190.c275 dma_addr_t rx_dma; member
888 SIS_W32(RxDescStartAddr, tp->rx_dma); in sis190_hw_start()
1078 tp->RxDescRing = pci_alloc_consistent(pdev, RX_RING_BYTES, &tp->rx_dma); in sis190_open()
1100 pci_free_consistent(pdev, RX_RING_BYTES, tp->RxDescRing, tp->rx_dma); in sis190_open()
1165 pci_free_consistent(pdev, RX_RING_BYTES, tp->RxDescRing, tp->rx_dma); in sis190_close()

12