Home
last modified time | relevance | path

Searched refs:dmadev (Results 1 – 24 of 24) sorted by relevance

/drivers/dma/qcom/
Dhidma.c74 static inline struct hidma_dev *to_hidma_dev(struct dma_device *dmadev) in to_hidma_dev() argument
76 return container_of(dmadev, struct hidma_dev, ddev); in to_hidma_dev()
96 static void hidma_free(struct hidma_dev *dmadev) in hidma_free() argument
98 INIT_LIST_HEAD(&dmadev->ddev.channels); in hidma_free()
168 struct hidma_dev *dmadev = to_hidma_dev(ddev); in hidma_callback() local
187 pm_runtime_mark_last_busy(dmadev->ddev.dev); in hidma_callback()
188 pm_runtime_put_autosuspend(dmadev->ddev.dev); in hidma_callback()
192 static int hidma_chan_init(struct hidma_dev *dmadev, u32 dma_sig) in hidma_chan_init() argument
197 mchan = devm_kzalloc(dmadev->ddev.dev, sizeof(*mchan), GFP_KERNEL); in hidma_chan_init()
201 ddev = &dmadev->ddev; in hidma_chan_init()
[all …]
Dhidma_dbg.c95 struct hidma_dev *dmadev = mchan->dmadev; in hidma_chan_stats() local
97 pm_runtime_get_sync(dmadev->ddev.dev); in hidma_chan_stats()
102 hidma_ll_chstats(s, mchan->dmadev->lldev, mdesc->tre_ch); in hidma_chan_stats()
106 hidma_ll_chstats(s, mchan->dmadev->lldev, mdesc->tre_ch); in hidma_chan_stats()
110 hidma_ll_chstats(s, mchan->dmadev->lldev, mdesc->tre_ch); in hidma_chan_stats()
112 hidma_ll_devstats(s, mchan->dmadev->lldev); in hidma_chan_stats()
113 pm_runtime_mark_last_busy(dmadev->ddev.dev); in hidma_chan_stats()
114 pm_runtime_put_autosuspend(dmadev->ddev.dev); in hidma_chan_stats()
125 struct hidma_dev *dmadev = s->private; in hidma_dma_info() local
128 seq_printf(s, "nr_descriptors=%d\n", dmadev->nr_descriptors); in hidma_dma_info()
[all …]
Dhidma.h100 struct hidma_dev *dmadev; member
156 int hidma_debug_init(struct hidma_dev *dmadev);
157 void hidma_debug_uninit(struct hidma_dev *dmadev);
/drivers/dma/
Ddma-jz4740.c149 static inline uint32_t jz4740_dma_read(struct jz4740_dma_dev *dmadev, in jz4740_dma_read() argument
152 return readl(dmadev->base + reg); in jz4740_dma_read()
155 static inline void jz4740_dma_write(struct jz4740_dma_dev *dmadev, in jz4740_dma_write() argument
158 writel(val, dmadev->base + reg); in jz4740_dma_write()
161 static inline void jz4740_dma_write_mask(struct jz4740_dma_dev *dmadev, in jz4740_dma_write_mask() argument
166 tmp = jz4740_dma_read(dmadev, reg); in jz4740_dma_write_mask()
169 jz4740_dma_write(dmadev, reg, tmp); in jz4740_dma_write_mask()
210 struct jz4740_dma_dev *dmadev = jz4740_dma_chan_get_dev(chan); in jz4740_dma_slave_config() local
260 jz4740_dma_write(dmadev, JZ_REG_DMA_CMD(chan->id), cmd); in jz4740_dma_slave_config()
261 jz4740_dma_write(dmadev, JZ_REG_DMA_STATUS_CTRL(chan->id), 0); in jz4740_dma_slave_config()
[all …]
Dstm32-dma.c205 static u32 stm32_dma_read(struct stm32_dma_device *dmadev, u32 reg) in stm32_dma_read() argument
207 return readl_relaxed(dmadev->base + reg); in stm32_dma_read()
210 static void stm32_dma_write(struct stm32_dma_device *dmadev, u32 reg, u32 val) in stm32_dma_write() argument
212 writel_relaxed(val, dmadev->base + reg); in stm32_dma_write()
284 struct stm32_dma_device *dmadev = stm32_dma_get_dev(chan); in stm32_dma_irq_status() local
296 dma_isr = stm32_dma_read(dmadev, STM32_DMA_HISR); in stm32_dma_irq_status()
298 dma_isr = stm32_dma_read(dmadev, STM32_DMA_LISR); in stm32_dma_irq_status()
307 struct stm32_dma_device *dmadev = stm32_dma_get_dev(chan); in stm32_dma_irq_clear() local
320 stm32_dma_write(dmadev, STM32_DMA_HIFCR, dma_ifcr); in stm32_dma_irq_clear()
322 stm32_dma_write(dmadev, STM32_DMA_LIFCR, dma_ifcr); in stm32_dma_irq_clear()
[all …]
Dsa11x0-dma.c135 static struct sa11x0_dma_dev *to_sa11x0_dma(struct dma_device *dmadev) in to_sa11x0_dma() argument
137 return container_of(dmadev, struct sa11x0_dma_dev, slave); in to_sa11x0_dma()
826 static int sa11x0_dma_init_dmadev(struct dma_device *dmadev, in sa11x0_dma_init_dmadev() argument
831 INIT_LIST_HEAD(&dmadev->channels); in sa11x0_dma_init_dmadev()
832 dmadev->dev = dev; in sa11x0_dma_init_dmadev()
833 dmadev->device_free_chan_resources = sa11x0_dma_free_chan_resources; in sa11x0_dma_init_dmadev()
834 dmadev->device_config = sa11x0_dma_device_config; in sa11x0_dma_init_dmadev()
835 dmadev->device_pause = sa11x0_dma_device_pause; in sa11x0_dma_init_dmadev()
836 dmadev->device_resume = sa11x0_dma_device_resume; in sa11x0_dma_init_dmadev()
837 dmadev->device_terminate_all = sa11x0_dma_device_terminate_all; in sa11x0_dma_init_dmadev()
[all …]
Dvirt-dma.c136 void vchan_init(struct virt_dma_chan *vc, struct dma_device *dmadev) in vchan_init() argument
148 vc->chan.device = dmadev; in vchan_init()
149 list_add_tail(&vc->chan.device_node, &dmadev->channels); in vchan_init()
Dmv_xor_v2.c157 struct dma_device dmadev; member
283 dev_dbg(xor_dev->dmadev.dev, in mv_xor_v2_tx_submit()
358 dev_dbg(xor_dev->dmadev.dev, in mv_xor_v2_prep_dma_memcpy()
414 dev_dbg(xor_dev->dmadev.dev, in mv_xor_v2_prep_dma_xor()
536 dev_dbg(xor_dev->dmadev.dev, "%s %d\n", __func__, __LINE__); in mv_xor_v2_tasklet()
772 dma_dev = &xor_dev->dmadev; in mv_xor_v2_probe()
824 dma_async_device_unregister(&xor_dev->dmadev); in mv_xor_v2_remove()
Ds3c24xx-dma.c1090 struct dma_device *dmadev, unsigned int channels, bool slave) in s3c24xx_dma_init_virtual_channels() argument
1095 INIT_LIST_HEAD(&dmadev->channels); in s3c24xx_dma_init_virtual_channels()
1103 chan = devm_kzalloc(dmadev->dev, sizeof(*chan), GFP_KERNEL); in s3c24xx_dma_init_virtual_channels()
1121 dev_dbg(dmadev->dev, in s3c24xx_dma_init_virtual_channels()
1126 vchan_init(&chan->vc, dmadev); in s3c24xx_dma_init_virtual_channels()
1128 dev_info(dmadev->dev, "initialized %d virtual %s channels\n", in s3c24xx_dma_init_virtual_channels()
1133 static void s3c24xx_dma_free_virtual_channels(struct dma_device *dmadev) in s3c24xx_dma_free_virtual_channels() argument
1139 next, &dmadev->channels, vc.chan.device_node) { in s3c24xx_dma_free_virtual_channels()
Dvirt-dma.h46 void vchan_init(struct virt_dma_chan *vc, struct dma_device *dmadev);
Dmv_xor.c54 ((chan)->dmadev.dev)
1005 struct device *dev = mv_chan->dmadev.dev; in mv_xor_channel_remove()
1007 dma_async_device_unregister(&mv_chan->dmadev); in mv_xor_channel_remove()
1016 list_for_each_entry_safe(chan, _chan, &mv_chan->dmadev.channels, in mv_xor_channel_remove()
1046 dma_dev = &mv_chan->dmadev; in mv_xor_channel_add()
Dmv_xor.h126 struct dma_device dmadev; member
Damba-pl08x.c1888 struct dma_device *dmadev, unsigned int channels, bool slave) in pl08x_dma_init_virtual_channels() argument
1893 INIT_LIST_HEAD(&dmadev->channels); in pl08x_dma_init_virtual_channels()
1931 vchan_init(&chan->vc, dmadev); in pl08x_dma_init_virtual_channels()
1938 static void pl08x_free_virtual_channels(struct dma_device *dmadev) in pl08x_free_virtual_channels() argument
1944 next, &dmadev->channels, vc.chan.device_node) { in pl08x_free_virtual_channels()
Dpxa_dma.c144 #define to_pxad_dev(dmadev) \ argument
145 container_of(dmadev, struct pxad_device, slave)
1286 static void pxad_free_channels(struct dma_device *dmadev) in pxad_free_channels() argument
1290 list_for_each_entry_safe(c, cn, &dmadev->channels, in pxad_free_channels()
Dzx296702_dma.c134 #define to_zx_dma(dmadev) container_of(dmadev, struct zx_dma_dev, slave) argument
Dmmp_pdma.c140 #define to_mmp_pdma_dev(dmadev) \ argument
141 container_of(dmadev, struct mmp_pdma_device, device)
Dk3dma.c119 #define to_k3_dma(dmadev) container_of(dmadev, struct k3_dma_dev, slave) argument
Dfsl-edma.c998 static void fsl_edma_cleanup_vchan(struct dma_device *dmadev) in fsl_edma_cleanup_vchan() argument
1003 &dmadev->channels, vchan.chan.device_node) { in fsl_edma_cleanup_vchan()
Dedma.c2414 static void edma_cleanupp_vchan(struct dma_device *dmadev) in edma_cleanupp_vchan() argument
2419 &dmadev->channels, vchan.chan.device_node) { in edma_cleanupp_vchan()
/drivers/net/ethernet/amazon/ena/
Dena_com.c158 aenq->entries = dma_zalloc_coherent(dev->dmadev, size, &aenq->dma_addr, in ena_com_admin_init_aenq()
342 dev_node = dev_to_node(ena_dev->dmadev); in ena_com_init_io_sq()
343 set_dev_node(ena_dev->dmadev, ctx->numa_node); in ena_com_init_io_sq()
345 dma_zalloc_coherent(ena_dev->dmadev, size, in ena_com_init_io_sq()
348 set_dev_node(ena_dev->dmadev, dev_node); in ena_com_init_io_sq()
351 dma_zalloc_coherent(ena_dev->dmadev, size, in ena_com_init_io_sq()
356 dev_node = dev_to_node(ena_dev->dmadev); in ena_com_init_io_sq()
357 set_dev_node(ena_dev->dmadev, ctx->numa_node); in ena_com_init_io_sq()
359 devm_kzalloc(ena_dev->dmadev, size, GFP_KERNEL); in ena_com_init_io_sq()
360 set_dev_node(ena_dev->dmadev, dev_node); in ena_com_init_io_sq()
[all …]
Dena_com.h313 void *dmadev; member
Dena_netdev.c2912 ena_dev->dmadev = &pdev->dev; in ena_probe()
/drivers/net/wireless/broadcom/brcm80211/brcmsmac/
Ddma.c213 struct device *dmadev; member
451 return dma_alloc_coherent(di->dmadev, size, pap, GFP_ATOMIC); in dma_alloc_consistent()
486 dma_free_coherent(di->dmadev, size, va, *descpa); in dma_ringalloc()
590 di->dmadev = core->dma_dev; in dma_attach()
755 dma_free_coherent(di->dmadev, di->txdalloc, in dma_detach()
759 dma_free_coherent(di->dmadev, di->rxdalloc, in dma_detach()
897 dma_unmap_single(di->dmadev, pa, di->rxbufsize, DMA_FROM_DEVICE); in dma64_getnextrxp()
1080 pa = dma_map_single(di->dmadev, p->data, di->rxbufsize, in dma_rxfill()
1082 if (dma_mapping_error(di->dmadev, pa)) { in dma_rxfill()
1291 pa = dma_map_single(di->dmadev, data, len, DMA_TO_DEVICE); in dma_txenq()
[all …]
/drivers/media/platform/
Dm2m-deinterlace.c230 struct dma_device *dmadev = chan->device; in deinterlace_issue_dma() local
349 tx = dmadev->device_prep_interleaved_dma(chan, ctx->xt, flags); in deinterlace_issue_dma()