/kernel/linux/linux-5.10/drivers/staging/ralink-gdma/ |
D | ralink-gdma.c | 134 void (*init)(struct gdma_dma_dev *dma_dev); 155 static inline uint32_t gdma_dma_read(struct gdma_dma_dev *dma_dev, in gdma_dma_read() argument 158 return readl(dma_dev->base + reg); in gdma_dma_read() 161 static inline void gdma_dma_write(struct gdma_dma_dev *dma_dev, in gdma_dma_write() argument 164 writel(val, dma_dev->base + reg); in gdma_dma_write() 185 struct gdma_dma_dev *dma_dev = gdma_dma_chan_get_dev(chan); in gdma_dma_config() local 188 dev_err(dma_dev->ddev.dev, "not support flow controller\n"); in gdma_dma_config() 195 dev_err(dma_dev->ddev.dev, "only support 4 byte buswidth\n"); in gdma_dma_config() 204 dev_err(dma_dev->ddev.dev, "only support 4 byte buswidth\n"); in gdma_dma_config() 212 dev_err(dma_dev->ddev.dev, "direction type %d error\n", in gdma_dma_config() [all …]
|
/kernel/linux/linux-5.10/drivers/dma/ |
D | mcf-edma.c | 222 INIT_LIST_HEAD(&mcf_edma->dma_dev.channels); in mcf_edma_probe() 231 vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev); in mcf_edma_probe() 242 dma_cap_set(DMA_PRIVATE, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe() 243 dma_cap_set(DMA_SLAVE, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe() 244 dma_cap_set(DMA_CYCLIC, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe() 246 mcf_edma->dma_dev.dev = &pdev->dev; in mcf_edma_probe() 247 mcf_edma->dma_dev.device_alloc_chan_resources = in mcf_edma_probe() 249 mcf_edma->dma_dev.device_free_chan_resources = in mcf_edma_probe() 251 mcf_edma->dma_dev.device_config = fsl_edma_slave_config; in mcf_edma_probe() 252 mcf_edma->dma_dev.device_prep_dma_cyclic = in mcf_edma_probe() [all …]
|
D | fsl-edma.c | 116 list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels, device_node) { in fsl_edma_xlate() 351 INIT_LIST_HEAD(&fsl_edma->dma_dev.channels); in fsl_edma_probe() 361 vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev); in fsl_edma_probe() 372 dma_cap_set(DMA_PRIVATE, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe() 373 dma_cap_set(DMA_SLAVE, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe() 374 dma_cap_set(DMA_CYCLIC, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe() 376 fsl_edma->dma_dev.dev = &pdev->dev; in fsl_edma_probe() 377 fsl_edma->dma_dev.device_alloc_chan_resources in fsl_edma_probe() 379 fsl_edma->dma_dev.device_free_chan_resources in fsl_edma_probe() 381 fsl_edma->dma_dev.device_tx_status = fsl_edma_tx_status; in fsl_edma_probe() [all …]
|
D | dma-axi-dmac.c | 146 struct dma_device dma_dev; member 153 dma_dev); in chan_to_axi_dmac() 320 dev_dbg(dmac->dma_dev.dev, in axi_dmac_dequeue_partial_xfers() 324 dev_warn(dmac->dma_dev.dev, in axi_dmac_dequeue_partial_xfers() 878 dev_err(dmac->dma_dev.dev, in axi_dmac_detect_caps() 886 dev_err(dmac->dma_dev.dev, in axi_dmac_detect_caps() 907 struct dma_device *dma_dev; in axi_dmac_probe() local 951 dma_dev = &dmac->dma_dev; in axi_dmac_probe() 952 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in axi_dmac_probe() 953 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in axi_dmac_probe() [all …]
|
D | altera-msgdma.c | 796 struct dma_device *dma_dev; in msgdma_probe() local 844 dma_dev = &mdev->dmadev; in msgdma_probe() 847 dma_cap_zero(dma_dev->cap_mask); in msgdma_probe() 848 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in msgdma_probe() 849 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in msgdma_probe() 851 dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe() 852 dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe() 853 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) | in msgdma_probe() 855 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in msgdma_probe() 858 INIT_LIST_HEAD(&dma_dev->channels); in msgdma_probe() [all …]
|
D | img-mdc-dma.c | 134 struct dma_device dma_dev; member 183 return mdma->dma_dev.dev; in mdma2dev() 817 list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) { in mdc_of_xlate() 915 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe() 916 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 917 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 918 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe() 919 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe() 949 mdma->dma_dev.dev = &pdev->dev; in mdc_dma_probe() 950 mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg; in mdc_dma_probe() [all …]
|
D | nbpfaxi.c | 230 struct dma_device dma_dev; member 326 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_read() 335 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_write() 401 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): next 0x%x, cur 0x%x\n", __func__, in nbpf_start() 857 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): force-free desc %p cookie %d\n", in nbpf_chan_idle() 1098 dchan = dma_get_any_slave_channel(&nbpf->dma_dev); in nbpf_of_xlate() 1229 dev_warn(nbpf->dma_dev.dev, "DMA error IRQ %u\n", irq); in nbpf_err_irq() 1247 struct dma_device *dma_dev = &nbpf->dma_dev; in nbpf_chan_probe() local 1255 chan->dma_chan.device = dma_dev; in nbpf_chan_probe() 1259 dev_dbg(dma_dev->dev, "%s(): channel %d: -> %p\n", __func__, n, chan->base); in nbpf_chan_probe() [all …]
|
D | fsl_raid.c | 628 struct dma_device *dma_dev; in fsl_re_chan_probe() local 636 dma_dev = &re_priv->dma_dev; in fsl_re_chan_probe() 682 chan->chan.device = dma_dev; in fsl_re_chan_probe() 752 struct dma_device *dma_dev; in fsl_re_probe() local 781 dma_dev = &re_priv->dma_dev; in fsl_re_probe() 782 dma_dev->dev = dev; in fsl_re_probe() 783 INIT_LIST_HEAD(&dma_dev->channels); in fsl_re_probe() 786 dma_dev->device_alloc_chan_resources = fsl_re_alloc_chan_resources; in fsl_re_probe() 787 dma_dev->device_tx_status = fsl_re_tx_status; in fsl_re_probe() 788 dma_dev->device_issue_pending = fsl_re_issue_pending; in fsl_re_probe() [all …]
|
D | iop-adma.c | 1271 struct dma_device *dma_dev; in iop_adma_probe() local 1285 dma_dev = &adev->common; in iop_adma_probe() 1306 dma_dev->cap_mask = plat_data->cap_mask; in iop_adma_probe() 1311 INIT_LIST_HEAD(&dma_dev->channels); in iop_adma_probe() 1314 dma_dev->device_alloc_chan_resources = iop_adma_alloc_chan_resources; in iop_adma_probe() 1315 dma_dev->device_free_chan_resources = iop_adma_free_chan_resources; in iop_adma_probe() 1316 dma_dev->device_tx_status = iop_adma_status; in iop_adma_probe() 1317 dma_dev->device_issue_pending = iop_adma_issue_pending; in iop_adma_probe() 1318 dma_dev->dev = &pdev->dev; in iop_adma_probe() 1321 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in iop_adma_probe() [all …]
|
D | tegra210-adma.c | 153 struct dma_device dma_dev; member 703 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate() 852 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_adma_probe() 865 vchan_init(&tdc->vc, &tdma->dma_dev); in tegra_adma_probe() 882 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_adma_probe() 883 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_adma_probe() 884 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_adma_probe() 886 tdma->dma_dev.dev = &pdev->dev; in tegra_adma_probe() 887 tdma->dma_dev.device_alloc_chan_resources = in tegra_adma_probe() 889 tdma->dma_dev.device_free_chan_resources = in tegra_adma_probe() [all …]
|
D | sprd-dma.c | 209 struct dma_device dma_dev; member 358 dev_warn(sdev->dma_dev.dev, in sprd_dma_pause_resume() 422 dev_warn(sdev->dma_dev.dev, "incorrect dma interrupt type\n"); in sprd_dma_get_int_type() 481 dev_err(sdev->dma_dev.dev, "invalid channel mode setting %d\n", in sprd_dma_set_2stage_config() 764 dev_err(sdev->dma_dev.dev, "invalid source step\n"); in sprd_dma_fill_desc() 780 dev_err(sdev->dma_dev.dev, "invalid destination step\n"); in sprd_dma_fill_desc() 788 dev_err(sdev->dma_dev.dev, "invalid source datawidth\n"); in sprd_dma_fill_desc() 794 dev_err(sdev->dma_dev.dev, "invalid destination datawidth\n"); in sprd_dma_fill_desc() 1169 dma_cap_set(DMA_MEMCPY, sdev->dma_dev.cap_mask); in sprd_dma_probe() 1171 sdev->dma_dev.chancnt = chn_count; in sprd_dma_probe() [all …]
|
D | ep93xx_dma.c | 204 struct dma_device dma_dev; member 1322 struct dma_device *dma_dev; in ep93xx_dma_probe() local 1331 dma_dev = &edma->dma_dev; in ep93xx_dma_probe() 1335 INIT_LIST_HEAD(&dma_dev->channels); in ep93xx_dma_probe() 1340 edmac->chan.device = dma_dev; in ep93xx_dma_probe() 1359 &dma_dev->channels); in ep93xx_dma_probe() 1362 dma_cap_zero(dma_dev->cap_mask); in ep93xx_dma_probe() 1363 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in ep93xx_dma_probe() 1364 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in ep93xx_dma_probe() 1366 dma_dev->dev = &pdev->dev; in ep93xx_dma_probe() [all …]
|
D | hisi_dma.c | 105 struct dma_device dma_dev; member 416 vchan_init(&hdma_dev->chan[i].vc, &hdma_dev->dma_dev); in hisi_dma_enable_qps() 513 struct dma_device *dma_dev; in hisi_dma_probe() local 559 dma_dev = &hdma_dev->dma_dev; in hisi_dma_probe() 560 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in hisi_dma_probe() 561 dma_dev->device_free_chan_resources = hisi_dma_free_chan_resources; in hisi_dma_probe() 562 dma_dev->device_prep_dma_memcpy = hisi_dma_prep_dma_memcpy; in hisi_dma_probe() 563 dma_dev->device_tx_status = hisi_dma_tx_status; in hisi_dma_probe() 564 dma_dev->device_issue_pending = hisi_dma_issue_pending; in hisi_dma_probe() 565 dma_dev->device_terminate_all = hisi_dma_terminate_all; in hisi_dma_probe() [all …]
|
D | mv_xor_v2.c | 719 struct dma_device *dma_dev; in mv_xor_v2_probe() local 834 dma_dev = &xor_dev->dmadev; in mv_xor_v2_probe() 837 dma_cap_zero(dma_dev->cap_mask); in mv_xor_v2_probe() 838 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in mv_xor_v2_probe() 839 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in mv_xor_v2_probe() 840 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in mv_xor_v2_probe() 843 INIT_LIST_HEAD(&dma_dev->channels); in mv_xor_v2_probe() 846 dma_dev->device_tx_status = dma_cookie_status; in mv_xor_v2_probe() 847 dma_dev->device_issue_pending = mv_xor_v2_issue_pending; in mv_xor_v2_probe() 848 dma_dev->dev = &pdev->dev; in mv_xor_v2_probe() [all …]
|
D | xgene-dma.c | 310 struct dma_device dma_dev[XGENE_DMA_MAX_CHANNEL]; member 1485 struct dma_device *dma_dev) in xgene_dma_set_caps() argument 1488 dma_cap_zero(dma_dev->cap_mask); in xgene_dma_set_caps() 1503 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in xgene_dma_set_caps() 1504 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps() 1507 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps() 1511 dma_dev->dev = chan->dev; in xgene_dma_set_caps() 1512 dma_dev->device_alloc_chan_resources = xgene_dma_alloc_chan_resources; in xgene_dma_set_caps() 1513 dma_dev->device_free_chan_resources = xgene_dma_free_chan_resources; in xgene_dma_set_caps() 1514 dma_dev->device_issue_pending = xgene_dma_issue_pending; in xgene_dma_set_caps() [all …]
|
/kernel/linux/linux-5.10/drivers/spi/ |
D | spi-pxa2xx-pci.c | 70 if (dws->dma_dev != chan->device->dev) in lpss_dma_filter() 77 static void lpss_dma_put_device(void *dma_dev) in lpss_dma_put_device() argument 79 pci_dev_put(dma_dev); in lpss_dma_put_device() 84 struct pci_dev *dma_dev; in lpss_spi_setup() local 90 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(PCI_SLOT(dev->devfn), 0)); in lpss_spi_setup() 91 ret = devm_add_action_or_reset(&dev->dev, lpss_dma_put_device, dma_dev); in lpss_spi_setup() 98 slave->dma_dev = &dma_dev->dev; in lpss_spi_setup() 106 slave->dma_dev = &dma_dev->dev; in lpss_spi_setup() 118 struct pci_dev *dma_dev; in mrfld_spi_setup() local 144 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(21, 0)); in mrfld_spi_setup() [all …]
|
/kernel/linux/linux-5.10/drivers/crypto/ccp/ |
D | ccp-dmaengine.c | 657 struct dma_device *dma_dev = &ccp->dma_dev; in ccp_dmaengine_register() local 703 dma_dev->dev = ccp->dev; in ccp_dmaengine_register() 704 dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register() 705 dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register() 706 dma_dev->directions = DMA_MEM_TO_MEM; in ccp_dmaengine_register() 707 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in ccp_dmaengine_register() 708 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ccp_dmaengine_register() 709 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in ccp_dmaengine_register() 718 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ccp_dmaengine_register() 720 INIT_LIST_HEAD(&dma_dev->channels); in ccp_dmaengine_register() [all …]
|
/kernel/linux/linux-5.10/drivers/mmc/host/ |
D | mmc_spi.c | 123 struct device *dma_dev; member 162 if (host->dma_dev) in mmc_spi_readbytes() 163 dma_sync_single_for_device(host->dma_dev, in mmc_spi_readbytes() 169 if (host->dma_dev) in mmc_spi_readbytes() 170 dma_sync_single_for_cpu(host->dma_dev, in mmc_spi_readbytes() 521 if (host->dma_dev) { in mmc_spi_command_send() 523 dma_sync_single_for_device(host->dma_dev, in mmc_spi_command_send() 529 if (host->dma_dev) in mmc_spi_command_send() 530 dma_sync_single_for_cpu(host->dma_dev, in mmc_spi_command_send() 663 if (host->dma_dev) in mmc_spi_writeblock() [all …]
|
/kernel/linux/linux-5.10/drivers/dma/sf-pdma/ |
D | sf-pdma.c | 90 dev_err(chan->pdma->dma_dev.dev, in sf_pdma_prep_dma_memcpy() 258 dev_err(chan->pdma->dma_dev.dev, "NULL desc.\n"); in sf_pdma_xfer_desc() 460 INIT_LIST_HEAD(&pdma->dma_dev.channels); in sf_pdma_setup_chans() 491 vchan_init(&chan->vchan, &pdma->dma_dev); in sf_pdma_setup_chans() 532 pdma->dma_dev.dev = &pdev->dev; in sf_pdma_probe() 535 dma_cap_set(DMA_MEMCPY, pdma->dma_dev.cap_mask); in sf_pdma_probe() 536 pdma->dma_dev.copy_align = 2; in sf_pdma_probe() 537 pdma->dma_dev.src_addr_widths = widths; in sf_pdma_probe() 538 pdma->dma_dev.dst_addr_widths = widths; in sf_pdma_probe() 539 pdma->dma_dev.directions = BIT(DMA_MEM_TO_MEM); in sf_pdma_probe() [all …]
|
/kernel/linux/linux-5.10/drivers/tty/serial/8250/ |
D | 8250_mid.c | 45 struct pci_dev *dma_dev; member 71 mid->dma_dev = pci_get_slot(pdev->bus, in pnw_setup() 78 pci_dev_put(mid->dma_dev); in pnw_exit() 90 chip = pci_get_drvdata(mid->dma_dev); in tng_handle_irq() 126 mid->dma_dev = pci_get_slot(pdev->bus, PCI_DEVFN(5, 0)); in tng_setup() 134 pci_dev_put(mid->dma_dev); in tng_exit() 194 mid->dma_dev = pdev; in dnv_setup() 202 if (!mid->dma_dev) in dnv_exit() 249 if (s->dma_dev != chan->device->dev || s->chan_id != chan->chan_id) in mid8250_dma_filter() 263 if (!mid->dma_dev) in mid8250_dma_setup() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/exynos/ |
D | exynos_drm_dma.c | 49 if (get_dma_ops(priv->dma_dev) != get_dma_ops(subdrv_dev)) { in drm_iommu_attach_device() 101 if (!priv->dma_dev) { in exynos_drm_register_dma() 102 priv->dma_dev = dev; in exynos_drm_register_dma() 117 mapping = iommu_get_domain_for_dev(priv->dma_dev); in exynos_drm_register_dma() 145 priv->dma_dev = NULL; in exynos_drm_cleanup_dma()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/broadcom/ |
D | bgmac.c | 135 struct device *dma_dev = bgmac->dma_dev; in bgmac_dma_tx_add() local 162 slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb), in bgmac_dma_tx_add() 164 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) in bgmac_dma_tx_add() 180 slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0, in bgmac_dma_tx_add() 182 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) in bgmac_dma_tx_add() 211 dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb), in bgmac_dma_tx_add() 220 dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE); in bgmac_dma_tx_add() 237 struct device *dma_dev = bgmac->dma_dev; in bgmac_dma_tx_free() local 262 dma_unmap_single(dma_dev, slot->dma_addr, len, in bgmac_dma_tx_free() 265 dma_unmap_page(dma_dev, slot->dma_addr, len, in bgmac_dma_tx_free() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/mediatek/ |
D | mtk_drm_drv.c | 171 struct device *dma_dev; in mtk_drm_kms_init() local 237 dma_dev = &pdev->dev; in mtk_drm_kms_init() 238 private->dma_dev = dma_dev; in mtk_drm_kms_init() 244 if (!dma_dev->dma_parms) { in mtk_drm_kms_init() 246 dma_dev->dma_parms = in mtk_drm_kms_init() 247 devm_kzalloc(drm->dev, sizeof(*dma_dev->dma_parms), in mtk_drm_kms_init() 250 if (!dma_dev->dma_parms) { in mtk_drm_kms_init() 255 ret = dma_set_max_seg_size(dma_dev, (unsigned int)DMA_BIT_MASK(32)); in mtk_drm_kms_init() 257 dev_err(dma_dev, "Failed to set DMA segment size\n"); in mtk_drm_kms_init() 278 dma_dev->dma_parms = NULL; in mtk_drm_kms_init() [all …]
|
/kernel/linux/linux-5.10/sound/soc/samsung/ |
D | dmaengine.c | 19 struct device *dma_dev) in samsung_asoc_dma_platform_register() argument 29 pcm_conf->dma_dev = dma_dev; in samsung_asoc_dma_platform_register()
|
/kernel/linux/linux-5.10/drivers/dma/sh/ |
D | shdmac.c | 165 dev_warn(shdev->shdma_dev.dma_dev.dev, "Can't initialize DMAOR.\n"); in sh_dmae_rst() 169 dev_warn(shdev->shdma_dev.dma_dev.dev, in sh_dmae_rst() 522 struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev); in sh_dmae_chan_probe() 527 sh_chan = devm_kzalloc(sdev->dma_dev.dev, sizeof(struct sh_dmae_chan), in sh_dmae_chan_probe() 549 dev_err(sdev->dma_dev.dev, in sh_dmae_chan_probe() 680 struct dma_device *dma_dev; in sh_dmae_probe() local 720 dma_dev = &shdev->shdma_dev.dma_dev; in sh_dmae_probe() 731 dma_dev->src_addr_widths = widths; in sh_dmae_probe() 732 dma_dev->dst_addr_widths = widths; in sh_dmae_probe() 733 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM); in sh_dmae_probe() [all …]
|