Home
last modified time | relevance | path

Searched refs:dma_dev (Results 1 – 25 of 87) sorted by relevance

1234

/drivers/dma/
Ddma-axi-dmac.c120 struct dma_device dma_dev; member
129 dma_dev); in chan_to_axi_dmac()
575 struct dma_device *dma_dev; in axi_dmac_probe() local
618 dma_dev = &dmac->dma_dev; in axi_dmac_probe()
619 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in axi_dmac_probe()
620 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in axi_dmac_probe()
621 dma_dev->device_free_chan_resources = axi_dmac_free_chan_resources; in axi_dmac_probe()
622 dma_dev->device_tx_status = dma_cookie_status; in axi_dmac_probe()
623 dma_dev->device_issue_pending = axi_dmac_issue_pending; in axi_dmac_probe()
624 dma_dev->device_prep_slave_sg = axi_dmac_prep_slave_sg; in axi_dmac_probe()
[all …]
Daltera-msgdma.c798 struct dma_device *dma_dev; in msgdma_probe() local
846 dma_dev = &mdev->dmadev; in msgdma_probe()
849 dma_cap_zero(dma_dev->cap_mask); in msgdma_probe()
850 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in msgdma_probe()
851 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in msgdma_probe()
853 dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe()
854 dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe()
855 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) | in msgdma_probe()
857 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in msgdma_probe()
860 INIT_LIST_HEAD(&dma_dev->channels); in msgdma_probe()
[all …]
Dtegra210-adma.c142 struct dma_device dma_dev; member
627 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate()
747 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_adma_probe()
759 vchan_init(&tdc->vc, &tdma->dma_dev); in tegra_adma_probe()
774 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_adma_probe()
775 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_adma_probe()
776 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_adma_probe()
778 tdma->dma_dev.dev = &pdev->dev; in tegra_adma_probe()
779 tdma->dma_dev.device_alloc_chan_resources = in tegra_adma_probe()
781 tdma->dma_dev.device_free_chan_resources = in tegra_adma_probe()
[all …]
Dnbpfaxi.c226 struct dma_device dma_dev; member
322 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_read()
331 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_write()
397 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): next 0x%x, cur 0x%x\n", __func__, in nbpf_start()
852 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): force-free desc %p cookie %d\n", in nbpf_chan_idle()
1093 dchan = dma_get_any_slave_channel(&nbpf->dma_dev); in nbpf_of_xlate()
1224 dev_warn(nbpf->dma_dev.dev, "DMA error IRQ %u\n", irq); in nbpf_err_irq()
1242 struct dma_device *dma_dev = &nbpf->dma_dev; in nbpf_chan_probe() local
1250 chan->dma_chan.device = dma_dev; in nbpf_chan_probe()
1254 dev_dbg(dma_dev->dev, "%s(): channel %d: -> %p\n", __func__, n, chan->base); in nbpf_chan_probe()
[all …]
Dimg-mdc-dma.c136 struct dma_device dma_dev; member
185 return mdma->dma_dev.dev; in mdma2dev()
804 list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) { in mdc_of_xlate()
890 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe()
891 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe()
892 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe()
893 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe()
894 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe()
924 mdma->dma_dev.dev = &pdev->dev; in mdc_dma_probe()
925 mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg; in mdc_dma_probe()
[all …]
Dfsl_raid.c630 struct dma_device *dma_dev; in fsl_re_chan_probe() local
638 dma_dev = &re_priv->dma_dev; in fsl_re_chan_probe()
684 chan->chan.device = dma_dev; in fsl_re_chan_probe()
754 struct dma_device *dma_dev; in fsl_re_probe() local
783 dma_dev = &re_priv->dma_dev; in fsl_re_probe()
784 dma_dev->dev = dev; in fsl_re_probe()
785 INIT_LIST_HEAD(&dma_dev->channels); in fsl_re_probe()
788 dma_dev->device_alloc_chan_resources = fsl_re_alloc_chan_resources; in fsl_re_probe()
789 dma_dev->device_tx_status = fsl_re_tx_status; in fsl_re_probe()
790 dma_dev->device_issue_pending = fsl_re_issue_pending; in fsl_re_probe()
[all …]
Diop-adma.c1282 struct dma_device *dma_dev; in iop_adma_probe() local
1296 dma_dev = &adev->common; in iop_adma_probe()
1318 dma_dev->cap_mask = plat_data->cap_mask; in iop_adma_probe()
1323 INIT_LIST_HEAD(&dma_dev->channels); in iop_adma_probe()
1326 dma_dev->device_alloc_chan_resources = iop_adma_alloc_chan_resources; in iop_adma_probe()
1327 dma_dev->device_free_chan_resources = iop_adma_free_chan_resources; in iop_adma_probe()
1328 dma_dev->device_tx_status = iop_adma_status; in iop_adma_probe()
1329 dma_dev->device_issue_pending = iop_adma_issue_pending; in iop_adma_probe()
1330 dma_dev->dev = &pdev->dev; in iop_adma_probe()
1333 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in iop_adma_probe()
[all …]
Dmv_xor_v2.c726 struct dma_device *dma_dev; in mv_xor_v2_probe() local
838 dma_dev = &xor_dev->dmadev; in mv_xor_v2_probe()
841 dma_cap_zero(dma_dev->cap_mask); in mv_xor_v2_probe()
842 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in mv_xor_v2_probe()
843 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in mv_xor_v2_probe()
844 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in mv_xor_v2_probe()
847 INIT_LIST_HEAD(&dma_dev->channels); in mv_xor_v2_probe()
850 dma_dev->device_tx_status = dma_cookie_status; in mv_xor_v2_probe()
851 dma_dev->device_issue_pending = mv_xor_v2_issue_pending; in mv_xor_v2_probe()
852 dma_dev->dev = &pdev->dev; in mv_xor_v2_probe()
[all …]
Dep93xx_dma.c201 struct dma_device dma_dev; member
1304 struct dma_device *dma_dev; in ep93xx_dma_probe() local
1313 dma_dev = &edma->dma_dev; in ep93xx_dma_probe()
1317 INIT_LIST_HEAD(&dma_dev->channels); in ep93xx_dma_probe()
1322 edmac->chan.device = dma_dev; in ep93xx_dma_probe()
1342 &dma_dev->channels); in ep93xx_dma_probe()
1345 dma_cap_zero(dma_dev->cap_mask); in ep93xx_dma_probe()
1346 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in ep93xx_dma_probe()
1347 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in ep93xx_dma_probe()
1349 dma_dev->dev = &pdev->dev; in ep93xx_dma_probe()
[all …]
Dxgene-dma.c320 struct dma_device dma_dev[XGENE_DMA_MAX_CHANNEL]; member
1495 struct dma_device *dma_dev) in xgene_dma_set_caps() argument
1498 dma_cap_zero(dma_dev->cap_mask); in xgene_dma_set_caps()
1513 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in xgene_dma_set_caps()
1514 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps()
1517 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps()
1521 dma_dev->dev = chan->dev; in xgene_dma_set_caps()
1522 dma_dev->device_alloc_chan_resources = xgene_dma_alloc_chan_resources; in xgene_dma_set_caps()
1523 dma_dev->device_free_chan_resources = xgene_dma_free_chan_resources; in xgene_dma_set_caps()
1524 dma_dev->device_issue_pending = xgene_dma_issue_pending; in xgene_dma_set_caps()
[all …]
Dmic_x100_dma.c474 dma_async_device_unregister(&mic_dma_dev->dma_dev); in mic_dma_unregister_dma_device()
606 dma_cap_zero(mic_dma_dev->dma_dev.cap_mask); in mic_dma_register_dma_device()
611 dma_cap_set(DMA_MEMCPY, mic_dma_dev->dma_dev.cap_mask); in mic_dma_register_dma_device()
614 dma_cap_set(DMA_PRIVATE, mic_dma_dev->dma_dev.cap_mask); in mic_dma_register_dma_device()
615 mic_dma_dev->dma_dev.device_alloc_chan_resources = in mic_dma_register_dma_device()
617 mic_dma_dev->dma_dev.device_free_chan_resources = in mic_dma_register_dma_device()
619 mic_dma_dev->dma_dev.device_tx_status = mic_dma_tx_status; in mic_dma_register_dma_device()
620 mic_dma_dev->dma_dev.device_prep_dma_memcpy = mic_dma_prep_memcpy_lock; in mic_dma_register_dma_device()
621 mic_dma_dev->dma_dev.device_prep_dma_imm_data = in mic_dma_register_dma_device()
623 mic_dma_dev->dma_dev.device_prep_dma_interrupt = in mic_dma_register_dma_device()
[all …]
Dfsl-edma.c172 struct dma_device dma_dev; member
764 list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels, device_node) { in fsl_edma_xlate()
931 INIT_LIST_HEAD(&fsl_edma->dma_dev.channels); in fsl_edma_probe()
940 vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev); in fsl_edma_probe()
951 dma_cap_set(DMA_PRIVATE, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
952 dma_cap_set(DMA_SLAVE, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
953 dma_cap_set(DMA_CYCLIC, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
955 fsl_edma->dma_dev.dev = &pdev->dev; in fsl_edma_probe()
956 fsl_edma->dma_dev.device_alloc_chan_resources in fsl_edma_probe()
958 fsl_edma->dma_dev.device_free_chan_resources in fsl_edma_probe()
[all …]
Dmv_xor.c1048 struct dma_device *dma_dev; in mv_xor_channel_add() local
1061 dma_dev = &mv_chan->dmadev; in mv_xor_channel_add()
1062 dma_dev->dev = &pdev->dev; in mv_xor_channel_add()
1070 mv_chan->dummy_src_addr = dma_map_single(dma_dev->dev, in mv_xor_channel_add()
1072 mv_chan->dummy_dst_addr = dma_map_single(dma_dev->dev, in mv_xor_channel_add()
1086 dma_dev->cap_mask = cap_mask; in mv_xor_channel_add()
1088 INIT_LIST_HEAD(&dma_dev->channels); in mv_xor_channel_add()
1091 dma_dev->device_alloc_chan_resources = mv_xor_alloc_chan_resources; in mv_xor_channel_add()
1092 dma_dev->device_free_chan_resources = mv_xor_free_chan_resources; in mv_xor_channel_add()
1093 dma_dev->device_tx_status = mv_xor_status; in mv_xor_channel_add()
[all …]
Dtegra20-apb-dma.c217 struct dma_device dma_dev; member
1252 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate()
1363 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_dma_probe()
1387 tdc->dma_chan.device = &tdma->dma_dev; in tegra_dma_probe()
1390 &tdma->dma_dev.channels); in tegra_dma_probe()
1405 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1406 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1407 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1410 tdma->dma_dev.dev = &pdev->dev; in tegra_dma_probe()
1411 tdma->dma_dev.device_alloc_chan_resources = in tegra_dma_probe()
[all …]
/drivers/mmc/host/
Dmmc_spi.c143 struct device *dma_dev; member
182 if (host->dma_dev) in mmc_spi_readbytes()
183 dma_sync_single_for_device(host->dma_dev, in mmc_spi_readbytes()
189 if (host->dma_dev) in mmc_spi_readbytes()
190 dma_sync_single_for_cpu(host->dma_dev, in mmc_spi_readbytes()
537 if (host->dma_dev) { in mmc_spi_command_send()
539 dma_sync_single_for_device(host->dma_dev, in mmc_spi_command_send()
545 if (host->dma_dev) in mmc_spi_command_send()
546 dma_sync_single_for_cpu(host->dma_dev, in mmc_spi_command_send()
682 if (host->dma_dev) in mmc_spi_writeblock()
[all …]
/drivers/crypto/ccp/
Dccp-dmaengine.c638 struct dma_device *dma_dev = &ccp->dma_dev; in ccp_dmaengine_register() local
681 dma_dev->dev = ccp->dev; in ccp_dmaengine_register()
682 dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register()
683 dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register()
684 dma_dev->directions = DMA_MEM_TO_MEM; in ccp_dmaengine_register()
685 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in ccp_dmaengine_register()
686 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ccp_dmaengine_register()
687 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in ccp_dmaengine_register()
696 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ccp_dmaengine_register()
698 INIT_LIST_HEAD(&dma_dev->channels); in ccp_dmaengine_register()
[all …]
/drivers/spi/
Dspi-pxa2xx-pci.c65 if (dws->dma_dev != chan->device->dev) in lpss_dma_filter()
74 struct pci_dev *dma_dev; in lpss_spi_setup() local
79 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(PCI_SLOT(dev->devfn), 0)); in lpss_spi_setup()
84 slave->dma_dev = &dma_dev->dev; in lpss_spi_setup()
92 slave->dma_dev = &dma_dev->dev; in lpss_spi_setup()
103 struct pci_dev *dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(21, 0)); in mrfld_spi_setup() local
130 tx->dma_dev = &dma_dev->dev; in mrfld_spi_setup()
133 rx->dma_dev = &dma_dev->dev; in mrfld_spi_setup()
Dspi-dw-mid.c39 if (s->dma_dev != chan->device->dev) in mid_spi_dma_chan_filter()
48 struct pci_dev *dma_dev; in mid_spi_dma_init() local
57 dma_dev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x0827, NULL); in mid_spi_dma_init()
58 if (!dma_dev) in mid_spi_dma_init()
65 rx->dma_dev = &dma_dev->dev; in mid_spi_dma_init()
72 tx->dma_dev = &dma_dev->dev; in mid_spi_dma_init()
/drivers/base/
Ddma-mapping.c341 struct device *bridge = NULL, *dma_dev = dev; in dma_configure() local
347 dma_dev = bridge; in dma_configure()
348 if (IS_ENABLED(CONFIG_OF) && dma_dev->parent && in dma_configure()
349 dma_dev->parent->of_node) in dma_configure()
350 dma_dev = dma_dev->parent; in dma_configure()
353 if (dma_dev->of_node) { in dma_configure()
354 ret = of_dma_configure(dev, dma_dev->of_node); in dma_configure()
355 } else if (has_acpi_companion(dma_dev)) { in dma_configure()
356 attr = acpi_get_dma_attr(to_acpi_device_node(dma_dev->fwnode)); in dma_configure()
/drivers/net/ethernet/broadcom/
Dbgmac.c134 struct device *dma_dev = bgmac->dma_dev; in bgmac_dma_tx_add() local
161 slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb), in bgmac_dma_tx_add()
163 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) in bgmac_dma_tx_add()
179 slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0, in bgmac_dma_tx_add()
181 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) in bgmac_dma_tx_add()
210 dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb), in bgmac_dma_tx_add()
219 dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE); in bgmac_dma_tx_add()
236 struct device *dma_dev = bgmac->dma_dev; in bgmac_dma_tx_free() local
262 dma_unmap_single(dma_dev, slot->dma_addr, len, in bgmac_dma_tx_free()
265 dma_unmap_page(dma_dev, slot->dma_addr, len, in bgmac_dma_tx_free()
[all …]
/drivers/tty/serial/8250/
D8250_mid.c47 struct pci_dev *dma_dev; member
73 mid->dma_dev = pci_get_slot(pdev->bus, in pnw_setup()
87 chip = pci_get_drvdata(mid->dma_dev); in tng_handle_irq()
123 mid->dma_dev = pci_get_slot(pdev->bus, PCI_DEVFN(5, 0)); in tng_setup()
186 mid->dma_dev = pdev; in dnv_setup()
194 if (!mid->dma_dev) in dnv_exit()
241 if (s->dma_dev != chan->device->dev || s->chan_id != chan->chan_id) in mid8250_dma_filter()
255 if (!mid->dma_dev) in mid8250_dma_setup()
272 rx_param->dma_dev = &mid->dma_dev->dev; in mid8250_dma_setup()
273 tx_param->dma_dev = &mid->dma_dev->dev; in mid8250_dma_setup()
/drivers/dma/sh/
Dsudmac.c55 struct sudmac_device, shdma_dev.dma_dev)
242 struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev); in sudmac_chan_probe()
280 dev_err(sdev->dma_dev.dev, in sudmac_chan_probe()
335 struct dma_device *dma_dev; in sudmac_probe() local
352 dma_dev = &su_dev->shdma_dev.dma_dev; in sudmac_probe()
359 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in sudmac_probe()
379 err = dma_async_device_register(&su_dev->shdma_dev.dma_dev); in sudmac_probe()
396 struct dma_device *dma_dev = &su_dev->shdma_dev.dma_dev; in sudmac_remove() local
398 dma_async_device_unregister(dma_dev); in sudmac_remove()
Dshdmac.c169 dev_warn(shdev->shdma_dev.dma_dev.dev, "Can't initialize DMAOR.\n"); in sh_dmae_rst()
173 dev_warn(shdev->shdma_dev.dma_dev.dev, in sh_dmae_rst()
528 struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev); in sh_dmae_chan_probe()
533 sh_chan = devm_kzalloc(sdev->dma_dev.dev, sizeof(struct sh_dmae_chan), in sh_dmae_chan_probe()
555 dev_err(sdev->dma_dev.dev, in sh_dmae_chan_probe()
695 struct dma_device *dma_dev; in sh_dmae_probe() local
735 dma_dev = &shdev->shdma_dev.dma_dev; in sh_dmae_probe()
746 dma_dev->src_addr_widths = widths; in sh_dmae_probe()
747 dma_dev->dst_addr_widths = widths; in sh_dmae_probe()
748 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM); in sh_dmae_probe()
[all …]
Dshdma-base.c40 #define to_shdma_dev(d) container_of(d, struct shdma_dev, dma_dev)
309 dev_warn(sdev->dma_dev.dev, "invalid slave ID passed to dma_request_slave\n"); in shdma_chan_filter()
964 schan->dma_chan.device = &sdev->dma_dev; in shdma_chan_probe()
967 schan->dev = sdev->dma_dev.dev; in shdma_chan_probe()
981 &sdev->dma_dev.channels); in shdma_chan_probe()
995 struct dma_device *dma_dev = &sdev->dma_dev; in shdma_init() local
1018 INIT_LIST_HEAD(&dma_dev->channels); in shdma_init()
1021 dma_dev->device_alloc_chan_resources in shdma_init()
1023 dma_dev->device_free_chan_resources = shdma_free_chan_resources; in shdma_init()
1024 dma_dev->device_prep_dma_memcpy = shdma_prep_memcpy; in shdma_init()
[all …]
/drivers/gpu/drm/mediatek/
Dmtk_drm_drv.c188 struct device *dma_dev; in mtk_drm_kms_init() local
246 dma_dev = &pdev->dev; in mtk_drm_kms_init()
247 private->dma_dev = dma_dev; in mtk_drm_kms_init()
253 if (!dma_dev->dma_parms) { in mtk_drm_kms_init()
255 dma_dev->dma_parms = in mtk_drm_kms_init()
256 devm_kzalloc(drm->dev, sizeof(*dma_dev->dma_parms), in mtk_drm_kms_init()
259 if (!dma_dev->dma_parms) { in mtk_drm_kms_init()
264 ret = dma_set_max_seg_size(dma_dev, (unsigned int)DMA_BIT_MASK(32)); in mtk_drm_kms_init()
266 dev_err(dma_dev, "Failed to set DMA segment size\n"); in mtk_drm_kms_init()
287 dma_dev->dma_parms = NULL; in mtk_drm_kms_init()
[all …]

1234