/drivers/dma/ |
D | dmaengine.c | 188 bitmap_and(has.bits, want->bits, device->cap_mask.bits, in __dma_device_satisfies_mask() 241 if (!dma_has_cap(DMA_PRIVATE, chan->device->cap_mask)) in dma_chan_get() 375 if (dma_has_cap(DMA_PRIVATE, device->cap_mask)) in dma_issue_pending_all() 412 if (!dma_has_cap(cap, device->cap_mask) || in min_chan() 413 dma_has_cap(DMA_PRIVATE, device->cap_mask)) in min_chan() 457 if (dma_has_cap(DMA_PRIVATE, device->cap_mask)) in dma_channel_rebalance() 485 if (!(test_bit(DMA_SLAVE, device->cap_mask.bits) || in dma_get_slave_caps() 486 test_bit(DMA_CYCLIC, device->cap_mask.bits))) in dma_get_slave_caps() 528 if (dev->chancnt > 1 && !dma_has_cap(DMA_PRIVATE, dev->cap_mask)) in private_candidate() 565 dma_cap_set(DMA_PRIVATE, device->cap_mask); in find_candidate() [all …]
|
D | iop-adma.c | 482 iop_chan->device->common.cap_mask)) in iop_adma_alloc_chan_resources() 485 iop_chan->device->common.cap_mask)) in iop_adma_alloc_chan_resources() 1013 if (!dma_has_cap(DMA_XOR_VAL, dma_chan->device->cap_mask)) in iop_adma_xor_val_self_test() 1318 dma_dev->cap_mask = plat_data->cap_mask; in iop_adma_probe() 1333 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in iop_adma_probe() 1335 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in iop_adma_probe() 1339 if (dma_has_cap(DMA_XOR_VAL, dma_dev->cap_mask)) in iop_adma_probe() 1342 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in iop_adma_probe() 1346 if (dma_has_cap(DMA_PQ_VAL, dma_dev->cap_mask)) in iop_adma_probe() 1349 if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) in iop_adma_probe() [all …]
|
D | mv_xor.c | 1044 int idx, dma_cap_mask_t cap_mask, int irq) in mv_xor_channel_add() argument 1086 dma_dev->cap_mask = cap_mask; in mv_xor_channel_add() 1097 if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) in mv_xor_channel_add() 1099 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in mv_xor_channel_add() 1101 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in mv_xor_channel_add() 1136 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { in mv_xor_channel_add() 1143 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in mv_xor_channel_add() 1152 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "", in mv_xor_channel_add() 1153 dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "", in mv_xor_channel_add() 1154 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : ""); in mv_xor_channel_add() [all …]
|
D | at_hdmac.c | 1811 dma_cap_set(DMA_MEMCPY, at91sam9rl_config.cap_mask); in at_dma_probe() 1812 dma_cap_set(DMA_INTERLEAVE, at91sam9g45_config.cap_mask); in at_dma_probe() 1813 dma_cap_set(DMA_MEMCPY, at91sam9g45_config.cap_mask); in at_dma_probe() 1814 dma_cap_set(DMA_MEMSET, at91sam9g45_config.cap_mask); in at_dma_probe() 1815 dma_cap_set(DMA_MEMSET_SG, at91sam9g45_config.cap_mask); in at_dma_probe() 1816 dma_cap_set(DMA_PRIVATE, at91sam9g45_config.cap_mask); in at_dma_probe() 1817 dma_cap_set(DMA_SLAVE, at91sam9g45_config.cap_mask); in at_dma_probe() 1839 atdma->dma_common.cap_mask = plat_dat->cap_mask; in at_dma_probe() 1928 if (dma_has_cap(DMA_INTERLEAVE, atdma->dma_common.cap_mask)) in at_dma_probe() 1931 if (dma_has_cap(DMA_MEMCPY, atdma->dma_common.cap_mask)) in at_dma_probe() [all …]
|
D | bcm-sba-raid.c | 1578 dma_cap_zero(dma_dev->cap_mask); in sba_async_register() 1579 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in sba_async_register() 1580 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in sba_async_register() 1581 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in sba_async_register() 1582 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in sba_async_register() 1598 if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) in sba_async_register() 1602 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in sba_async_register() 1606 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in sba_async_register() 1612 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in sba_async_register() 1630 dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "interrupt " : "", in sba_async_register() [all …]
|
D | xgene-dma.c | 1498 dma_cap_zero(dma_dev->cap_mask); in xgene_dma_set_caps() 1513 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in xgene_dma_set_caps() 1514 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps() 1517 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps() 1527 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in xgene_dma_set_caps() 1533 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in xgene_dma_set_caps() 1578 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "XOR " : "", in xgene_dma_async_register() 1579 dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "PQ " : ""); in xgene_dma_async_register()
|
D | mv_xor_v2.c | 841 dma_cap_zero(dma_dev->cap_mask); in mv_xor_v2_probe() 842 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in mv_xor_v2_probe() 843 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in mv_xor_v2_probe() 844 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in mv_xor_v2_probe()
|
D | bcm2835-dma.c | 913 dma_cap_set(DMA_SLAVE, od->ddev.cap_mask); in bcm2835_dma_probe() 914 dma_cap_set(DMA_PRIVATE, od->ddev.cap_mask); in bcm2835_dma_probe() 915 dma_cap_set(DMA_CYCLIC, od->ddev.cap_mask); in bcm2835_dma_probe() 916 dma_cap_set(DMA_SLAVE, od->ddev.cap_mask); in bcm2835_dma_probe() 917 dma_cap_set(DMA_MEMCPY, od->ddev.cap_mask); in bcm2835_dma_probe()
|
D | img-mdc-dma.c | 890 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe() 891 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 892 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 893 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe() 894 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe()
|
D | ep93xx_dma.c | 1345 dma_cap_zero(dma_dev->cap_mask); in ep93xx_dma_probe() 1346 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in ep93xx_dma_probe() 1347 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in ep93xx_dma_probe() 1363 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ep93xx_dma_probe() 1371 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ep93xx_dma_probe()
|
D | sun4i-dma.c | 1153 dma_cap_zero(priv->slave.cap_mask); in sun4i_dma_probe() 1154 dma_cap_set(DMA_PRIVATE, priv->slave.cap_mask); in sun4i_dma_probe() 1155 dma_cap_set(DMA_MEMCPY, priv->slave.cap_mask); in sun4i_dma_probe() 1156 dma_cap_set(DMA_CYCLIC, priv->slave.cap_mask); in sun4i_dma_probe() 1157 dma_cap_set(DMA_SLAVE, priv->slave.cap_mask); in sun4i_dma_probe()
|
D | dma-jz4780.c | 713 dma_cap_mask_t mask = jzdma->dma_device.cap_mask; in jz4780_of_dma_xlate() 808 dma_cap_set(DMA_MEMCPY, dd->cap_mask); in jz4780_dma_probe() 809 dma_cap_set(DMA_SLAVE, dd->cap_mask); in jz4780_dma_probe() 810 dma_cap_set(DMA_CYCLIC, dd->cap_mask); in jz4780_dma_probe()
|
D | mmp_tdma.c | 597 dma_cap_mask_t mask = tdev->device.cap_mask; in mmp_tdma_xlate() 680 dma_cap_set(DMA_SLAVE, tdev->device.cap_mask); in mmp_tdma_probe() 681 dma_cap_set(DMA_CYCLIC, tdev->device.cap_mask); in mmp_tdma_probe()
|
D | moxart-dma.c | 592 dma_cap_zero(mdc->dma_slave.cap_mask); in moxart_probe() 593 dma_cap_set(DMA_SLAVE, mdc->dma_slave.cap_mask); in moxart_probe() 594 dma_cap_set(DMA_PRIVATE, mdc->dma_slave.cap_mask); in moxart_probe()
|
D | s3c24xx-dma.c | 1275 dma_cap_set(DMA_MEMCPY, s3cdma->memcpy.cap_mask); in s3c24xx_dma_probe() 1276 dma_cap_set(DMA_PRIVATE, s3cdma->memcpy.cap_mask); in s3c24xx_dma_probe() 1287 dma_cap_set(DMA_SLAVE, s3cdma->slave.cap_mask); in s3c24xx_dma_probe() 1288 dma_cap_set(DMA_CYCLIC, s3cdma->slave.cap_mask); in s3c24xx_dma_probe() 1289 dma_cap_set(DMA_PRIVATE, s3cdma->slave.cap_mask); in s3c24xx_dma_probe()
|
D | ste_dma40.c | 1953 dma_cap_mask_t cap = d40c->chan.device->cap_mask; in d40_config_memcpy() 2796 if (dma_has_cap(DMA_SLAVE, dev->cap_mask)) { in d40_ops_init() 2801 if (dma_has_cap(DMA_MEMCPY, dev->cap_mask)) { in d40_ops_init() 2811 if (dma_has_cap(DMA_CYCLIC, dev->cap_mask)) in d40_ops_init() 2834 dma_cap_zero(base->dma_slave.cap_mask); in d40_dmaengine_init() 2835 dma_cap_set(DMA_SLAVE, base->dma_slave.cap_mask); in d40_dmaengine_init() 2836 dma_cap_set(DMA_CYCLIC, base->dma_slave.cap_mask); in d40_dmaengine_init() 2850 dma_cap_zero(base->dma_memcpy.cap_mask); in d40_dmaengine_init() 2851 dma_cap_set(DMA_MEMCPY, base->dma_memcpy.cap_mask); in d40_dmaengine_init() 2866 dma_cap_zero(base->dma_both.cap_mask); in d40_dmaengine_init() [all …]
|
D | imx-dma.c | 1045 return dma_request_channel(imxdma->dma_device.cap_mask, in imxdma_xlate() 1130 dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); in imxdma_probe() 1131 dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); in imxdma_probe() 1132 dma_cap_set(DMA_MEMCPY, imxdma->dma_device.cap_mask); in imxdma_probe() 1133 dma_cap_set(DMA_INTERLEAVE, imxdma->dma_device.cap_mask); in imxdma_probe()
|
D | altera-msgdma.c | 849 dma_cap_zero(dma_dev->cap_mask); in msgdma_probe() 850 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in msgdma_probe() 851 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in msgdma_probe()
|
/drivers/pci/pcie/ |
D | portdrv_core.c | 234 int cap_mask = 0; in get_port_device_capability() local 239 cap_mask = PCIE_PORT_SERVICE_PME | PCIE_PORT_SERVICE_HP in get_port_device_capability() 242 cap_mask |= PCIE_PORT_SERVICE_AER; in get_port_device_capability() 245 pcie_port_platform_notify(dev, &cap_mask); in get_port_device_capability() 248 if ((cap_mask & PCIE_PORT_SERVICE_HP) && dev->is_hotplug_bridge) { in get_port_device_capability() 258 if ((cap_mask & PCIE_PORT_SERVICE_AER) in get_port_device_capability() 271 if ((cap_mask & PCIE_PORT_SERVICE_PME) in get_port_device_capability()
|
/drivers/dma/ioat/ |
D | sysfs.c | 34 dma_has_cap(DMA_PQ, dma->cap_mask) ? " pq" : "", in cap_show() 35 dma_has_cap(DMA_PQ_VAL, dma->cap_mask) ? " pq_val" : "", in cap_show() 36 dma_has_cap(DMA_XOR, dma->cap_mask) ? " xor" : "", in cap_show() 37 dma_has_cap(DMA_XOR_VAL, dma->cap_mask) ? " xor_val" : "", in cap_show() 38 dma_has_cap(DMA_INTERRUPT, dma->cap_mask) ? " intr" : ""); in cap_show()
|
D | init.c | 521 dma_cap_set(DMA_MEMCPY, dma->cap_mask); in ioat_probe() 790 if (!dma_has_cap(DMA_XOR, dma->cap_mask)) in ioat_xor_val_self_test() 898 if (!dma_has_cap(DMA_XOR_VAL, dma_chan->device->cap_mask)) in ioat_xor_val_self_test() 1087 dma_cap_set(DMA_INTERRUPT, dma->cap_mask); in ioat3_dma_probe() 1103 dma_cap_set(DMA_XOR, dma->cap_mask); in ioat3_dma_probe() 1106 dma_cap_set(DMA_XOR_VAL, dma->cap_mask); in ioat3_dma_probe() 1114 dma_cap_set(DMA_PQ, dma->cap_mask); in ioat3_dma_probe() 1115 dma_cap_set(DMA_PQ_VAL, dma->cap_mask); in ioat3_dma_probe() 1125 dma_cap_set(DMA_XOR, dma->cap_mask); in ioat3_dma_probe() 1126 dma_cap_set(DMA_XOR_VAL, dma->cap_mask); in ioat3_dma_probe() [all …]
|
/drivers/infiniband/hw/hfi1/ |
D | driver.c | 94 module_param_cb(cap_mask, &cap_ops, &hfi1_cap_mask, S_IWUSR | S_IRUGO); 95 MODULE_PARM_DESC(cap_mask, "Bit mask of enabled/disabled HW features"); 117 cap_mask = *cap_mask_ptr, value, diff, in hfi1_caps_set() local 127 diff = value ^ (cap_mask & ~HFI1_CAP_LOCKED_SMASK); in hfi1_caps_set() 139 cap_mask &= ~diff; in hfi1_caps_set() 141 cap_mask |= (value & diff); in hfi1_caps_set() 143 diff = (cap_mask & (HFI1_CAP_MUST_HAVE_KERN << HFI1_CAP_USER_SHIFT)) ^ in hfi1_caps_set() 144 ((cap_mask & HFI1_CAP_MUST_HAVE_KERN) << HFI1_CAP_USER_SHIFT); in hfi1_caps_set() 145 cap_mask &= ~diff; in hfi1_caps_set() 147 *cap_mask_ptr = cap_mask; in hfi1_caps_set() [all …]
|
/drivers/net/ethernet/qlogic/qlcnic/ |
D | qlcnic_minidump.c | 296 hdr->drv_cap_mask = hdr->cap_mask; in qlcnic_82xx_cache_tmpl_hdr_values() 297 fw_dump->cap_mask = hdr->cap_mask; in qlcnic_82xx_cache_tmpl_hdr_values() 350 hdr->drv_cap_mask = hdr->cap_mask; in qlcnic_83xx_cache_tmpl_hdr_values() 351 fw_dump->cap_mask = hdr->cap_mask; in qlcnic_83xx_cache_tmpl_hdr_values() 1279 fw_dump->cap_mask); in qlcnic_fw_cmd_get_minidump_temp() 1323 if (i & fw_dump->cap_mask) in qlcnic_dump_fw() 1354 if (!(entry->hdr.mask & fw_dump->cap_mask)) { in qlcnic_dump_fw() 1448 fw_dump->cap_mask = 0x1f; in qlcnic_83xx_get_minidump_template()
|
/drivers/dma/ppc4xx/ |
D | adma.c | 1414 if (dma_has_cap(cap, ref->chan->device->cap_mask)) { in ppc440spe_async_tx_find_best_channel() 3787 dma_cap_set(DMA_MEMCPY, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3788 dma_cap_set(DMA_INTERRUPT, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3789 dma_cap_set(DMA_PQ, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3790 dma_cap_set(DMA_PQ_VAL, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3791 dma_cap_set(DMA_XOR_VAL, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3794 dma_cap_set(DMA_XOR, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3795 dma_cap_set(DMA_PQ, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3796 dma_cap_set(DMA_INTERRUPT, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3797 adev->common.cap_mask = adev->common.cap_mask; in ppc440spe_adma_init_capabilities() [all …]
|
/drivers/crypto/ccp/ |
D | ccp-dmaengine.c | 686 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ccp_dmaengine_register() 687 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in ccp_dmaengine_register() 696 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ccp_dmaengine_register()
|