/drivers/vdpa/vdpa_user/ |
D | iova_domain.c | 20 static int vduse_iotlb_add_range(struct vduse_iova_domain *domain, in vduse_iotlb_add_range() argument 35 ret = vhost_iotlb_add_range_ctx(domain->iotlb, start, last, in vduse_iotlb_add_range() 45 static void vduse_iotlb_del_range(struct vduse_iova_domain *domain, in vduse_iotlb_del_range() argument 51 while ((map = vhost_iotlb_itree_first(domain->iotlb, start, last))) { in vduse_iotlb_del_range() 55 vhost_iotlb_map_free(domain->iotlb, map); in vduse_iotlb_del_range() 59 int vduse_domain_set_map(struct vduse_iova_domain *domain, in vduse_domain_set_map() argument 67 spin_lock(&domain->iotlb_lock); in vduse_domain_set_map() 68 vduse_iotlb_del_range(domain, start, last); in vduse_domain_set_map() 73 ret = vduse_iotlb_add_range(domain, map->start, map->last, in vduse_domain_set_map() 80 spin_unlock(&domain->iotlb_lock); in vduse_domain_set_map() [all …]
|
/drivers/scsi/elx/libefc/ |
D | efc_domain.c | 17 struct efc_domain *domain = NULL; in efc_domain_cb() local 22 domain = data; in efc_domain_cb() 37 domain = efc->domain; in efc_domain_cb() 38 if (!domain) { in efc_domain_cb() 39 domain = efc_domain_alloc(efc, fcf_wwn); in efc_domain_cb() 40 if (!domain) { in efc_domain_cb() 45 efc_sm_transition(&domain->drvsm, __efc_domain_init, in efc_domain_cb() 48 efc_domain_post_event(domain, EFC_EVT_DOMAIN_FOUND, drec); in efc_domain_cb() 53 domain_trace(domain, "EFC_HW_DOMAIN_LOST:\n"); in efc_domain_cb() 55 efc_domain_post_event(domain, EFC_EVT_DOMAIN_LOST, NULL); in efc_domain_cb() [all …]
|
D | efc_cmds.c | 132 nport->indicator, nport->domain->indicator); in efc_nport_alloc_init_vpi() 207 struct efc_domain *domain, u8 *wwpn) in efc_cmd_nport_alloc() argument 227 if (domain) { in efc_cmd_nport_alloc() 278 nport->domain->indicator, false); in efc_cmd_nport_attach() 317 efc_domain_get_mbox_status(struct efc_domain *domain, u8 *mqe, int status) in efc_domain_get_mbox_status() argument 319 struct efc *efc = domain->efc; in efc_domain_get_mbox_status() 325 domain->indicator, status, in efc_domain_get_mbox_status() 334 efc_domain_free_resources(struct efc_domain *domain, int evt, void *data) in efc_domain_free_resources() argument 336 struct efc *efc = domain->efc; in efc_domain_free_resources() 339 if (domain->dma.virt) { in efc_domain_free_resources() [all …]
|
D | efc_nport.c | 45 efc_nport_find_wwn(struct efc_domain *domain, uint64_t wwnn, uint64_t wwpn) in efc_nport_find_wwn() argument 50 list_for_each_entry(nport, &domain->nport_list, list_entry) { in efc_nport_find_wwn() 66 efc_nport_alloc(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, in efc_nport_alloc() argument 71 if (domain->efc->enable_ini) in efc_nport_alloc() 76 nport = efc_nport_find_wwn(domain, wwnn, wwpn); in efc_nport_alloc() 78 efc_log_err(domain->efc, in efc_nport_alloc() 93 nport->efc = domain->efc; in efc_nport_alloc() 95 nport->domain = domain; in efc_nport_alloc() 97 nport->instance_index = domain->nport_count++; in efc_nport_alloc() 105 memcpy(nport->service_params, domain->service_params, in efc_nport_alloc() [all …]
|
/drivers/soc/imx/ |
D | imx93-pd.c | 37 struct imx93_power_domain *domain = to_imx93_pd(genpd); in imx93_pd_on() local 38 void __iomem *addr = domain->addr; in imx93_pd_on() 42 ret = clk_bulk_prepare_enable(domain->num_clks, domain->clks); in imx93_pd_on() 44 dev_err(domain->dev, "failed to enable clocks for domain: %s\n", genpd->name); in imx93_pd_on() 55 dev_err(domain->dev, "pd_on timeout: name: %s, stat: %x\n", genpd->name, val); in imx93_pd_on() 64 struct imx93_power_domain *domain = to_imx93_pd(genpd); in imx93_pd_off() local 65 void __iomem *addr = domain->addr; in imx93_pd_off() 77 dev_err(domain->dev, "pd_off timeout: name: %s, stat: %x\n", genpd->name, val); in imx93_pd_off() 81 clk_bulk_disable_unprepare(domain->num_clks, domain->clks); in imx93_pd_off() 88 struct imx93_power_domain *domain = platform_get_drvdata(pdev); in imx93_pd_remove() local [all …]
|
D | gpcv2.c | 318 struct imx_pgc_domain *domain = to_imx_pgc_domain(genpd); in imx_pgc_power_up() local 322 ret = pm_runtime_get_sync(domain->dev); in imx_pgc_power_up() 324 pm_runtime_put_noidle(domain->dev); in imx_pgc_power_up() 328 if (!IS_ERR(domain->regulator)) { in imx_pgc_power_up() 329 ret = regulator_enable(domain->regulator); in imx_pgc_power_up() 331 dev_err(domain->dev, in imx_pgc_power_up() 338 reset_control_assert(domain->reset); in imx_pgc_power_up() 341 ret = clk_bulk_prepare_enable(domain->num_clks, domain->clks); in imx_pgc_power_up() 343 dev_err(domain->dev, "failed to enable reset clocks\n"); in imx_pgc_power_up() 350 if (domain->bits.pxx) { in imx_pgc_power_up() [all …]
|
D | imx8mp-blk-ctrl.c | 34 void (*power_off) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain); 35 void (*power_on) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain); 64 void (*power_off) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain); 65 void (*power_on) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain); 77 struct imx8mp_blk_ctrl_domain *domain) in imx8mp_hsio_blk_ctrl_power_on() argument 79 switch (domain->id) { in imx8mp_hsio_blk_ctrl_power_on() 96 struct imx8mp_blk_ctrl_domain *domain) in imx8mp_hsio_blk_ctrl_power_off() argument 98 switch (domain->id) { in imx8mp_hsio_blk_ctrl_power_off() 206 struct imx8mp_blk_ctrl_domain *domain) in imx8mp_hdmi_blk_ctrl_power_on() argument 208 switch (domain->id) { in imx8mp_hdmi_blk_ctrl_power_on() [all …]
|
D | gpc.c | 130 static int imx_pgc_get_clocks(struct device *dev, struct imx_pm_domain *domain) in imx_pgc_get_clocks() argument 143 domain->clk[i] = clk; in imx_pgc_get_clocks() 145 domain->num_clks = i; in imx_pgc_get_clocks() 151 clk_put(domain->clk[i]); in imx_pgc_get_clocks() 156 static void imx_pgc_put_clocks(struct imx_pm_domain *domain) in imx_pgc_put_clocks() argument 160 for (i = domain->num_clks - 1; i >= 0; i--) in imx_pgc_put_clocks() 161 clk_put(domain->clk[i]); in imx_pgc_put_clocks() 164 static int imx_pgc_parse_dt(struct device *dev, struct imx_pm_domain *domain) in imx_pgc_parse_dt() argument 167 domain->supply = devm_regulator_get_optional(dev, "power"); in imx_pgc_parse_dt() 168 if (IS_ERR(domain->supply)) { in imx_pgc_parse_dt() [all …]
|
/drivers/iommu/ |
D | ipmmu-vmsa.c | 212 static u32 ipmmu_ctx_read_root(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_read_root() argument 215 return ipmmu_ctx_read(domain->mmu->root, domain->context_id, reg); in ipmmu_ctx_read_root() 218 static void ipmmu_ctx_write_root(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_write_root() argument 221 ipmmu_ctx_write(domain->mmu->root, domain->context_id, reg, data); in ipmmu_ctx_write_root() 224 static void ipmmu_ctx_write_all(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_write_all() argument 227 if (domain->mmu != domain->mmu->root) in ipmmu_ctx_write_all() 228 ipmmu_ctx_write(domain->mmu, domain->context_id, reg, data); in ipmmu_ctx_write_all() 230 ipmmu_ctx_write(domain->mmu->root, domain->context_id, reg, data); in ipmmu_ctx_write_all() 255 static void ipmmu_tlb_sync(struct ipmmu_vmsa_domain *domain) in ipmmu_tlb_sync() argument 259 while (ipmmu_ctx_read_root(domain, IMCTR) & IMCTR_FLUSH) { in ipmmu_tlb_sync() [all …]
|
D | exynos-iommu.c | 230 struct iommu_domain *domain; /* domain this device is attached */ member 246 struct iommu_domain domain; /* generic domain data structure */ member 281 struct exynos_iommu_domain *domain; /* domain we belong to */ member 331 return container_of(dom, struct exynos_iommu_domain, domain); in to_exynos_domain() 506 if (data->domain) in exynos_sysmmu_irq() 507 ret = report_iommu_fault(&data->domain->domain, in exynos_sysmmu_irq() 764 if (data->domain) { in exynos_sysmmu_suspend() 782 if (data->domain) { in exynos_sysmmu_resume() 823 struct exynos_iommu_domain *domain; in exynos_iommu_domain_alloc() local 833 domain = kzalloc(sizeof(*domain), GFP_KERNEL); in exynos_iommu_domain_alloc() [all …]
|
D | iommu.c | 58 struct iommu_domain *domain; member 94 static int __iommu_attach_device(struct iommu_domain *domain, 96 static int __iommu_attach_group(struct iommu_domain *domain, 900 struct iommu_domain *domain = group->default_domain; in iommu_create_device_direct_mappings() local 906 if (!domain || !iommu_is_dma_domain(domain)) in iommu_create_device_direct_mappings() 909 BUG_ON(!domain->pgsize_bitmap); in iommu_create_device_direct_mappings() 911 pg_size = 1UL << __ffs(domain->pgsize_bitmap); in iommu_create_device_direct_mappings() 934 phys_addr = iommu_iova_to_phys(domain, addr); in iommu_create_device_direct_mappings() 942 ret = iommu_map(domain, addr - map_size, in iommu_create_device_direct_mappings() 953 iommu_flush_iotlb_all(domain); in iommu_create_device_direct_mappings() [all …]
|
D | iommu-sva.c | 92 struct iommu_domain *domain; in iommu_sva_bind_device() local 112 domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, in iommu_sva_bind_device() 114 if (IS_ERR(domain)) { in iommu_sva_bind_device() 115 ret = PTR_ERR(domain); in iommu_sva_bind_device() 119 if (domain) { in iommu_sva_bind_device() 120 domain->users++; in iommu_sva_bind_device() 125 domain = iommu_sva_domain_alloc(dev, mm); in iommu_sva_bind_device() 126 if (!domain) { in iommu_sva_bind_device() 131 ret = iommu_attach_device_pasid(domain, dev, mm->pasid); in iommu_sva_bind_device() 134 domain->users = 1; in iommu_sva_bind_device() [all …]
|
D | sprd-iommu.c | 77 struct iommu_domain domain; member 87 return container_of(dom, struct sprd_iommu_domain, domain); in to_sprd_domain() 128 sprd_iommu_pgt_size(struct iommu_domain *domain) in sprd_iommu_pgt_size() argument 130 return ((domain->geometry.aperture_end - in sprd_iommu_pgt_size() 131 domain->geometry.aperture_start + 1) >> in sprd_iommu_pgt_size() 148 dom->domain.geometry.aperture_start = 0; in sprd_iommu_domain_alloc() 149 dom->domain.geometry.aperture_end = SZ_256M - 1; in sprd_iommu_domain_alloc() 150 dom->domain.geometry.force_aperture = true; in sprd_iommu_domain_alloc() 152 return &dom->domain; in sprd_iommu_domain_alloc() 155 static void sprd_iommu_domain_free(struct iommu_domain *domain) in sprd_iommu_domain_free() argument [all …]
|
D | s390-iommu.c | 25 struct iommu_domain domain; member 39 return container_of(dom, struct s390_domain, domain); in to_s390_domain() 75 return &s390_domain->domain; in s390_domain_alloc() 78 static void s390_domain_free(struct iommu_domain *domain) in s390_domain_free() argument 80 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_domain_free() 112 static int s390_iommu_attach_device(struct iommu_domain *domain, in s390_iommu_attach_device() argument 115 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_attach_device() 144 domain->geometry.aperture_start = zdev->start_dma; in s390_iommu_attach_device() 145 domain->geometry.aperture_end = zdev->end_dma; in s390_iommu_attach_device() 146 domain->geometry.force_aperture = true; in s390_iommu_attach_device() [all …]
|
/drivers/dca/ |
D | dca-core.c | 44 struct dca_domain *domain; in dca_allocate_domain() local 46 domain = kzalloc(sizeof(*domain), GFP_NOWAIT); in dca_allocate_domain() 47 if (!domain) in dca_allocate_domain() 50 INIT_LIST_HEAD(&domain->dca_providers); in dca_allocate_domain() 51 domain->pci_rc = rc; in dca_allocate_domain() 53 return domain; in dca_allocate_domain() 56 static void dca_free_domain(struct dca_domain *domain) in dca_free_domain() argument 58 list_del(&domain->node); in dca_free_domain() 59 kfree(domain); in dca_free_domain() 81 struct dca_domain *domain; in unregister_dca_providers() local [all …]
|
/drivers/soc/dove/ |
D | pmu.c | 143 static int pmu_domain_power_off(struct generic_pm_domain *domain) in pmu_domain_power_off() argument 145 struct pmu_domain *pmu_dom = to_pmu_domain(domain); in pmu_domain_power_off() 177 static int pmu_domain_power_on(struct generic_pm_domain *domain) in pmu_domain_power_on() argument 179 struct pmu_domain *pmu_dom = to_pmu_domain(domain); in pmu_domain_power_on() 211 static void __pmu_domain_register(struct pmu_domain *domain, in __pmu_domain_register() argument 214 unsigned int val = readl_relaxed(domain->pmu->pmu_base + PMU_PWR); in __pmu_domain_register() 216 domain->base.power_off = pmu_domain_power_off; in __pmu_domain_register() 217 domain->base.power_on = pmu_domain_power_on; in __pmu_domain_register() 219 pm_genpd_init(&domain->base, NULL, !(val & domain->pwr_mask)); in __pmu_domain_register() 222 of_genpd_add_provider_simple(np, &domain->base); in __pmu_domain_register() [all …]
|
/drivers/firmware/arm_scmi/ |
D | perf.c | 76 __le32 domain; member 81 __le32 domain; member 92 __le32 domain; member 97 __le32 domain; member 191 u32 domain, struct perf_dom_info *dom_info, in scmi_perf_domain_attributes_get() argument 200 sizeof(domain), sizeof(*attr), &t); in scmi_perf_domain_attributes_get() 204 put_unaligned_le32(domain, t->tx.buf); in scmi_perf_domain_attributes_get() 239 ph->hops->extended_name_get(ph, PERF_DOMAIN_NAME_GET, domain, in scmi_perf_domain_attributes_get() 253 u32 domain; member 264 msg->domain = cpu_to_le32(p->domain); in iter_perf_levels_prepare_message() [all …]
|
/drivers/irqchip/ |
D | irq-atmel-aic5.c | 88 struct irq_domain *domain = d->domain; in aic5_mask() local 89 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_mask() 105 struct irq_domain *domain = d->domain; in aic5_unmask() local 106 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_unmask() 122 struct irq_domain *domain = d->domain; in aic5_retrigger() local 123 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_retrigger() 136 struct irq_domain *domain = d->domain; in aic5_set_type() local 137 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_set_type() 157 struct irq_domain *domain = d->domain; in aic5_suspend() local 158 struct irq_domain_chip_generic *dgc = domain->gc; in aic5_suspend() [all …]
|
/drivers/iommu/intel/ |
D | iommu.c | 380 static inline int domain_type_is_si(struct dmar_domain *domain) in domain_type_is_si() argument 382 return domain->domain.type == IOMMU_DOMAIN_IDENTITY; in domain_type_is_si() 385 static inline bool domain_use_first_level(struct dmar_domain *domain) in domain_use_first_level() argument 387 return domain->flags & DOMAIN_FLAG_USE_FIRST_LEVEL; in domain_use_first_level() 390 static inline int domain_pfn_supported(struct dmar_domain *domain, in domain_pfn_supported() argument 393 int addr_width = agaw_to_width(domain->agaw) - VTD_PAGE_SHIFT; in domain_pfn_supported() 459 static void domain_update_iommu_coherency(struct dmar_domain *domain) in domain_update_iommu_coherency() argument 467 domain->iommu_coherency = true; in domain_update_iommu_coherency() 468 xa_for_each(&domain->iommu_array, i, info) { in domain_update_iommu_coherency() 471 domain->iommu_coherency = false; in domain_update_iommu_coherency() [all …]
|
/drivers/iommu/amd/ |
D | iommu.c | 82 static int domain_enable_v2(struct protection_domain *domain, int pasids); 182 return container_of(dom, struct protection_domain, domain); in to_pdomain() 436 if (dev_data->domain) in amd_iommu_uninit_device() 562 if (dev_data->domain == NULL) { in amd_iommu_report_page_fault() 570 if (!report_iommu_fault(&dev_data->domain->domain, in amd_iommu_report_page_fault() 1356 static void __domain_flush_pages(struct protection_domain *domain, in __domain_flush_pages() argument 1363 build_inv_iommu_pages(&cmd, address, size, domain->id, pde); in __domain_flush_pages() 1366 if (!domain->dev_iommu[i]) in __domain_flush_pages() 1376 list_for_each_entry(dev_data, &domain->dev_list, list) { in __domain_flush_pages() 1387 static void domain_flush_pages(struct protection_domain *domain, in domain_flush_pages() argument [all …]
|
/drivers/soc/xilinx/ |
D | zynqmp_pm_domains.c | 73 static int zynqmp_gpd_power_on(struct generic_pm_domain *domain) in zynqmp_gpd_power_on() argument 75 struct zynqmp_pm_domain *pd = to_zynqmp_pm_domain(domain); in zynqmp_gpd_power_on() 83 dev_err(&domain->dev, in zynqmp_gpd_power_on() 89 dev_dbg(&domain->dev, "set requirement to 0x%x for PM node id %d\n", in zynqmp_gpd_power_on() 104 static int zynqmp_gpd_power_off(struct generic_pm_domain *domain) in zynqmp_gpd_power_off() argument 106 struct zynqmp_pm_domain *pd = to_zynqmp_pm_domain(domain); in zynqmp_gpd_power_off() 114 dev_dbg(&domain->dev, "PM node id %d is already released\n", in zynqmp_gpd_power_off() 119 list_for_each_entry_safe(pdd, tmp, &domain->dev_list, list_node) { in zynqmp_gpd_power_off() 124 domain->name); in zynqmp_gpd_power_off() 133 dev_err(&domain->dev, in zynqmp_gpd_power_off() [all …]
|
/drivers/base/ |
D | platform-msi.c | 52 static int platform_msi_init(struct irq_domain *domain, in platform_msi_init() argument 57 return irq_domain_set_hwirq_and_chip(domain, virq, hwirq, in platform_msi_init() 126 struct irq_domain *domain; in platform_msi_create_irq_domain() local 135 domain = msi_create_irq_domain(fwnode, info, parent); in platform_msi_create_irq_domain() 136 if (domain) in platform_msi_create_irq_domain() 137 irq_domain_update_bus_token(domain, DOMAIN_BUS_PLATFORM_MSI); in platform_msi_create_irq_domain() 139 return domain; in platform_msi_create_irq_domain() 155 if (!dev->msi.domain || !write_msi_msg || !nvec || nvec > MAX_DEV_MSIS) in platform_msi_alloc_priv_data() 158 if (dev->msi.domain->bus_token != DOMAIN_BUS_PLATFORM_MSI) { in platform_msi_alloc_priv_data() 216 err = msi_domain_alloc_irqs(dev->msi.domain, dev, nvec); in platform_msi_domain_alloc_irqs() [all …]
|
/drivers/gpu/drm/i915/display/ |
D | intel_display_power.h | 176 intel_display_power_domain_str(enum intel_display_power_domain domain); 179 enum intel_display_power_domain domain); 181 enum intel_display_power_domain domain); 183 enum intel_display_power_domain domain); 186 enum intel_display_power_domain domain); 188 enum intel_display_power_domain domain, 193 enum intel_display_power_domain domain, 197 enum intel_display_power_domain domain, in intel_display_power_put_async() argument 200 __intel_display_power_put_async(i915, domain, wakeref); in intel_display_power_put_async() 204 enum intel_display_power_domain domain); [all …]
|
/drivers/media/platform/qcom/venus/ |
D | hfi_parser.c | 28 cap->domain = VIDC_SESSION_TYPE_DEC; in init_codecs() 35 cap->domain = VIDC_SESSION_TYPE_ENC; in init_codecs() 41 u32 codecs, u32 domain, func cb, void *data, in for_each_codec() argument 49 if (cap->valid && cap->domain == domain) in for_each_codec() 51 if (cap->codec & codecs && cap->domain == domain) in for_each_codec() 66 parse_alloc_mode(struct venus_core *core, u32 codecs, u32 domain, void *data) in parse_alloc_mode() argument 81 codecs, domain, fill_buf_mode, type, 1); in parse_alloc_mode() 100 parse_profile_level(struct venus_core *core, u32 codecs, u32 domain, void *data) in parse_profile_level() argument 111 for_each_codec(core->caps, ARRAY_SIZE(core->caps), codecs, domain, in parse_profile_level() 128 parse_caps(struct venus_core *core, u32 codecs, u32 domain, void *data) in parse_caps() argument [all …]
|
/drivers/gpu/drm/etnaviv/ |
D | etnaviv_perfmon.c | 19 const struct etnaviv_pm_domain *domain, 41 const struct etnaviv_pm_domain *domain, in perf_reg_read() argument 44 gpu_write(gpu, domain->profile_config, signal->data); in perf_reg_read() 46 return gpu_read(gpu, domain->profile_read); in perf_reg_read() 58 const struct etnaviv_pm_domain *domain, in pipe_perf_reg_read() argument 67 value += perf_reg_read(gpu, domain, signal); in pipe_perf_reg_read() 77 const struct etnaviv_pm_domain *domain, in pipe_reg_read() argument 96 const struct etnaviv_pm_domain *domain, in hi_total_cycle_read() argument 110 const struct etnaviv_pm_domain *domain, in hi_total_idle_cycle_read() argument 479 const struct etnaviv_pm_domain *domain = NULL; in pm_domain() local [all …]
|