Home
last modified time | relevance | path

Searched refs:domain (Results 1 – 25 of 672) sorted by relevance

12345678910>>...27

/drivers/vdpa/vdpa_user/
Diova_domain.c20 static int vduse_iotlb_add_range(struct vduse_iova_domain *domain, in vduse_iotlb_add_range() argument
35 ret = vhost_iotlb_add_range_ctx(domain->iotlb, start, last, in vduse_iotlb_add_range()
45 static void vduse_iotlb_del_range(struct vduse_iova_domain *domain, in vduse_iotlb_del_range() argument
51 while ((map = vhost_iotlb_itree_first(domain->iotlb, start, last))) { in vduse_iotlb_del_range()
55 vhost_iotlb_map_free(domain->iotlb, map); in vduse_iotlb_del_range()
59 int vduse_domain_set_map(struct vduse_iova_domain *domain, in vduse_domain_set_map() argument
67 spin_lock(&domain->iotlb_lock); in vduse_domain_set_map()
68 vduse_iotlb_del_range(domain, start, last); in vduse_domain_set_map()
73 ret = vduse_iotlb_add_range(domain, map->start, map->last, in vduse_domain_set_map()
80 spin_unlock(&domain->iotlb_lock); in vduse_domain_set_map()
[all …]
/drivers/scsi/elx/libefc/
Defc_domain.c17 struct efc_domain *domain = NULL; in efc_domain_cb() local
22 domain = data; in efc_domain_cb()
37 domain = efc->domain; in efc_domain_cb()
38 if (!domain) { in efc_domain_cb()
39 domain = efc_domain_alloc(efc, fcf_wwn); in efc_domain_cb()
40 if (!domain) { in efc_domain_cb()
45 efc_sm_transition(&domain->drvsm, __efc_domain_init, in efc_domain_cb()
48 efc_domain_post_event(domain, EFC_EVT_DOMAIN_FOUND, drec); in efc_domain_cb()
53 domain_trace(domain, "EFC_HW_DOMAIN_LOST:\n"); in efc_domain_cb()
55 efc_domain_post_event(domain, EFC_EVT_DOMAIN_LOST, NULL); in efc_domain_cb()
[all …]
Defc_cmds.c132 nport->indicator, nport->domain->indicator); in efc_nport_alloc_init_vpi()
207 struct efc_domain *domain, u8 *wwpn) in efc_cmd_nport_alloc() argument
227 if (domain) { in efc_cmd_nport_alloc()
278 nport->domain->indicator, false); in efc_cmd_nport_attach()
317 efc_domain_get_mbox_status(struct efc_domain *domain, u8 *mqe, int status) in efc_domain_get_mbox_status() argument
319 struct efc *efc = domain->efc; in efc_domain_get_mbox_status()
325 domain->indicator, status, in efc_domain_get_mbox_status()
334 efc_domain_free_resources(struct efc_domain *domain, int evt, void *data) in efc_domain_free_resources() argument
336 struct efc *efc = domain->efc; in efc_domain_free_resources()
339 if (domain->dma.virt) { in efc_domain_free_resources()
[all …]
Defc_nport.c45 efc_nport_find_wwn(struct efc_domain *domain, uint64_t wwnn, uint64_t wwpn) in efc_nport_find_wwn() argument
50 list_for_each_entry(nport, &domain->nport_list, list_entry) { in efc_nport_find_wwn()
66 efc_nport_alloc(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, in efc_nport_alloc() argument
71 if (domain->efc->enable_ini) in efc_nport_alloc()
76 nport = efc_nport_find_wwn(domain, wwnn, wwpn); in efc_nport_alloc()
78 efc_log_err(domain->efc, in efc_nport_alloc()
93 nport->efc = domain->efc; in efc_nport_alloc()
95 nport->domain = domain; in efc_nport_alloc()
97 nport->instance_index = domain->nport_count++; in efc_nport_alloc()
105 memcpy(nport->service_params, domain->service_params, in efc_nport_alloc()
[all …]
/drivers/soc/imx/
Dimx93-pd.c37 struct imx93_power_domain *domain = to_imx93_pd(genpd); in imx93_pd_on() local
38 void __iomem *addr = domain->addr; in imx93_pd_on()
42 ret = clk_bulk_prepare_enable(domain->num_clks, domain->clks); in imx93_pd_on()
44 dev_err(domain->dev, "failed to enable clocks for domain: %s\n", genpd->name); in imx93_pd_on()
55 dev_err(domain->dev, "pd_on timeout: name: %s, stat: %x\n", genpd->name, val); in imx93_pd_on()
64 struct imx93_power_domain *domain = to_imx93_pd(genpd); in imx93_pd_off() local
65 void __iomem *addr = domain->addr; in imx93_pd_off()
77 dev_err(domain->dev, "pd_off timeout: name: %s, stat: %x\n", genpd->name, val); in imx93_pd_off()
81 clk_bulk_disable_unprepare(domain->num_clks, domain->clks); in imx93_pd_off()
88 struct imx93_power_domain *domain = platform_get_drvdata(pdev); in imx93_pd_remove() local
[all …]
Dgpcv2.c318 struct imx_pgc_domain *domain = to_imx_pgc_domain(genpd); in imx_pgc_power_up() local
322 ret = pm_runtime_get_sync(domain->dev); in imx_pgc_power_up()
324 pm_runtime_put_noidle(domain->dev); in imx_pgc_power_up()
328 if (!IS_ERR(domain->regulator)) { in imx_pgc_power_up()
329 ret = regulator_enable(domain->regulator); in imx_pgc_power_up()
331 dev_err(domain->dev, in imx_pgc_power_up()
338 reset_control_assert(domain->reset); in imx_pgc_power_up()
341 ret = clk_bulk_prepare_enable(domain->num_clks, domain->clks); in imx_pgc_power_up()
343 dev_err(domain->dev, "failed to enable reset clocks\n"); in imx_pgc_power_up()
350 if (domain->bits.pxx) { in imx_pgc_power_up()
[all …]
Dimx8mp-blk-ctrl.c34 void (*power_off) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain);
35 void (*power_on) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain);
64 void (*power_off) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain);
65 void (*power_on) (struct imx8mp_blk_ctrl *bc, struct imx8mp_blk_ctrl_domain *domain);
77 struct imx8mp_blk_ctrl_domain *domain) in imx8mp_hsio_blk_ctrl_power_on() argument
79 switch (domain->id) { in imx8mp_hsio_blk_ctrl_power_on()
96 struct imx8mp_blk_ctrl_domain *domain) in imx8mp_hsio_blk_ctrl_power_off() argument
98 switch (domain->id) { in imx8mp_hsio_blk_ctrl_power_off()
206 struct imx8mp_blk_ctrl_domain *domain) in imx8mp_hdmi_blk_ctrl_power_on() argument
208 switch (domain->id) { in imx8mp_hdmi_blk_ctrl_power_on()
[all …]
Dgpc.c130 static int imx_pgc_get_clocks(struct device *dev, struct imx_pm_domain *domain) in imx_pgc_get_clocks() argument
143 domain->clk[i] = clk; in imx_pgc_get_clocks()
145 domain->num_clks = i; in imx_pgc_get_clocks()
151 clk_put(domain->clk[i]); in imx_pgc_get_clocks()
156 static void imx_pgc_put_clocks(struct imx_pm_domain *domain) in imx_pgc_put_clocks() argument
160 for (i = domain->num_clks - 1; i >= 0; i--) in imx_pgc_put_clocks()
161 clk_put(domain->clk[i]); in imx_pgc_put_clocks()
164 static int imx_pgc_parse_dt(struct device *dev, struct imx_pm_domain *domain) in imx_pgc_parse_dt() argument
167 domain->supply = devm_regulator_get_optional(dev, "power"); in imx_pgc_parse_dt()
168 if (IS_ERR(domain->supply)) { in imx_pgc_parse_dt()
[all …]
/drivers/iommu/
Dipmmu-vmsa.c212 static u32 ipmmu_ctx_read_root(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_read_root() argument
215 return ipmmu_ctx_read(domain->mmu->root, domain->context_id, reg); in ipmmu_ctx_read_root()
218 static void ipmmu_ctx_write_root(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_write_root() argument
221 ipmmu_ctx_write(domain->mmu->root, domain->context_id, reg, data); in ipmmu_ctx_write_root()
224 static void ipmmu_ctx_write_all(struct ipmmu_vmsa_domain *domain, in ipmmu_ctx_write_all() argument
227 if (domain->mmu != domain->mmu->root) in ipmmu_ctx_write_all()
228 ipmmu_ctx_write(domain->mmu, domain->context_id, reg, data); in ipmmu_ctx_write_all()
230 ipmmu_ctx_write(domain->mmu->root, domain->context_id, reg, data); in ipmmu_ctx_write_all()
255 static void ipmmu_tlb_sync(struct ipmmu_vmsa_domain *domain) in ipmmu_tlb_sync() argument
259 while (ipmmu_ctx_read_root(domain, IMCTR) & IMCTR_FLUSH) { in ipmmu_tlb_sync()
[all …]
Dexynos-iommu.c230 struct iommu_domain *domain; /* domain this device is attached */ member
246 struct iommu_domain domain; /* generic domain data structure */ member
281 struct exynos_iommu_domain *domain; /* domain we belong to */ member
331 return container_of(dom, struct exynos_iommu_domain, domain); in to_exynos_domain()
506 if (data->domain) in exynos_sysmmu_irq()
507 ret = report_iommu_fault(&data->domain->domain, in exynos_sysmmu_irq()
764 if (data->domain) { in exynos_sysmmu_suspend()
782 if (data->domain) { in exynos_sysmmu_resume()
823 struct exynos_iommu_domain *domain; in exynos_iommu_domain_alloc() local
833 domain = kzalloc(sizeof(*domain), GFP_KERNEL); in exynos_iommu_domain_alloc()
[all …]
Diommu.c58 struct iommu_domain *domain; member
94 static int __iommu_attach_device(struct iommu_domain *domain,
96 static int __iommu_attach_group(struct iommu_domain *domain,
900 struct iommu_domain *domain = group->default_domain; in iommu_create_device_direct_mappings() local
906 if (!domain || !iommu_is_dma_domain(domain)) in iommu_create_device_direct_mappings()
909 BUG_ON(!domain->pgsize_bitmap); in iommu_create_device_direct_mappings()
911 pg_size = 1UL << __ffs(domain->pgsize_bitmap); in iommu_create_device_direct_mappings()
934 phys_addr = iommu_iova_to_phys(domain, addr); in iommu_create_device_direct_mappings()
942 ret = iommu_map(domain, addr - map_size, in iommu_create_device_direct_mappings()
953 iommu_flush_iotlb_all(domain); in iommu_create_device_direct_mappings()
[all …]
Diommu-sva.c92 struct iommu_domain *domain; in iommu_sva_bind_device() local
112 domain = iommu_get_domain_for_dev_pasid(dev, mm->pasid, in iommu_sva_bind_device()
114 if (IS_ERR(domain)) { in iommu_sva_bind_device()
115 ret = PTR_ERR(domain); in iommu_sva_bind_device()
119 if (domain) { in iommu_sva_bind_device()
120 domain->users++; in iommu_sva_bind_device()
125 domain = iommu_sva_domain_alloc(dev, mm); in iommu_sva_bind_device()
126 if (!domain) { in iommu_sva_bind_device()
131 ret = iommu_attach_device_pasid(domain, dev, mm->pasid); in iommu_sva_bind_device()
134 domain->users = 1; in iommu_sva_bind_device()
[all …]
Dsprd-iommu.c77 struct iommu_domain domain; member
87 return container_of(dom, struct sprd_iommu_domain, domain); in to_sprd_domain()
128 sprd_iommu_pgt_size(struct iommu_domain *domain) in sprd_iommu_pgt_size() argument
130 return ((domain->geometry.aperture_end - in sprd_iommu_pgt_size()
131 domain->geometry.aperture_start + 1) >> in sprd_iommu_pgt_size()
148 dom->domain.geometry.aperture_start = 0; in sprd_iommu_domain_alloc()
149 dom->domain.geometry.aperture_end = SZ_256M - 1; in sprd_iommu_domain_alloc()
150 dom->domain.geometry.force_aperture = true; in sprd_iommu_domain_alloc()
152 return &dom->domain; in sprd_iommu_domain_alloc()
155 static void sprd_iommu_domain_free(struct iommu_domain *domain) in sprd_iommu_domain_free() argument
[all …]
Ds390-iommu.c25 struct iommu_domain domain; member
39 return container_of(dom, struct s390_domain, domain); in to_s390_domain()
75 return &s390_domain->domain; in s390_domain_alloc()
78 static void s390_domain_free(struct iommu_domain *domain) in s390_domain_free() argument
80 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_domain_free()
112 static int s390_iommu_attach_device(struct iommu_domain *domain, in s390_iommu_attach_device() argument
115 struct s390_domain *s390_domain = to_s390_domain(domain); in s390_iommu_attach_device()
144 domain->geometry.aperture_start = zdev->start_dma; in s390_iommu_attach_device()
145 domain->geometry.aperture_end = zdev->end_dma; in s390_iommu_attach_device()
146 domain->geometry.force_aperture = true; in s390_iommu_attach_device()
[all …]
/drivers/dca/
Ddca-core.c44 struct dca_domain *domain; in dca_allocate_domain() local
46 domain = kzalloc(sizeof(*domain), GFP_NOWAIT); in dca_allocate_domain()
47 if (!domain) in dca_allocate_domain()
50 INIT_LIST_HEAD(&domain->dca_providers); in dca_allocate_domain()
51 domain->pci_rc = rc; in dca_allocate_domain()
53 return domain; in dca_allocate_domain()
56 static void dca_free_domain(struct dca_domain *domain) in dca_free_domain() argument
58 list_del(&domain->node); in dca_free_domain()
59 kfree(domain); in dca_free_domain()
81 struct dca_domain *domain; in unregister_dca_providers() local
[all …]
/drivers/soc/dove/
Dpmu.c143 static int pmu_domain_power_off(struct generic_pm_domain *domain) in pmu_domain_power_off() argument
145 struct pmu_domain *pmu_dom = to_pmu_domain(domain); in pmu_domain_power_off()
177 static int pmu_domain_power_on(struct generic_pm_domain *domain) in pmu_domain_power_on() argument
179 struct pmu_domain *pmu_dom = to_pmu_domain(domain); in pmu_domain_power_on()
211 static void __pmu_domain_register(struct pmu_domain *domain, in __pmu_domain_register() argument
214 unsigned int val = readl_relaxed(domain->pmu->pmu_base + PMU_PWR); in __pmu_domain_register()
216 domain->base.power_off = pmu_domain_power_off; in __pmu_domain_register()
217 domain->base.power_on = pmu_domain_power_on; in __pmu_domain_register()
219 pm_genpd_init(&domain->base, NULL, !(val & domain->pwr_mask)); in __pmu_domain_register()
222 of_genpd_add_provider_simple(np, &domain->base); in __pmu_domain_register()
[all …]
/drivers/firmware/arm_scmi/
Dperf.c76 __le32 domain; member
81 __le32 domain; member
92 __le32 domain; member
97 __le32 domain; member
191 u32 domain, struct perf_dom_info *dom_info, in scmi_perf_domain_attributes_get() argument
200 sizeof(domain), sizeof(*attr), &t); in scmi_perf_domain_attributes_get()
204 put_unaligned_le32(domain, t->tx.buf); in scmi_perf_domain_attributes_get()
239 ph->hops->extended_name_get(ph, PERF_DOMAIN_NAME_GET, domain, in scmi_perf_domain_attributes_get()
253 u32 domain; member
264 msg->domain = cpu_to_le32(p->domain); in iter_perf_levels_prepare_message()
[all …]
/drivers/irqchip/
Dirq-atmel-aic5.c88 struct irq_domain *domain = d->domain; in aic5_mask() local
89 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_mask()
105 struct irq_domain *domain = d->domain; in aic5_unmask() local
106 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_unmask()
122 struct irq_domain *domain = d->domain; in aic5_retrigger() local
123 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_retrigger()
136 struct irq_domain *domain = d->domain; in aic5_set_type() local
137 struct irq_chip_generic *bgc = irq_get_domain_generic_chip(domain, 0); in aic5_set_type()
157 struct irq_domain *domain = d->domain; in aic5_suspend() local
158 struct irq_domain_chip_generic *dgc = domain->gc; in aic5_suspend()
[all …]
/drivers/iommu/intel/
Diommu.c380 static inline int domain_type_is_si(struct dmar_domain *domain) in domain_type_is_si() argument
382 return domain->domain.type == IOMMU_DOMAIN_IDENTITY; in domain_type_is_si()
385 static inline bool domain_use_first_level(struct dmar_domain *domain) in domain_use_first_level() argument
387 return domain->flags & DOMAIN_FLAG_USE_FIRST_LEVEL; in domain_use_first_level()
390 static inline int domain_pfn_supported(struct dmar_domain *domain, in domain_pfn_supported() argument
393 int addr_width = agaw_to_width(domain->agaw) - VTD_PAGE_SHIFT; in domain_pfn_supported()
459 static void domain_update_iommu_coherency(struct dmar_domain *domain) in domain_update_iommu_coherency() argument
467 domain->iommu_coherency = true; in domain_update_iommu_coherency()
468 xa_for_each(&domain->iommu_array, i, info) { in domain_update_iommu_coherency()
471 domain->iommu_coherency = false; in domain_update_iommu_coherency()
[all …]
/drivers/iommu/amd/
Diommu.c82 static int domain_enable_v2(struct protection_domain *domain, int pasids);
182 return container_of(dom, struct protection_domain, domain); in to_pdomain()
436 if (dev_data->domain) in amd_iommu_uninit_device()
562 if (dev_data->domain == NULL) { in amd_iommu_report_page_fault()
570 if (!report_iommu_fault(&dev_data->domain->domain, in amd_iommu_report_page_fault()
1356 static void __domain_flush_pages(struct protection_domain *domain, in __domain_flush_pages() argument
1363 build_inv_iommu_pages(&cmd, address, size, domain->id, pde); in __domain_flush_pages()
1366 if (!domain->dev_iommu[i]) in __domain_flush_pages()
1376 list_for_each_entry(dev_data, &domain->dev_list, list) { in __domain_flush_pages()
1387 static void domain_flush_pages(struct protection_domain *domain, in domain_flush_pages() argument
[all …]
/drivers/soc/xilinx/
Dzynqmp_pm_domains.c73 static int zynqmp_gpd_power_on(struct generic_pm_domain *domain) in zynqmp_gpd_power_on() argument
75 struct zynqmp_pm_domain *pd = to_zynqmp_pm_domain(domain); in zynqmp_gpd_power_on()
83 dev_err(&domain->dev, in zynqmp_gpd_power_on()
89 dev_dbg(&domain->dev, "set requirement to 0x%x for PM node id %d\n", in zynqmp_gpd_power_on()
104 static int zynqmp_gpd_power_off(struct generic_pm_domain *domain) in zynqmp_gpd_power_off() argument
106 struct zynqmp_pm_domain *pd = to_zynqmp_pm_domain(domain); in zynqmp_gpd_power_off()
114 dev_dbg(&domain->dev, "PM node id %d is already released\n", in zynqmp_gpd_power_off()
119 list_for_each_entry_safe(pdd, tmp, &domain->dev_list, list_node) { in zynqmp_gpd_power_off()
124 domain->name); in zynqmp_gpd_power_off()
133 dev_err(&domain->dev, in zynqmp_gpd_power_off()
[all …]
/drivers/base/
Dplatform-msi.c52 static int platform_msi_init(struct irq_domain *domain, in platform_msi_init() argument
57 return irq_domain_set_hwirq_and_chip(domain, virq, hwirq, in platform_msi_init()
126 struct irq_domain *domain; in platform_msi_create_irq_domain() local
135 domain = msi_create_irq_domain(fwnode, info, parent); in platform_msi_create_irq_domain()
136 if (domain) in platform_msi_create_irq_domain()
137 irq_domain_update_bus_token(domain, DOMAIN_BUS_PLATFORM_MSI); in platform_msi_create_irq_domain()
139 return domain; in platform_msi_create_irq_domain()
155 if (!dev->msi.domain || !write_msi_msg || !nvec || nvec > MAX_DEV_MSIS) in platform_msi_alloc_priv_data()
158 if (dev->msi.domain->bus_token != DOMAIN_BUS_PLATFORM_MSI) { in platform_msi_alloc_priv_data()
216 err = msi_domain_alloc_irqs(dev->msi.domain, dev, nvec); in platform_msi_domain_alloc_irqs()
[all …]
/drivers/gpu/drm/i915/display/
Dintel_display_power.h176 intel_display_power_domain_str(enum intel_display_power_domain domain);
179 enum intel_display_power_domain domain);
181 enum intel_display_power_domain domain);
183 enum intel_display_power_domain domain);
186 enum intel_display_power_domain domain);
188 enum intel_display_power_domain domain,
193 enum intel_display_power_domain domain,
197 enum intel_display_power_domain domain, in intel_display_power_put_async() argument
200 __intel_display_power_put_async(i915, domain, wakeref); in intel_display_power_put_async()
204 enum intel_display_power_domain domain);
[all …]
/drivers/media/platform/qcom/venus/
Dhfi_parser.c28 cap->domain = VIDC_SESSION_TYPE_DEC; in init_codecs()
35 cap->domain = VIDC_SESSION_TYPE_ENC; in init_codecs()
41 u32 codecs, u32 domain, func cb, void *data, in for_each_codec() argument
49 if (cap->valid && cap->domain == domain) in for_each_codec()
51 if (cap->codec & codecs && cap->domain == domain) in for_each_codec()
66 parse_alloc_mode(struct venus_core *core, u32 codecs, u32 domain, void *data) in parse_alloc_mode() argument
81 codecs, domain, fill_buf_mode, type, 1); in parse_alloc_mode()
100 parse_profile_level(struct venus_core *core, u32 codecs, u32 domain, void *data) in parse_profile_level() argument
111 for_each_codec(core->caps, ARRAY_SIZE(core->caps), codecs, domain, in parse_profile_level()
128 parse_caps(struct venus_core *core, u32 codecs, u32 domain, void *data) in parse_caps() argument
[all …]
/drivers/gpu/drm/etnaviv/
Detnaviv_perfmon.c19 const struct etnaviv_pm_domain *domain,
41 const struct etnaviv_pm_domain *domain, in perf_reg_read() argument
44 gpu_write(gpu, domain->profile_config, signal->data); in perf_reg_read()
46 return gpu_read(gpu, domain->profile_read); in perf_reg_read()
58 const struct etnaviv_pm_domain *domain, in pipe_perf_reg_read() argument
67 value += perf_reg_read(gpu, domain, signal); in pipe_perf_reg_read()
77 const struct etnaviv_pm_domain *domain, in pipe_reg_read() argument
96 const struct etnaviv_pm_domain *domain, in hi_total_cycle_read() argument
110 const struct etnaviv_pm_domain *domain, in hi_total_idle_cycle_read() argument
479 const struct etnaviv_pm_domain *domain = NULL; in pm_domain() local
[all …]

12345678910>>...27