Searched refs:dev_iommu_priv_get (Results 1 – 20 of 20) sorted by relevance
266 struct mtk_iommu_data *data = dev_iommu_priv_get(dev); in mtk_iommu_attach_device()292 struct mtk_iommu_data *data = dev_iommu_priv_get(dev); in mtk_iommu_detach_device()389 if (!dev_iommu_priv_get(dev)) { in mtk_iommu_create_mapping()402 data = dev_iommu_priv_get(dev); in mtk_iommu_create_mapping()448 data = dev_iommu_priv_get(dev); in mtk_iommu_probe_device()459 data = dev_iommu_priv_get(dev); in mtk_iommu_probe_finalize()
176 #define has_sysmmu(dev) (dev_iommu_priv_get(dev) != NULL)677 struct exynos_iommu_owner *owner = dev_iommu_priv_get(master); in exynos_sysmmu_suspend()695 struct exynos_iommu_owner *owner = dev_iommu_priv_get(master); in exynos_sysmmu_resume()845 struct exynos_iommu_owner *owner = dev_iommu_priv_get(dev); in exynos_iommu_detach_device()883 struct exynos_iommu_owner *owner = dev_iommu_priv_get(dev); in exynos_iommu_attach_device()1244 struct exynos_iommu_owner *owner = dev_iommu_priv_get(dev); in exynos_iommu_probe_device()1270 struct exynos_iommu_owner *owner = dev_iommu_priv_get(dev); in exynos_iommu_release_device()1295 struct exynos_iommu_owner *owner = dev_iommu_priv_get(dev); in exynos_iommu_of_xlate()
116 } else if (dev_iommu_priv_get(dev) != domain) { in gart_iommu_attach_dev()134 if (dev_iommu_priv_get(dev) == domain) { in gart_iommu_detach_dev()
74 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in omap_iommu_save_ctx()104 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in omap_iommu_restore_ctx()1401 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in omap_iommu_count()1462 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in omap_iommu_attach_dev()1527 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in _omap_iommu_detach_dev()1715 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in omap_iommu_release_device()1727 struct omap_iommu_arch_data *arch_data = dev_iommu_priv_get(dev); in omap_iommu_device_group()
388 struct mtk_iommu_data *data = dev_iommu_priv_get(dev); in mtk_iommu_attach_device()408 struct mtk_iommu_data *data = dev_iommu_priv_get(dev); in mtk_iommu_detach_device()479 data = dev_iommu_priv_get(dev); in mtk_iommu_probe_device()522 if (!dev_iommu_priv_get(dev)) { in mtk_iommu_of_xlate()
469 struct viommu_endpoint *vdev = dev_iommu_priv_get(dev); in viommu_probe_endpoint()663 struct viommu_endpoint *vdev = dev_iommu_priv_get(dev); in viommu_attach_dev()822 struct viommu_endpoint *vdev = dev_iommu_priv_get(dev); in viommu_get_resv_regions()915 vdev = dev_iommu_priv_get(dev); in viommu_release_device()
839 struct rk_iommudata *data = dev_iommu_priv_get(dev); in rk_iommu_from_dev()1062 data = dev_iommu_priv_get(dev); in rk_iommu_probe_device()1076 struct rk_iommudata *data = dev_iommu_priv_get(dev); in rk_iommu_release_device()
133 return dev_iommu_priv_get(dev); in sun50i_iommu_from_dev()705 struct sun50i_iommu *iommu = dev_iommu_priv_get(dev); in sun50i_iommu_detach_device()
355 old_domain_info = dev_iommu_priv_get(dev); in attach_device()374 if (!dev_iommu_priv_get(dev)) in attach_device()
479 struct tegra_smmu *smmu = dev_iommu_priv_get(dev); in tegra_smmu_attach_dev()956 struct tegra_smmu *smmu = dev_iommu_priv_get(dev); in tegra_smmu_device_group()
91 return dev_iommu_priv_get(dev); in to_ipmmu()
596 struct msm_iommu_ctx_dev *master = dev_iommu_priv_get(dev); in insert_iommu_master()
16 return dev_iommu_priv_get(dev) != NULL; in exynos_is_iommu_available()
341 dev_data = dev_iommu_priv_get(&pdev->dev); in pdev_pri_erratum()376 if (dev_iommu_priv_get(dev)) in iommu_init_device()426 dev_data = dev_iommu_priv_get(dev); in amd_iommu_uninit_device()503 dev_data = dev_iommu_priv_get(&pdev->dev); in amd_iommu_report_rmp_hw_error()534 dev_data = dev_iommu_priv_get(&pdev->dev); in amd_iommu_report_rmp_fault()558 dev_data = dev_iommu_priv_get(&pdev->dev); in amd_iommu_report_page_fault()2126 dev_data = dev_iommu_priv_get(dev); in attach_device()2190 dev_data = dev_iommu_priv_get(dev); in detach_device()2239 if (dev_iommu_priv_get(dev)) in amd_iommu_probe_device()2524 struct iommu_dev_data *dev_data = dev_iommu_priv_get(dev); in amd_iommu_detach_device()[all …]
91 return dev_iommu_priv_get(dev); in to_iommu()592 if (!dev_iommu_priv_get(dev)) { in qcom_iommu_of_xlate()599 if (WARN_ON(qcom_iommu != dev_iommu_priv_get(dev))) { in qcom_iommu_of_xlate()
1053 struct arm_smmu_master_cfg *cfg = dev_iommu_priv_get(dev); in arm_smmu_master_alloc_smes()1161 cfg = dev_iommu_priv_get(dev); in arm_smmu_attach_dev()1454 cfg = dev_iommu_priv_get(dev); in arm_smmu_release_device()1472 struct arm_smmu_master_cfg *cfg = dev_iommu_priv_get(dev); in arm_smmu_device_group()1609 struct arm_smmu_master_cfg *cfg = dev_iommu_priv_get(dev); in arm_smmu_def_domain_type()
2177 master = dev_iommu_priv_get(dev); in arm_smmu_attach_dev()2322 if (WARN_ON_ONCE(dev_iommu_priv_get(dev))) in arm_smmu_probe_device()2389 master = dev_iommu_priv_get(dev); in arm_smmu_release_device()2511 struct arm_smmu_master *master = dev_iommu_priv_get(dev); in arm_smmu_dev_has_feature()2527 struct arm_smmu_master *master = dev_iommu_priv_get(dev); in arm_smmu_dev_feature_enabled()2551 return arm_smmu_master_enable_sva(dev_iommu_priv_get(dev)); in arm_smmu_dev_enable_feature()2565 return arm_smmu_master_disable_sva(dev_iommu_priv_get(dev)); in arm_smmu_dev_disable_feature()
615 static inline void *dev_iommu_priv_get(struct device *dev) in dev_iommu_priv_get() function
375 info = dev_iommu_priv_get(dev); in get_domain_info()821 return dev_iommu_priv_get(dev) == DEFER_DEVICE_DOMAIN_INFO; in attach_deferred()
40190 struct rk_iommudata *data = dev_iommu_priv_get(dev);40358 + struct rk_iommudata *data = dev_iommu_priv_get(dev);