/drivers/iommu/ |
D | tegra-gart.c | 60 #define for_each_gart_pte(gart, iova) \ argument 66 unsigned long iova, unsigned long pte) in gart_set_pte() 73 unsigned long iova) in gart_read_pte() 85 unsigned long iova; in do_gart_setup() local 95 unsigned long iova, size_t bytes) in gart_iova_range_invalid() 101 static inline bool gart_pte_valid(struct gart_device *gart, unsigned long iova) in gart_pte_valid() 167 static inline int __gart_iommu_map(struct gart_device *gart, unsigned long iova, in __gart_iommu_map() 180 static int gart_iommu_map(struct iommu_domain *domain, unsigned long iova, in gart_iommu_map() 197 unsigned long iova) in __gart_iommu_unmap() 209 static size_t gart_iommu_unmap(struct iommu_domain *domain, unsigned long iova, in gart_iommu_unmap() [all …]
|
D | iova.c | 154 iova_insert_rbtree(struct rb_root *root, struct iova *iova, in iova_insert_rbtree() 240 void free_iova_mem(struct iova *iova) in free_iova_mem() 325 struct iova *iova = rb_entry(node, struct iova, node); in private_find_iova() local 338 static void private_free_iova(struct iova_domain *iovad, struct iova *iova) in private_free_iova() 356 struct iova *iova; in find_iova() local 373 __free_iova(struct iova_domain *iovad, struct iova *iova) in __free_iova() 393 struct iova *iova = find_iova(iovad, pfn); in free_iova() local 595 struct iova *iova, *tmp; in put_iova_domain() local 608 struct iova *iova = rb_entry(node, struct iova, node); in __is_range_overlap() local 618 struct iova *iova; in alloc_and_init_iova() local [all …]
|
D | exynos-iommu.c | 99 #define section_offs(iova) (iova & (SECT_SIZE - 1)) argument 101 #define lpage_offs(iova) (iova & (LPAGE_SIZE - 1)) argument 103 #define spage_offs(iova) (iova & (SPAGE_SIZE - 1)) argument 108 static u32 lv1ent_offset(sysmmu_iova_t iova) in lv1ent_offset() 113 static u32 lv2ent_offset(sysmmu_iova_t iova) in lv2ent_offset() 183 static sysmmu_pte_t *section_entry(sysmmu_pte_t *pgtable, sysmmu_iova_t iova) in section_entry() 188 static sysmmu_pte_t *page_entry(sysmmu_pte_t *sent, sysmmu_iova_t iova) in page_entry() 316 sysmmu_iova_t iova, unsigned int num_inv) in __sysmmu_tlb_invalidate_entry() 515 sysmmu_iova_t iova) in sysmmu_tlb_invalidate_flpdcache() 535 sysmmu_iova_t iova, size_t size) in sysmmu_tlb_invalidate_entry() [all …]
|
D | tegra-smmu.c | 146 static unsigned int iova_pd_index(unsigned long iova) in iova_pd_index() 151 static unsigned int iova_pt_index(unsigned long iova) in iova_pt_index() 213 unsigned long iova) in smmu_flush_tlb_section() 228 unsigned long iova) in smmu_flush_tlb_group() 520 static void tegra_smmu_set_pde(struct tegra_smmu_as *as, unsigned long iova, in tegra_smmu_set_pde() 541 static u32 *tegra_smmu_pte_offset(struct page *pt_page, unsigned long iova) in tegra_smmu_pte_offset() 548 static u32 *tegra_smmu_pte_lookup(struct tegra_smmu_as *as, unsigned long iova, in tegra_smmu_pte_lookup() 566 static u32 *as_get_pte(struct tegra_smmu_as *as, dma_addr_t iova, in as_get_pte() 609 static void tegra_smmu_pte_get_use(struct tegra_smmu_as *as, unsigned long iova) in tegra_smmu_pte_get_use() 616 static void tegra_smmu_pte_put_use(struct tegra_smmu_as *as, unsigned long iova) in tegra_smmu_pte_put_use() [all …]
|
D | io-pgtable-arm-v7s.c | 413 unsigned long iova, phys_addr_t paddr, int prot, in arm_v7s_init_pte() 473 static int __arm_v7s_map(struct arm_v7s_io_pgtable *data, unsigned long iova, in __arm_v7s_map() 520 static int arm_v7s_map(struct io_pgtable_ops *ops, unsigned long iova, in arm_v7s_map() 568 unsigned long iova, int idx, int lvl, in arm_v7s_split_cont() 595 unsigned long iova, size_t size, in arm_v7s_split_blk_unmap() 640 unsigned long iova, size_t size, int lvl, in __arm_v7s_unmap() 717 static size_t arm_v7s_unmap(struct io_pgtable_ops *ops, unsigned long iova, in arm_v7s_unmap() 729 unsigned long iova) in arm_v7s_iova_to_phys() 856 static void dummy_tlb_flush(unsigned long iova, size_t size, size_t granule, in dummy_tlb_flush() 864 unsigned long iova, size_t granule, void *cookie) in dummy_tlb_add_page() [all …]
|
D | rockchip-iommu.c | 260 static u32 rk_iova_dte_index(dma_addr_t iova) in rk_iova_dte_index() 265 static u32 rk_iova_pte_index(dma_addr_t iova) in rk_iova_pte_index() 270 static u32 rk_iova_page_offset(dma_addr_t iova) in rk_iova_page_offset() 307 dma_addr_t iova; in rk_iommu_zap_lines() local 473 static void log_iova(struct rk_iommu *iommu, int index, dma_addr_t iova) in log_iova() 525 dma_addr_t iova; in rk_iommu_irq() local 590 dma_addr_t iova) in rk_iommu_iova_to_phys() 618 dma_addr_t iova, size_t size) in rk_iommu_zap_iova() 647 dma_addr_t iova, size_t size) in rk_iommu_zap_iova_first_last() 656 dma_addr_t iova) in rk_dte_get_page_table() [all …]
|
D | qcom_iommu.c | 147 static void qcom_iommu_tlb_inv_range_nosync(unsigned long iova, size_t size, in qcom_iommu_tlb_inv_range_nosync() 168 static void qcom_iommu_tlb_flush_walk(unsigned long iova, size_t size, in qcom_iommu_tlb_flush_walk() 175 static void qcom_iommu_tlb_flush_leaf(unsigned long iova, size_t size, in qcom_iommu_tlb_flush_leaf() 183 unsigned long iova, size_t granule, in qcom_iommu_tlb_add_page() 200 u64 iova; in qcom_iommu_fault() local 425 static int qcom_iommu_map(struct iommu_domain *domain, unsigned long iova, in qcom_iommu_map() 442 static size_t qcom_iommu_unmap(struct iommu_domain *domain, unsigned long iova, in qcom_iommu_unmap() 487 dma_addr_t iova) in qcom_iommu_iova_to_phys()
|
D | io-pgtable-arm.c | 322 unsigned long iova, phys_addr_t paddr, in arm_lpae_init_pte() 382 static int __arm_lpae_map(struct arm_lpae_io_pgtable *data, unsigned long iova, in __arm_lpae_map() 478 static int arm_lpae_map(struct io_pgtable_ops *ops, unsigned long iova, in arm_lpae_map() 546 unsigned long iova, size_t size, in arm_lpae_split_blk_unmap() 600 unsigned long iova, size_t size, int lvl, in __arm_lpae_unmap() 651 static size_t arm_lpae_unmap(struct io_pgtable_ops *ops, unsigned long iova, in arm_lpae_unmap() 665 unsigned long iova) in arm_lpae_iova_to_phys() 1107 static void dummy_tlb_flush(unsigned long iova, size_t size, size_t granule, in dummy_tlb_flush() 1115 unsigned long iova, size_t granule, void *cookie) in dummy_tlb_add_page() 1154 unsigned long iova; in arm_lpae_run_tests() local
|
D | msm_iommu.c | 139 static void __flush_iotlb_range(unsigned long iova, size_t size, in __flush_iotlb_range() 171 static void __flush_iotlb_walk(unsigned long iova, size_t size, in __flush_iotlb_walk() 177 static void __flush_iotlb_leaf(unsigned long iova, size_t size, in __flush_iotlb_leaf() 184 unsigned long iova, size_t granule, void *cookie) in __flush_iotlb_page() 506 static int msm_iommu_map(struct iommu_domain *domain, unsigned long iova, in msm_iommu_map() 520 static size_t msm_iommu_unmap(struct iommu_domain *domain, unsigned long iova, in msm_iommu_unmap()
|
D | mtk_iommu.c | 176 static void mtk_iommu_tlb_add_flush_nosync(unsigned long iova, size_t size, in mtk_iommu_tlb_add_flush_nosync() 219 static void mtk_iommu_tlb_flush_walk(unsigned long iova, size_t size, in mtk_iommu_tlb_flush_walk() 231 static void mtk_iommu_tlb_flush_leaf(unsigned long iova, size_t size, in mtk_iommu_tlb_flush_leaf() 244 unsigned long iova, size_t granule, in mtk_iommu_tlb_flush_page_nosync() 429 static int mtk_iommu_map(struct iommu_domain *domain, unsigned long iova, in mtk_iommu_map() 449 unsigned long iova, size_t size, in mtk_iommu_unmap() 480 dma_addr_t iova) in mtk_iommu_iova_to_phys()
|
D | arm-smmu.c | 329 static void arm_smmu_tlb_inv_range_s1(unsigned long iova, size_t size, in arm_smmu_tlb_inv_range_s1() 359 static void arm_smmu_tlb_inv_range_s2(unsigned long iova, size_t size, in arm_smmu_tlb_inv_range_s2() 386 static void arm_smmu_tlb_inv_vmid_nosync(unsigned long iova, size_t size, in arm_smmu_tlb_inv_vmid_nosync() 398 static void arm_smmu_tlb_inv_walk(unsigned long iova, size_t size, in arm_smmu_tlb_inv_walk() 408 static void arm_smmu_tlb_inv_leaf(unsigned long iova, size_t size, in arm_smmu_tlb_inv_leaf() 419 unsigned long iova, size_t granule, in arm_smmu_tlb_add_page() 464 unsigned long iova; in arm_smmu_context_fault() local 1203 static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova, in arm_smmu_map() 1220 static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long iova, in arm_smmu_unmap() 1263 dma_addr_t iova) in arm_smmu_iova_to_phys_hard() [all …]
|
D | s390-iommu.c | 267 static int s390_iommu_map(struct iommu_domain *domain, unsigned long iova, in s390_iommu_map() 286 dma_addr_t iova) in s390_iommu_iova_to_phys() 317 unsigned long iova, size_t size, in s390_iommu_unmap()
|
/drivers/vfio/ |
D | vfio_iommu_type1.c | 85 dma_addr_t iova; /* Device address */ member 112 dma_addr_t iova; /* Device address */ member 119 dma_addr_t iova; member 180 static struct vfio_pfn *vfio_find_vpfn(struct vfio_dma *dma, dma_addr_t iova) in vfio_find_vpfn() 224 static int vfio_add_to_pfn_list(struct vfio_dma *dma, dma_addr_t iova, in vfio_add_to_pfn_list() 248 unsigned long iova) in vfio_iova_get_vfio_pfn() 404 dma_addr_t iova = vaddr - dma->vaddr + dma->iova; in vfio_pin_pages_remote() local 476 static long vfio_unpin_pages_remote(struct vfio_dma *dma, dma_addr_t iova, in vfio_unpin_pages_remote() 524 static int vfio_unpin_page_external(struct vfio_dma *dma, dma_addr_t iova, in vfio_unpin_page_external() 575 dma_addr_t iova; in vfio_iommu_type1_pin_pages() local [all …]
|
/drivers/fpga/ |
D | dfl-afu-dma-region.c | 134 u64 iova, u64 size) in dma_region_check_iova() 254 afu_dma_region_find(struct dfl_feature_platform_data *pdata, u64 iova, u64 size) in afu_dma_region_find() 294 afu_dma_region_find_iova(struct dfl_feature_platform_data *pdata, u64 iova) in afu_dma_region_find_iova() 311 u64 user_addr, u64 length, u64 *iova) in afu_dma_map_region() 393 int afu_dma_unmap_region(struct dfl_feature_platform_data *pdata, u64 iova) in afu_dma_unmap_region()
|
/drivers/infiniband/sw/rxe/ |
D | rxe_mr.c | 54 int mem_check_range(struct rxe_mem *mem, u64 iova, size_t length) in mem_check_range() 161 u64 length, u64 iova, int access, struct ib_udata *udata, in rxe_mem_init_user() 265 u64 iova, in lookup_iova() 303 void *iova_to_vaddr(struct rxe_mem *mem, u64 iova, int length) in iova_to_vaddr() 344 int rxe_mem_copy(struct rxe_mem *mem, u64 iova, void *addr, int length, in rxe_mem_copy() 449 u64 iova; in copy_data() local 592 u64 *page, int num_pages, u64 iova) in rxe_mem_map_pages()
|
/drivers/gpu/drm/panfrost/ |
D | panfrost_mmu.c | 55 u64 iova, size_t size) in lock_region() 84 u64 iova, size_t size, u32 op) in mmu_hw_do_operation_locked() 101 u64 iova, size_t size, u32 op) in mmu_hw_do_operation() 229 u64 iova, size_t size) in panfrost_mmu_flush_range() 244 u64 iova, int prot, struct sg_table *sgt) in mmu_map_sg() 303 u64 iova = mapping->mmnode.start << PAGE_SHIFT; in panfrost_mmu_unmap() local 339 static void mmu_tlb_flush_walk(unsigned long iova, size_t size, size_t granule, in mmu_tlb_flush_walk() 345 static void mmu_tlb_flush_leaf(unsigned long iova, size_t size, size_t granule, in mmu_tlb_flush_leaf()
|
/drivers/staging/media/tegra-vde/ |
D | iommu.c | 24 struct iova *iova; in tegra_vde_iommu_map() local 51 void tegra_vde_iommu_unmap(struct tegra_vde *vde, struct iova *iova) in tegra_vde_iommu_unmap() 64 struct iova *iova; in tegra_vde_iommu_init() local
|
/drivers/staging/media/ipu3/ |
D | ipu3-mmu.c | 154 static inline void address_to_pte_idx(unsigned long iova, u32 *l1pt_idx, in address_to_pte_idx() 208 static int __imgu_mmu_map(struct imgu_mmu *mmu, unsigned long iova, in __imgu_mmu_map() 249 int imgu_mmu_map(struct imgu_mmu_info *info, unsigned long iova, in imgu_mmu_map() 297 size_t imgu_mmu_map_sg(struct imgu_mmu_info *info, unsigned long iova, in imgu_mmu_map_sg() 337 unsigned long iova, size_t size) in __imgu_mmu_unmap() 377 size_t imgu_mmu_unmap(struct imgu_mmu_info *info, unsigned long iova, in imgu_mmu_unmap()
|
D | ipu3-dmamap.c | 101 struct iova *iova; in imgu_dmamap_alloc() local 162 struct iova *iova; in imgu_dmamap_unmap() local 203 struct iova *iova; in imgu_dmamap_map_sg() local
|
/drivers/gpu/drm/etnaviv/ |
D | etnaviv_mmu.c | 17 unsigned long iova, size_t size) in etnaviv_context_unmap() 40 unsigned long iova, phys_addr_t paddr, in etnaviv_context_map() 72 static int etnaviv_iommu_map(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_map() 109 static void etnaviv_iommu_unmap(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_unmap() 245 u32 iova; in etnaviv_iommu_map_gem() local
|
D | etnaviv_iommu.c | 48 unsigned long iova, phys_addr_t paddr, in etnaviv_iommuv1_map() 63 unsigned long iova, size_t size) in etnaviv_iommuv1_unmap()
|
/drivers/gpu/drm/msm/ |
D | msm_iommu.c | 17 unsigned long iova, int flags, void *arg) in msm_fault_handler() 42 static int msm_iommu_map(struct msm_mmu *mmu, uint64_t iova, in msm_iommu_map() 54 static int msm_iommu_unmap(struct msm_mmu *mmu, uint64_t iova, unsigned len) in msm_iommu_unmap()
|
D | msm_gem_submit.c | 258 uint64_t iova; in submit_pin_objects() local 283 struct msm_gem_object **obj, uint64_t *iova, bool *valid) in submit_bo() 333 uint64_t iova; in submit_reloc() local 509 uint64_t iova; in msm_ioctl_gem_submit() local
|
/drivers/s390/cio/ |
D | vfio_ccw_cp.c | 58 static int pfn_array_alloc(struct pfn_array *pa, u64 iova, unsigned int len) in pfn_array_alloc() 133 static bool pfn_array_iova_pinned(struct pfn_array *pa, unsigned long iova) in pfn_array_iova_pinned() 193 void *to, u64 iova, in copy_from_iova() 370 static int ccwchain_calc_length(u64 iova, struct channel_program *cp) in ccwchain_calc_length() 512 u64 iova; in ccwchain_fetch_direct() local 851 bool cp_iova_pinned(struct channel_program *cp, u64 iova) in cp_iova_pinned()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_mr.c | 430 u64 iova, u64 total_size, u32 access, struct mthca_mr *mr) in mthca_mr_alloc() 521 int list_len, u64 iova, u64 total_size, in mthca_mr_alloc_phys() 690 int list_len, u64 iova) in mthca_check_fmr() 718 int list_len, u64 iova) in mthca_tavor_map_phys_fmr() 759 int list_len, u64 iova) in mthca_arbel_map_phys_fmr()
|