Lines Matching refs:PAGE_SHIFT
80 mem->pageshift = __ffs(dev_hpa | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
91 mem->pageshift = __ffs(ua | (entries << PAGE_SHIFT)); in mm_iommu_do_alloc()
100 chunk = (1UL << (PAGE_SHIFT + MAX_ORDER - 1)) / in mm_iommu_do_alloc()
106 ret = get_user_pages(ua + (entry << PAGE_SHIFT), n, in mm_iommu_do_alloc()
134 if ((mem2->ua < (ua + (entries << PAGE_SHIFT))) && in mm_iommu_do_alloc()
136 (mem2->entries << PAGE_SHIFT)))) { in mm_iommu_do_alloc()
149 pageshift = PAGE_SHIFT; in mm_iommu_do_alloc()
153 if ((mem->pageshift > PAGE_SHIFT) && PageHuge(page)) in mm_iommu_do_alloc()
160 mem->hpas[i] = page_to_pfn(page) << PAGE_SHIFT; in mm_iommu_do_alloc()
214 page = pfn_to_page(mem->hpas[i] >> PAGE_SHIFT); in mm_iommu_unpin()
295 (mem->entries << PAGE_SHIFT))) { in mm_iommu_lookup()
314 (mem->entries << PAGE_SHIFT))) { in mm_iommu_lookup_rm()
347 const long entry = (ua - mem->ua) >> PAGE_SHIFT; in mm_iommu_ua_to_hpa()
371 const long entry = (ua - mem->ua) >> PAGE_SHIFT; in mm_iommu_ua_to_hpa_rm()
408 entry = (ua - mem->ua) >> PAGE_SHIFT; in mm_iommu_ua_mark_dirty_rm()
428 end = mem->dev_hpa + (mem->entries << PAGE_SHIFT); in mm_iommu_is_devmem()