/arch/arm/mm/ |
D | flush.c | 105 flush_pfn_alias(page_to_pfn(page), uaddr); in flush_ptrace_access() 136 flush_pfn_alias(page_to_pfn(page), in __flush_dcache_page() 167 flush_cache_page(mpnt, mpnt->vm_start + offset, page_to_pfn(page)); in __flush_dcache_aliases() 229 pfn = page_to_pfn(page); in __flush_anon_page()
|
D | copypage-v6.c | 93 set_pte_ext(TOP_PTE(from_address) + offset, pfn_pte(page_to_pfn(from), PAGE_KERNEL), 0); in v6_copy_user_highpage_aliasing() 94 set_pte_ext(TOP_PTE(to_address) + offset, pfn_pte(page_to_pfn(to), PAGE_KERNEL), 0); in v6_copy_user_highpage_aliasing() 126 set_pte_ext(TOP_PTE(to_address) + offset, pfn_pte(page_to_pfn(page), PAGE_KERNEL), 0); in v6_clear_user_highpage_aliasing()
|
D | copypage-v4mc.c | 81 set_pte_ext(TOP_PTE(0xffff8000), pfn_pte(page_to_pfn(from), minicache_pgprot), 0); in v4_mc_copy_user_highpage()
|
D | copypage-xscale.c | 103 set_pte_ext(TOP_PTE(COPYPAGE_MINICACHE), pfn_pte(page_to_pfn(from), minicache_pgprot), 0); in xscale_mc_copy_user_highpage()
|
/arch/s390/mm/ |
D | page-states.c | 64 : "a" ((page_to_pfn(page) + i) << PAGE_SHIFT), in arch_free_page() 77 : "a" ((page_to_pfn(page) + i) << PAGE_SHIFT), in arch_alloc_page()
|
/arch/sh/include/asm/ |
D | cacheflush.h | 63 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 70 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
/arch/sparc/include/asm/ |
D | cacheflush_64.h | 51 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 58 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
|
D | cacheflush_32.h | 61 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 66 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
/arch/avr32/include/asm/ |
D | page.h | 82 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT) 89 #define page_to_pfn(page) ((unsigned long)((page) - mem_map) + PHYS_PFN_OFFSET) macro
|
/arch/alpha/include/asm/ |
D | mmzone.h | 86 pfn = page_to_pfn(page) << 32; \ 104 (page_to_pfn(page) << PAGE_SHIFT)
|
/arch/m68k/include/asm/ |
D | cacheflush_mm.h | 144 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 152 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
|
D | virtconvert.h | 37 page_to_pfn(__page) << PAGE_SHIFT; \
|
D | page_mm.h | 202 pfn_to_virt(page_to_pfn(page)); \ 211 #define page_to_pfn(_page) ({ \ macro
|
D | page_no.h | 67 #define page_to_pfn(page) virt_to_pfn(page_to_virt(page)) macro
|
/arch/xtensa/include/asm/ |
D | page.h | 161 #define page_to_virt(page) __va(page_to_pfn(page) << PAGE_SHIFT) 163 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
/arch/parisc/include/asm/ |
D | cacheflush.h | 64 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 71 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
|
/arch/arm/kernel/ |
D | machine_kexec.c | 59 page_to_pfn(image->control_code_page) << PAGE_SHIFT; in machine_kexec()
|
/arch/um/include/asm/ |
D | pgalloc.h | 19 ((unsigned long long)page_to_pfn(pte) << \
|
/arch/powerpc/include/asm/ |
D | pgalloc-32.h | 26 (pmd_val(*(pmd)) = (page_to_pfn(pte) << PAGE_SHIFT) | _PMD_PRESENT)
|
/arch/ia64/include/asm/ |
D | page.h | 106 # define page_to_pfn(page) ((unsigned long) (page - vmem_map)) macro 123 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
/arch/ia64/xen/ |
D | xencomm.c | 104 return (page_to_pfn(page) << PAGE_SHIFT) | (vaddr & ~PAGE_MASK); in xencomm_vtop()
|
/arch/arm/include/asm/ |
D | pgalloc.h | 130 __pmd_populate(pmdp, page_to_pfn(ptep) << PAGE_SHIFT | _PAGE_USER_TABLE); in pmd_populate()
|
/arch/blackfin/include/asm/ |
D | page.h | 78 #define page_to_pfn(page) virt_to_pfn(page_to_virt(page)) macro
|
/arch/s390/include/asm/ |
D | page.h | 146 #define page_to_phys(page) (page_to_pfn(page) << PAGE_SHIFT)
|
/arch/sparc/mm/ |
D | iommu.c | 184 ioptex = bit_map_string_get(&iommu->usemap, npages, page_to_pfn(page)); in iommu_get_one() 193 iopte_val(*iopte) = MKIOPTE(page_to_pfn(page), IOPERM); in iommu_get_one() 383 MKIOPTE(page_to_pfn(virt_to_page(page)), ioperm_noc); in iommu_map_dma_area()
|