/arch/m68k/sun3/ |
D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 29 j = *(volatile unsigned long *)kaddr; in dvma_page() 30 *(volatile unsigned long *)kaddr = j; in dvma_page() 32 ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); in dvma_page() 41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 45 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument 57 dvma_page(kaddr, vaddr); in dvma_map_iommu() 58 kaddr += PAGE_SIZE; in dvma_map_iommu()
|
D | sun3dvma.c | 280 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument 289 if(!kaddr || !len) { in dvma_map_align() 297 len, kaddr); in dvma_map_align() 299 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align() 300 kaddr &= PAGE_MASK; in dvma_map_align() 312 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align() 315 printk("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, baddr, len); in dvma_map_align() 339 unsigned long kaddr; in dvma_malloc_align() local 351 if((kaddr = __get_free_pages(GFP_ATOMIC, get_order(len))) == 0) in dvma_malloc_align() 354 if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) { in dvma_malloc_align() [all …]
|
/arch/tile/include/asm/ |
D | page.h | 250 unsigned long kaddr = (unsigned long)_kaddr; in kaddr_to_pfn() local 251 return pbase_map[kaddr >> HPAGE_SHIFT] + in kaddr_to_pfn() 252 ((kaddr & (HPAGE_SIZE - 1)) >> PAGE_SHIFT); in kaddr_to_pfn() 260 static inline phys_addr_t virt_to_phys(const volatile void *kaddr) in virt_to_phys() argument 262 unsigned long pfn = kaddr_to_pfn(kaddr); in virt_to_phys() 264 ((unsigned long)kaddr & (PAGE_SIZE-1)); in virt_to_phys() 273 static inline int virt_addr_valid(const volatile void *kaddr) in virt_addr_valid() argument 276 return ((unsigned long)kaddr >= PAGE_OFFSET && kaddr < high_memory); in virt_addr_valid() 281 static inline unsigned long kaddr_to_pfn(const volatile void *kaddr) in kaddr_to_pfn() argument 283 return ((unsigned long)kaddr - PAGE_OFFSET) >> PAGE_SHIFT; in kaddr_to_pfn() [all …]
|
/arch/alpha/include/asm/ |
D | mmzone.h | 58 #define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr)) argument 65 #define LOCAL_BASE_ADDR(kaddr) \ argument 66 ((unsigned long)__va(NODE_DATA(kvaddr_to_nid(kaddr))->node_start_pfn \ 70 #define kern_addr_valid(kaddr) (0) argument 72 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 108 #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT)) argument
|
/arch/arm/include/asm/ |
D | memory.h | 151 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 170 #define virt_to_pfn(kaddr) \ argument 171 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \ 245 #define virt_to_pfn(kaddr) \ argument 246 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \ 331 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument 332 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < … argument 333 && pfn_valid(virt_to_pfn(kaddr)))
|
D | tlbflush.h | 478 static inline void __local_flush_tlb_kernel_page(unsigned long kaddr) in __local_flush_tlb_kernel_page() argument 483 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 484 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 485 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 489 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 490 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 491 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 494 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument 498 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page() 503 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page() [all …]
|
/arch/arm64/include/asm/ |
D | memory.h | 203 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 204 #define _virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument 206 #define _virt_addr_is_linear(kaddr) (((u64)(kaddr)) >= PAGE_OFFSET) argument 207 #define virt_addr_valid(kaddr) (_virt_addr_is_linear(kaddr) && \ argument 208 _virt_addr_valid(kaddr))
|
/arch/m68k/sun3x/ |
D | dvma.c | 82 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument 89 kaddr &= PAGE_MASK; in dvma_map_cpu() 96 kaddr, vaddr); in dvma_map_cpu() 131 __pa(kaddr), vaddr); in dvma_map_cpu() 133 set_pte(pte, pfn_pte(virt_to_pfn(kaddr), in dvma_map_cpu() 136 kaddr += PAGE_SIZE; in dvma_map_cpu() 151 inline int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument 167 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu() 172 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
|
/arch/x86/include/asm/ |
D | page.h | 63 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 65 extern bool __virt_addr_valid(unsigned long kaddr); 66 #define virt_addr_valid(kaddr) __virt_addr_valid((unsigned long) (kaddr)) argument
|
/arch/unicore32/include/asm/ |
D | memory.h | 103 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 104 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && \ argument 105 (unsigned long)(kaddr) < (unsigned long)high_memory)
|
D | tlbflush.h | 106 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument 111 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page() 114 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page() 118 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
|
/arch/m68k/include/asm/ |
D | page_no.h | 25 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 35 #define virt_addr_valid(kaddr) (((void *)(kaddr) >= (void *)PAGE_OFFSET) && \ argument 36 ((void *)(kaddr) < (void *)memory_end))
|
/arch/hexagon/include/asm/ |
D | page.h | 106 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument 113 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument 143 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
|
/arch/arm64/mm/ |
D | flush.c | 37 static void sync_icache_aliases(void *kaddr, unsigned long len) in sync_icache_aliases() argument 39 unsigned long addr = (unsigned long)kaddr; in sync_icache_aliases() 42 __clean_dcache_area_pou(kaddr, len); in sync_icache_aliases() 50 unsigned long uaddr, void *kaddr, in flush_ptrace_access() argument 54 sync_icache_aliases(kaddr, len); in flush_ptrace_access()
|
/arch/sh/mm/ |
D | uncached.c | 23 int virt_addr_uncached(unsigned long kaddr) in virt_addr_uncached() argument 25 return (kaddr >= uncached_start) && (kaddr < uncached_end); in virt_addr_uncached()
|
/arch/powerpc/include/asm/ |
D | page.h | 132 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 133 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument 143 #define virt_addr_valid(kaddr) (REGION_ID(kaddr) == KERNEL_REGION_ID && \ argument 144 pfn_valid(virt_to_pfn(kaddr))) 146 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
|
/arch/arm/mm/ |
D | copypage-v4wt.c | 64 void *ptr, *kaddr = kmap_atomic(page); in v4wt_clear_user_highpage() local 79 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wt_clear_user_highpage() 81 kunmap_atomic(kaddr); in v4wt_clear_user_highpage()
|
D | copypage-fa.c | 60 void *ptr, *kaddr = kmap_atomic(page); in fa_clear_user_highpage() local 77 : "0" (kaddr), "I" (PAGE_SIZE / 32) in fa_clear_user_highpage() 79 kunmap_atomic(kaddr); in fa_clear_user_highpage()
|
D | copypage-v4wb.c | 69 void *ptr, *kaddr = kmap_atomic(page); in v4wb_clear_user_highpage() local 86 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wb_clear_user_highpage() 88 kunmap_atomic(kaddr); in v4wb_clear_user_highpage()
|
D | copypage-feroceon.c | 83 void *ptr, *kaddr = kmap_atomic(page); in feroceon_clear_user_highpage() local 101 : "0" (kaddr), "I" (PAGE_SIZE / 32) in feroceon_clear_user_highpage() 103 kunmap_atomic(kaddr); in feroceon_clear_user_highpage()
|
D | copypage-xsc3.c | 84 void *ptr, *kaddr = kmap_atomic(page); in xsc3_mc_clear_user_highpage() local 97 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xsc3_mc_clear_user_highpage() 99 kunmap_atomic(kaddr); in xsc3_mc_clear_user_highpage()
|
D | copypage-v4mc.c | 90 void *ptr, *kaddr = kmap_atomic(page); in v4_mc_clear_user_highpage() local 106 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4_mc_clear_user_highpage() 108 kunmap_atomic(kaddr); in v4_mc_clear_user_highpage()
|
D | copypage-xscale.c | 111 void *ptr, *kaddr = kmap_atomic(page); in xscale_mc_clear_user_highpage() local 126 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xscale_mc_clear_user_highpage() 128 kunmap_atomic(kaddr); in xscale_mc_clear_user_highpage()
|
/arch/cris/include/asm/ |
D | page.h | 53 #define virt_to_page(kaddr) (mem_map + (((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT)) argument 54 #define virt_addr_valid(kaddr) pfn_valid((unsigned)(kaddr) >> PAGE_SHIFT) argument
|
/arch/m32r/include/asm/ |
D | page.h | 78 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 79 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
|