/arch/arm64/kernel/ |
D | paravirt.c | 33 struct pvclock_vcpu_stolen_time __rcu *kaddr; member 50 struct pvclock_vcpu_stolen_time *kaddr = NULL; in pv_steal_clock() local 62 kaddr = rcu_dereference(reg->kaddr); in pv_steal_clock() 63 if (!kaddr) { in pv_steal_clock() 68 ret = le64_to_cpu(READ_ONCE(kaddr->stolen_time)); in pv_steal_clock() 75 struct pvclock_vcpu_stolen_time *kaddr = NULL; in stolen_time_cpu_down_prepare() local 79 if (!reg->kaddr) in stolen_time_cpu_down_prepare() 82 kaddr = rcu_replace_pointer(reg->kaddr, NULL, true); in stolen_time_cpu_down_prepare() 84 memunmap(kaddr); in stolen_time_cpu_down_prepare() 91 struct pvclock_vcpu_stolen_time *kaddr = NULL; in stolen_time_cpu_online() local [all …]
|
/arch/nds32/mm/ |
D | cacheflush.c | 31 unsigned long kaddr; in flush_icache_page() local 33 kaddr = (unsigned long)kmap_atomic(page); in flush_icache_page() 34 cpu_cache_wbinval_page(kaddr, vma->vm_flags & VM_EXEC); in flush_icache_page() 35 kunmap_atomic((void *)kaddr); in flush_icache_page() 42 unsigned long kaddr; in flush_icache_user_page() local 43 kaddr = (unsigned long)kmap_atomic(page) + (addr & ~PAGE_MASK); in flush_icache_user_page() 44 flush_icache_range(kaddr, kaddr + len); in flush_icache_user_page() 45 kunmap_atomic((void *)kaddr); in flush_icache_user_page() 69 unsigned long kaddr; in update_mmu_cache() local 71 kaddr = (unsigned long)kmap_atomic(page); in update_mmu_cache() [all …]
|
/arch/m68k/sun3/ |
D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 29 j = *(volatile unsigned long *)kaddr; in dvma_page() 30 *(volatile unsigned long *)kaddr = j; in dvma_page() 32 ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); in dvma_page() 40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 44 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument 56 dvma_page(kaddr, vaddr); in dvma_map_iommu() 57 kaddr += PAGE_SIZE; in dvma_map_iommu()
|
D | sun3dvma.c | 283 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument 292 if(!kaddr || !len) { in dvma_map_align() 298 pr_debug("dvma_map request %08x bytes from %08lx\n", len, kaddr); in dvma_map_align() 299 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align() 300 kaddr &= PAGE_MASK; in dvma_map_align() 312 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align() 315 pr_crit("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, baddr, in dvma_map_align() 340 unsigned long kaddr; in dvma_malloc_align() local 350 if((kaddr = __get_free_pages(GFP_ATOMIC, get_order(len))) == 0) in dvma_malloc_align() 353 if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) { in dvma_malloc_align() [all …]
|
/arch/alpha/include/asm/ |
D | mmzone.h | 57 #define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr)) argument 64 #define LOCAL_BASE_ADDR(kaddr) \ argument 65 ((unsigned long)__va(NODE_DATA(kvaddr_to_nid(kaddr))->node_start_pfn \ 69 #define kern_addr_valid(kaddr) (0) argument 71 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 106 #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT)) argument
|
/arch/m68k/sun3x/ |
D | dvma.c | 78 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument 87 kaddr &= PAGE_MASK; in dvma_map_cpu() 92 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu() 127 __pa(kaddr), vaddr); in dvma_map_cpu() 128 set_pte(pte, pfn_pte(virt_to_pfn(kaddr), in dvma_map_cpu() 131 kaddr += PAGE_SIZE; in dvma_map_cpu() 146 inline int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument 163 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu() 168 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
|
/arch/nds32/include/asm/ |
D | memory.h | 84 #define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)) argument 85 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (… argument
|
/arch/m68k/include/asm/ |
D | page_no.h | 23 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 33 #define virt_addr_valid(kaddr) (((void *)(kaddr) >= (void *)PAGE_OFFSET) && \ argument 34 ((void *)(kaddr) < (void *)memory_end))
|
D | page_mm.h | 124 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 169 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && (void *)(kaddr) < high_me… argument
|
/arch/x86/include/asm/ |
D | page.h | 69 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument 71 extern bool __virt_addr_valid(unsigned long kaddr); 72 #define virt_addr_valid(kaddr) __virt_addr_valid((unsigned long) (kaddr)) argument
|
/arch/hexagon/include/asm/ |
D | page.h | 93 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument 99 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument 129 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
|
/arch/arc/include/asm/ |
D | page.h | 96 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 114 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument 115 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
|
/arch/arm64/mm/ |
D | flush.c | 17 void sync_icache_aliases(void *kaddr, unsigned long len) in sync_icache_aliases() argument 19 unsigned long addr = (unsigned long)kaddr; in sync_icache_aliases() 22 __clean_dcache_area_pou(kaddr, len); in sync_icache_aliases() 34 unsigned long uaddr, void *kaddr, in flush_ptrace_access() argument 38 sync_icache_aliases(kaddr, len); in flush_ptrace_access()
|
/arch/arm/include/asm/ |
D | tlbflush.h | 479 static inline void __local_flush_tlb_kernel_page(unsigned long kaddr) in __local_flush_tlb_kernel_page() argument 484 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 485 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 486 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 490 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 491 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 492 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 495 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument 499 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page() 504 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page() [all …]
|
D | memory.h | 264 #define virt_to_pfn(kaddr) \ argument 265 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \ 369 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument 370 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < … argument 371 && pfn_valid(virt_to_pfn(kaddr)))
|
/arch/mips/include/asm/ |
D | page.h | 249 #define virt_to_pfn(kaddr) PFN_DOWN(virt_to_phys((void *)(kaddr))) argument 250 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument 252 extern bool __virt_addr_valid(const volatile void *kaddr); 253 #define virt_addr_valid(kaddr) \ argument 254 __virt_addr_valid((const volatile void *) (kaddr))
|
/arch/sh/mm/ |
D | uncached.c | 24 int virt_addr_uncached(unsigned long kaddr) in virt_addr_uncached() argument 26 return (kaddr >= uncached_start) && (kaddr < uncached_end); in virt_addr_uncached()
|
/arch/csky/include/asm/ |
D | page.h | 37 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 40 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \ argument 41 (void *)(kaddr) < high_memory)
|
/arch/arm/mm/ |
D | copypage-v4wt.c | 62 void *ptr, *kaddr = kmap_atomic(page); in v4wt_clear_user_highpage() local 77 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wt_clear_user_highpage() 79 kunmap_atomic(kaddr); in v4wt_clear_user_highpage()
|
D | copypage-fa.c | 57 void *ptr, *kaddr = kmap_atomic(page); in fa_clear_user_highpage() local 74 : "0" (kaddr), "I" (PAGE_SIZE / 32) in fa_clear_user_highpage() 76 kunmap_atomic(kaddr); in fa_clear_user_highpage()
|
D | copypage-xsc3.c | 81 void *ptr, *kaddr = kmap_atomic(page); in xsc3_mc_clear_user_highpage() local 94 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xsc3_mc_clear_user_highpage() 96 kunmap_atomic(kaddr); in xsc3_mc_clear_user_highpage()
|
D | copypage-feroceon.c | 80 void *ptr, *kaddr = kmap_atomic(page); in feroceon_clear_user_highpage() local 98 : "0" (kaddr), "I" (PAGE_SIZE / 32) in feroceon_clear_user_highpage() 100 kunmap_atomic(kaddr); in feroceon_clear_user_highpage()
|
D | copypage-v4wb.c | 67 void *ptr, *kaddr = kmap_atomic(page); in v4wb_clear_user_highpage() local 84 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wb_clear_user_highpage() 86 kunmap_atomic(kaddr); in v4wb_clear_user_highpage()
|
D | copypage-v4mc.c | 87 void *ptr, *kaddr = kmap_atomic(page); in v4_mc_clear_user_highpage() local 103 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4_mc_clear_user_highpage() 105 kunmap_atomic(kaddr); in v4_mc_clear_user_highpage()
|
/arch/openrisc/include/asm/ |
D | page.h | 75 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument 85 #define virt_addr_valid(kaddr) (pfn_valid(virt_to_pfn(kaddr))) argument
|