Home
last modified time | relevance | path

Searched refs:kaddr (Results 1 – 25 of 73) sorted by relevance

123

/arch/nds32/mm/
Dcacheflush.c31 unsigned long kaddr; in flush_icache_page() local
33 kaddr = (unsigned long)kmap_atomic(page); in flush_icache_page()
34 cpu_cache_wbinval_page(kaddr, vma->vm_flags & VM_EXEC); in flush_icache_page()
35 kunmap_atomic((void *)kaddr); in flush_icache_page()
43 unsigned long kaddr; in flush_icache_user_range() local
44 kaddr = (unsigned long)kmap_atomic(page) + (addr & ~PAGE_MASK); in flush_icache_user_range()
45 flush_icache_range(kaddr, kaddr + len); in flush_icache_user_range()
46 kunmap_atomic((void *)kaddr); in flush_icache_user_range()
70 unsigned long kaddr; in update_mmu_cache() local
72 kaddr = (unsigned long)kmap_atomic(page); in update_mmu_cache()
[all …]
/arch/m68k/sun3/
Ddvma.c24 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument
30 j = *(volatile unsigned long *)kaddr; in dvma_page()
31 *(volatile unsigned long *)kaddr = j; in dvma_page()
33 ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); in dvma_page()
41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
45 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
57 dvma_page(kaddr, vaddr); in dvma_map_iommu()
58 kaddr += PAGE_SIZE; in dvma_map_iommu()
Dsun3dvma.c284 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument
293 if(!kaddr || !len) { in dvma_map_align()
299 pr_debug("dvma_map request %08x bytes from %08lx\n", len, kaddr); in dvma_map_align()
300 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align()
301 kaddr &= PAGE_MASK; in dvma_map_align()
313 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align()
316 pr_crit("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, baddr, in dvma_map_align()
341 unsigned long kaddr; in dvma_malloc_align() local
351 if((kaddr = __get_free_pages(GFP_ATOMIC, get_order(len))) == 0) in dvma_malloc_align()
354 if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) { in dvma_malloc_align()
[all …]
/arch/alpha/include/asm/
Dmmzone.h59 #define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr)) argument
66 #define LOCAL_BASE_ADDR(kaddr) \ argument
67 ((unsigned long)__va(NODE_DATA(kvaddr_to_nid(kaddr))->node_start_pfn \
71 #define kern_addr_valid(kaddr) (0) argument
73 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
109 #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT)) argument
/arch/csky/abiv2/
Dcacheflush.c23 unsigned long kaddr; in flush_icache_user_range() local
25 kaddr = (unsigned long) kmap_atomic(page) + (vaddr & ~PAGE_MASK); in flush_icache_user_range()
27 cache_wbinv_range(kaddr, kaddr + len); in flush_icache_user_range()
29 kunmap_atomic((void *)kaddr); in flush_icache_user_range()
/arch/m68k/sun3x/
Ddvma.c79 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument
86 kaddr &= PAGE_MASK; in dvma_map_cpu()
91 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu()
124 __pa(kaddr), vaddr); in dvma_map_cpu()
125 set_pte(pte, pfn_pte(virt_to_pfn(kaddr), in dvma_map_cpu()
128 kaddr += PAGE_SIZE; in dvma_map_cpu()
143 inline int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
160 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu()
165 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
/arch/s390/include/asm/
Dpage.h164 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
168 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
171 #define phys_to_pfn(kaddr) ((kaddr) >> PAGE_SHIFT) argument
174 #define phys_to_page(kaddr) pfn_to_page(phys_to_pfn(kaddr)) argument
177 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
/arch/x86/include/asm/
Dpage.h69 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
71 extern bool __virt_addr_valid(unsigned long kaddr);
72 #define virt_addr_valid(kaddr) __virt_addr_valid((unsigned long) (kaddr)) argument
/arch/nds32/include/asm/
Dmemory.h82 #define virt_to_page(kaddr) (pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)) argument
83 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < (… argument
/arch/unicore32/include/asm/
Dmemory.h94 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
95 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && \ argument
96 (unsigned long)(kaddr) < (unsigned long)high_memory)
Dtlbflush.h103 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument
108 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
111 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
115 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
/arch/m68k/include/asm/
Dpage_no.h23 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
33 #define virt_addr_valid(kaddr) (((void *)(kaddr) >= (void *)PAGE_OFFSET) && \ argument
34 ((void *)(kaddr) < (void *)memory_end))
/arch/arm64/mm/
Dflush.c17 void sync_icache_aliases(void *kaddr, unsigned long len) in sync_icache_aliases() argument
19 unsigned long addr = (unsigned long)kaddr; in sync_icache_aliases()
22 __clean_dcache_area_pou(kaddr, len); in sync_icache_aliases()
34 unsigned long uaddr, void *kaddr, in flush_ptrace_access() argument
38 sync_icache_aliases(kaddr, len); in flush_ptrace_access()
/arch/arc/include/asm/
Dpage.h96 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
114 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
115 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
/arch/hexagon/include/asm/
Dpage.h93 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument
100 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
130 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
/arch/arm/include/asm/
Dtlbflush.h479 static inline void __local_flush_tlb_kernel_page(unsigned long kaddr) in __local_flush_tlb_kernel_page() argument
484 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
485 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
486 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
490 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
491 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
492 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
495 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument
499 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page()
504 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page()
[all …]
Dmemory.h264 #define virt_to_pfn(kaddr) \ argument
265 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \
369 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
370 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < … argument
371 && pfn_valid(virt_to_pfn(kaddr)))
/arch/sh/mm/
Duncached.c24 int virt_addr_uncached(unsigned long kaddr) in virt_addr_uncached() argument
26 return (kaddr >= uncached_start) && (kaddr < uncached_end); in virt_addr_uncached()
/arch/mips/include/asm/
Dpage.h249 #define virt_to_pfn(kaddr) PFN_DOWN(virt_to_phys((void *)(kaddr))) argument
250 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
252 extern bool __virt_addr_valid(const volatile void *kaddr);
253 #define virt_addr_valid(kaddr) \ argument
254 __virt_addr_valid((const volatile void *) (kaddr))
/arch/csky/include/asm/
Dpage.h37 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
40 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \ argument
41 (void *)(kaddr) < high_memory)
/arch/arm/mm/
Dcopypage-v4wt.c62 void *ptr, *kaddr = kmap_atomic(page); in v4wt_clear_user_highpage() local
77 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wt_clear_user_highpage()
79 kunmap_atomic(kaddr); in v4wt_clear_user_highpage()
Dcopypage-fa.c57 void *ptr, *kaddr = kmap_atomic(page); in fa_clear_user_highpage() local
74 : "0" (kaddr), "I" (PAGE_SIZE / 32) in fa_clear_user_highpage()
76 kunmap_atomic(kaddr); in fa_clear_user_highpage()
Dcopypage-xsc3.c81 void *ptr, *kaddr = kmap_atomic(page); in xsc3_mc_clear_user_highpage() local
94 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xsc3_mc_clear_user_highpage()
96 kunmap_atomic(kaddr); in xsc3_mc_clear_user_highpage()
Dcopypage-v4wb.c67 void *ptr, *kaddr = kmap_atomic(page); in v4wb_clear_user_highpage() local
84 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wb_clear_user_highpage()
86 kunmap_atomic(kaddr); in v4wb_clear_user_highpage()
/arch/openrisc/include/asm/
Dpage.h75 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
85 #define virt_addr_valid(kaddr) (pfn_valid(virt_to_pfn(kaddr))) argument

123