Home
last modified time | relevance | path

Searched refs:kaddr (Results 1 – 25 of 78) sorted by relevance

1234

/arch/m68k/sun3/
Ddvma.c23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument
29 j = *(volatile unsigned long *)kaddr; in dvma_page()
30 *(volatile unsigned long *)kaddr = j; in dvma_page()
32 ptep = pfn_pte(virt_to_pfn(kaddr), PAGE_KERNEL); in dvma_page()
41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
45 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
57 dvma_page(kaddr, vaddr); in dvma_map_iommu()
58 kaddr += PAGE_SIZE; in dvma_map_iommu()
Dsun3dvma.c280 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument
289 if(!kaddr || !len) { in dvma_map_align()
297 len, kaddr); in dvma_map_align()
299 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align()
300 kaddr &= PAGE_MASK; in dvma_map_align()
312 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align()
315 printk("dvma_map failed kaddr %lx baddr %lx len %x\n", kaddr, baddr, len); in dvma_map_align()
339 unsigned long kaddr; in dvma_malloc_align() local
351 if((kaddr = __get_free_pages(GFP_ATOMIC, get_order(len))) == 0) in dvma_malloc_align()
354 if((baddr = (unsigned long)dvma_map_align(kaddr, len, align)) == 0) { in dvma_malloc_align()
[all …]
/arch/tile/include/asm/
Dpage.h250 unsigned long kaddr = (unsigned long)_kaddr; in kaddr_to_pfn() local
251 return pbase_map[kaddr >> HPAGE_SHIFT] + in kaddr_to_pfn()
252 ((kaddr & (HPAGE_SIZE - 1)) >> PAGE_SHIFT); in kaddr_to_pfn()
260 static inline phys_addr_t virt_to_phys(const volatile void *kaddr) in virt_to_phys() argument
262 unsigned long pfn = kaddr_to_pfn(kaddr); in virt_to_phys()
264 ((unsigned long)kaddr & (PAGE_SIZE-1)); in virt_to_phys()
273 static inline int virt_addr_valid(const volatile void *kaddr) in virt_addr_valid() argument
276 return ((unsigned long)kaddr >= PAGE_OFFSET && kaddr < high_memory); in virt_addr_valid()
281 static inline unsigned long kaddr_to_pfn(const volatile void *kaddr) in kaddr_to_pfn() argument
283 return ((unsigned long)kaddr - PAGE_OFFSET) >> PAGE_SHIFT; in kaddr_to_pfn()
[all …]
/arch/alpha/include/asm/
Dmmzone.h58 #define kvaddr_to_nid(kaddr) pa_to_nid(__pa(kaddr)) argument
65 #define LOCAL_BASE_ADDR(kaddr) \ argument
66 ((unsigned long)__va(NODE_DATA(kvaddr_to_nid(kaddr))->node_start_pfn \
70 #define kern_addr_valid(kaddr) (0) argument
72 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
108 #define virt_addr_valid(kaddr) pfn_valid((__pa(kaddr) >> PAGE_SHIFT)) argument
/arch/arm/include/asm/
Dmemory.h151 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
170 #define virt_to_pfn(kaddr) \ argument
171 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \
245 #define virt_to_pfn(kaddr) \ argument
246 ((((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT) + \
331 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
332 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < … argument
333 && pfn_valid(virt_to_pfn(kaddr)))
Dtlbflush.h478 static inline void __local_flush_tlb_kernel_page(unsigned long kaddr) in __local_flush_tlb_kernel_page() argument
483 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
484 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
485 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
489 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
490 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
491 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
494 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument
498 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page()
503 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page()
[all …]
/arch/arm64/include/asm/
Dmemory.h203 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
204 #define _virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
206 #define _virt_addr_is_linear(kaddr) (((u64)(kaddr)) >= PAGE_OFFSET) argument
207 #define virt_addr_valid(kaddr) (_virt_addr_is_linear(kaddr) && \ argument
208 _virt_addr_valid(kaddr))
/arch/m68k/sun3x/
Ddvma.c82 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument
89 kaddr &= PAGE_MASK; in dvma_map_cpu()
96 kaddr, vaddr); in dvma_map_cpu()
131 __pa(kaddr), vaddr); in dvma_map_cpu()
133 set_pte(pte, pfn_pte(virt_to_pfn(kaddr), in dvma_map_cpu()
136 kaddr += PAGE_SIZE; in dvma_map_cpu()
151 inline int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
167 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu()
172 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
/arch/x86/include/asm/
Dpage.h63 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
65 extern bool __virt_addr_valid(unsigned long kaddr);
66 #define virt_addr_valid(kaddr) __virt_addr_valid((unsigned long) (kaddr)) argument
/arch/unicore32/include/asm/
Dmemory.h103 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
104 #define virt_addr_valid(kaddr) ((unsigned long)(kaddr) >= PAGE_OFFSET && \ argument
105 (unsigned long)(kaddr) < (unsigned long)high_memory)
Dtlbflush.h106 static inline void local_flush_tlb_kernel_page(unsigned long kaddr) in local_flush_tlb_kernel_page() argument
111 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
114 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
118 : : "r" (kaddr & PAGE_MASK) : "cc"); in local_flush_tlb_kernel_page()
/arch/m68k/include/asm/
Dpage_no.h25 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
35 #define virt_addr_valid(kaddr) (((void *)(kaddr) >= (void *)PAGE_OFFSET) && \ argument
36 ((void *)(kaddr) < (void *)memory_end))
/arch/hexagon/include/asm/
Dpage.h106 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument
113 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
143 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
/arch/arm64/mm/
Dflush.c37 static void sync_icache_aliases(void *kaddr, unsigned long len) in sync_icache_aliases() argument
39 unsigned long addr = (unsigned long)kaddr; in sync_icache_aliases()
42 __clean_dcache_area_pou(kaddr, len); in sync_icache_aliases()
50 unsigned long uaddr, void *kaddr, in flush_ptrace_access() argument
54 sync_icache_aliases(kaddr, len); in flush_ptrace_access()
/arch/sh/mm/
Duncached.c23 int virt_addr_uncached(unsigned long kaddr) in virt_addr_uncached() argument
25 return (kaddr >= uncached_start) && (kaddr < uncached_end); in virt_addr_uncached()
/arch/powerpc/include/asm/
Dpage.h132 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT) argument
133 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
143 #define virt_addr_valid(kaddr) (REGION_ID(kaddr) == KERNEL_REGION_ID && \ argument
144 pfn_valid(virt_to_pfn(kaddr)))
146 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
/arch/arm/mm/
Dcopypage-v4wt.c64 void *ptr, *kaddr = kmap_atomic(page); in v4wt_clear_user_highpage() local
79 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wt_clear_user_highpage()
81 kunmap_atomic(kaddr); in v4wt_clear_user_highpage()
Dcopypage-fa.c60 void *ptr, *kaddr = kmap_atomic(page); in fa_clear_user_highpage() local
77 : "0" (kaddr), "I" (PAGE_SIZE / 32) in fa_clear_user_highpage()
79 kunmap_atomic(kaddr); in fa_clear_user_highpage()
Dcopypage-v4wb.c69 void *ptr, *kaddr = kmap_atomic(page); in v4wb_clear_user_highpage() local
86 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4wb_clear_user_highpage()
88 kunmap_atomic(kaddr); in v4wb_clear_user_highpage()
Dcopypage-feroceon.c83 void *ptr, *kaddr = kmap_atomic(page); in feroceon_clear_user_highpage() local
101 : "0" (kaddr), "I" (PAGE_SIZE / 32) in feroceon_clear_user_highpage()
103 kunmap_atomic(kaddr); in feroceon_clear_user_highpage()
Dcopypage-xsc3.c84 void *ptr, *kaddr = kmap_atomic(page); in xsc3_mc_clear_user_highpage() local
97 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xsc3_mc_clear_user_highpage()
99 kunmap_atomic(kaddr); in xsc3_mc_clear_user_highpage()
Dcopypage-v4mc.c90 void *ptr, *kaddr = kmap_atomic(page); in v4_mc_clear_user_highpage() local
106 : "0" (kaddr), "I" (PAGE_SIZE / 64) in v4_mc_clear_user_highpage()
108 kunmap_atomic(kaddr); in v4_mc_clear_user_highpage()
Dcopypage-xscale.c111 void *ptr, *kaddr = kmap_atomic(page); in xscale_mc_clear_user_highpage() local
126 : "0" (kaddr), "I" (PAGE_SIZE / 32) in xscale_mc_clear_user_highpage()
128 kunmap_atomic(kaddr); in xscale_mc_clear_user_highpage()
/arch/cris/include/asm/
Dpage.h53 #define virt_to_page(kaddr) (mem_map + (((unsigned long)(kaddr) - PAGE_OFFSET) >> PAGE_SHIFT)) argument
54 #define virt_addr_valid(kaddr) pfn_valid((unsigned)(kaddr) >> PAGE_SHIFT) argument
/arch/m32r/include/asm/
Dpage.h78 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
79 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument

1234