/arch/nds32/mm/ |
D | cacheflush.c | 157 unsigned long vto, flags; in flush_cache_page() local 160 vto = kremap0(addr, pfn << PAGE_SHIFT); in flush_cache_page() 161 cpu_cache_wbinval_page(vto, vma->vm_flags & VM_EXEC); in flush_cache_page() 162 kunmap01(vto); in flush_cache_page() 178 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument 183 copy_page(vto, vfrom); in copy_user_page() 184 cpu_dcache_wbinval_page((unsigned long)vto); in copy_user_page() 185 cpu_icache_inval_page((unsigned long)vto); in copy_user_page() 200 unsigned long vto, vfrom, flags, kto, kfrom, pfrom, pto; in copy_user_highpage() local 209 vto = kremap0(vaddr, pto); in copy_user_highpage() [all …]
|
/arch/sh/mm/ |
D | cache.c | 66 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() local 67 memcpy(vto, src, len); in copy_to_user_page() 68 kunmap_coherent(vto); in copy_to_user_page() 98 void *vfrom, *vto; in copy_user_highpage() local 100 vto = kmap_atomic(to); in copy_user_highpage() 105 copy_page(vto, vfrom); in copy_user_highpage() 109 copy_page(vto, vfrom); in copy_user_highpage() 113 if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) || in copy_user_highpage() 115 __flush_purge_region(vto, PAGE_SIZE); in copy_user_highpage() 117 kunmap_atomic(vto); in copy_user_highpage()
|
/arch/mips/mm/ |
D | init.c | 174 void *vfrom, *vto; in copy_user_highpage() local 176 vto = kmap_atomic(to); in copy_user_highpage() 180 copy_page(vto, vfrom); in copy_user_highpage() 184 copy_page(vto, vfrom); in copy_user_highpage() 188 pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK)) in copy_user_highpage() 189 flush_data_cache_page((unsigned long)vto); in copy_user_highpage() 190 kunmap_atomic(vto); in copy_user_highpage() 201 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() local 202 memcpy(vto, src, len); in copy_to_user_page()
|
/arch/nios2/mm/ |
D | cacheflush.c | 233 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument 238 copy_page(vto, vfrom); in copy_user_page() 239 __flush_dcache((unsigned long)vto, (unsigned long)vto + PAGE_SIZE); in copy_user_page() 240 __flush_icache((unsigned long)vto, (unsigned long)vto + PAGE_SIZE); in copy_user_page()
|
/arch/parisc/include/asm/ |
D | page.h | 32 #define clear_user_page(vto, vaddr, page) clear_page_asm(vto) argument 33 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr,
|
/arch/riscv/include/asm/ |
D | page.h | 54 #define copy_user_page(vto, vfrom, vaddr, topg) \ argument 55 memcpy((vto), (vfrom), PAGE_SIZE)
|
/arch/microblaze/include/asm/ |
D | page.h | 82 # define copy_user_page(vto, vfrom, vaddr, topg) \ argument 83 memcpy((vto), (vfrom), PAGE_SIZE)
|
/arch/nds32/include/asm/ |
D | page.h | 30 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr,
|
/arch/nios2/include/asm/ |
D | page.h | 54 extern void copy_user_page(void *vto, void *vfrom, unsigned long vaddr,
|
/arch/s390/include/asm/ |
D | vx-insn.h | 366 .macro VLM vfrom, vto, disp, base, hint=3 368 VX_NUM v3, \vto 376 .macro VSTM vfrom, vto, disp, base, hint=3 378 VX_NUM v3, \vto
|
/arch/powerpc/mm/ |
D | mem.c | 402 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument 405 copy_page(vto, vfrom); in copy_user_page()
|
/arch/parisc/kernel/ |
D | cache.c | 469 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument 478 copy_page_asm(vto, vfrom); in copy_user_page()
|
/arch/sparc/mm/ |
D | init_64.c | 3131 char *vfrom, *vto; in copy_user_highpage() local 3134 vto = kmap_atomic(to); in copy_user_highpage() 3135 copy_user_page(vto, vfrom, vaddr, to); in copy_user_highpage() 3136 kunmap_atomic(vto); in copy_user_highpage() 3165 char *vfrom, *vto; in copy_highpage() local 3168 vto = kmap_atomic(to); in copy_highpage() 3169 copy_page(vto, vfrom); in copy_highpage() 3170 kunmap_atomic(vto); in copy_highpage()
|