Lines Matching refs:user_addr
56 unsigned long user_addr; in do_usercopy_stack() local
80 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in do_usercopy_stack()
83 if (user_addr >= TASK_SIZE) { in do_usercopy_stack()
90 if (copy_to_user((void __user *)user_addr, good_stack, in do_usercopy_stack()
97 if (copy_to_user((void __user *)user_addr, bad_stack, in do_usercopy_stack()
111 if (copy_from_user(good_stack, (void __user *)user_addr, in do_usercopy_stack()
118 if (copy_from_user(bad_stack, (void __user *)user_addr, in do_usercopy_stack()
126 vm_munmap(user_addr, PAGE_SIZE); in do_usercopy_stack()
135 unsigned long user_addr; in do_usercopy_heap_size() local
148 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in do_usercopy_heap_size()
151 if (user_addr >= TASK_SIZE) { in do_usercopy_heap_size()
159 test_user_addr = (void __user *)(user_addr + 16); in do_usercopy_heap_size()
189 vm_munmap(user_addr, PAGE_SIZE); in do_usercopy_heap_size()
203 unsigned char __user *user_addr; in do_usercopy_heap_whitelist() local
229 user_addr = (void __user *)user_alloc; in do_usercopy_heap_whitelist()
239 if (copy_to_user(user_addr, buf + offset, size)) { in do_usercopy_heap_whitelist()
245 if (copy_to_user(user_addr, buf + offset - 1, size)) { in do_usercopy_heap_whitelist()
251 if (copy_from_user(buf + offset, user_addr, size)) { in do_usercopy_heap_whitelist()
257 if (copy_from_user(buf + offset - 1, user_addr, size)) { in do_usercopy_heap_whitelist()
308 unsigned long user_addr; in lkdtm_USERCOPY_KERNEL() local
310 user_addr = vm_mmap(NULL, 0, PAGE_SIZE, in lkdtm_USERCOPY_KERNEL()
313 if (user_addr >= TASK_SIZE) { in lkdtm_USERCOPY_KERNEL()
320 if (copy_to_user((void __user *)user_addr, test_text, in lkdtm_USERCOPY_KERNEL()
328 if (copy_to_user((void __user *)user_addr, function_nocfi(vm_mmap), in lkdtm_USERCOPY_KERNEL()
336 vm_munmap(user_addr, PAGE_SIZE); in lkdtm_USERCOPY_KERNEL()