Home
last modified time | relevance | path

Searched refs:uaddr (Results 1 – 25 of 58) sorted by relevance

123

/arch/parisc/include/asm/
Dfutex.h16 _futex_spin_lock_irqsave(u32 __user *uaddr, unsigned long int *flags) in _futex_spin_lock_irqsave() argument
19 long index = ((long)uaddr & 0xf0) >> 2; in _futex_spin_lock_irqsave()
26 _futex_spin_unlock_irqrestore(u32 __user *uaddr, unsigned long int *flags) in _futex_spin_unlock_irqrestore() argument
29 long index = ((long)uaddr & 0xf0) >> 2; in _futex_spin_unlock_irqrestore()
36 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
42 _futex_spin_lock_irqsave(uaddr, &flags); in arch_futex_atomic_op_inuser()
46 if (unlikely(get_user(oldval, uaddr) != 0)) in arch_futex_atomic_op_inuser()
72 if (ret == 0 && unlikely(put_user(tmp, uaddr) != 0)) in arch_futex_atomic_op_inuser()
77 _futex_spin_unlock_irqrestore(uaddr, &flags); in arch_futex_atomic_op_inuser()
86 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/frv/kernel/
Dfutex.c21 static inline int atomic_futex_op_xchg_set(int oparg, u32 __user *uaddr, int *_oldval) in atomic_futex_op_xchg_set() argument
44 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_set()
53 static inline int atomic_futex_op_xchg_add(int oparg, u32 __user *uaddr, int *_oldval) in atomic_futex_op_xchg_add() argument
77 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_add()
86 static inline int atomic_futex_op_xchg_or(int oparg, u32 __user *uaddr, int *_oldval) in atomic_futex_op_xchg_or() argument
110 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_or()
119 static inline int atomic_futex_op_xchg_and(int oparg, u32 __user *uaddr, int *_oldval) in atomic_futex_op_xchg_and() argument
143 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_and()
152 static inline int atomic_futex_op_xchg_xor(int oparg, u32 __user *uaddr, int *_oldval) in atomic_futex_op_xchg_xor() argument
176 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_xor()
[all …]
/arch/mips/include/asm/
Dfutex.h21 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
48 "=" GCC_OFF_SMALL_ASM() (*uaddr) \
49 : "0" (0), GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oparg), \
77 "=" GCC_OFF_SMALL_ASM() (*uaddr) \
78 : "0" (0), GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oparg), \
86 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
94 __futex_atomic_op("move $1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
99 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
103 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
107 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/hexagon/include/asm/
Dfutex.h13 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
29 : "=&r" (oldval), "=&r" (ret), "+m" (*uaddr) \
30 : "r" (uaddr), "r" (oparg), "i" (-EFAULT) \
35 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
43 __futex_atomic_op("%1 = %4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
46 __futex_atomic_op("%1 = add(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
50 __futex_atomic_op("%1 = or(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
55 oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
58 __futex_atomic_op("%1 = xor(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
74 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, u32 oldval, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/s390/include/asm/
Dfutex.h10 #define __futex_atomic_op(insn, ret, oldval, newval, uaddr, oparg) \ argument
21 "=m" (*uaddr) \
22 : "0" (-EFAULT), "d" (oparg), "a" (uaddr), \
23 "m" (*uaddr) : "cc");
26 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
36 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
40 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
44 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
52 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/arm/include/asm/
Dfutex.h25 #define __futex_atomic_op(insn, ret, oldval, tmp, uaddr, oparg) \ argument
29 prefetchw(uaddr); \
40 : "r" (uaddr), "r" (oparg), "Ir" (-EFAULT) \
46 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
53 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
58 prefetchw(uaddr); in futex_atomic_cmpxchg_inatomic()
70 : "r" (oldval), "r" (newval), "r" (uaddr), "Ir" (-EFAULT) in futex_atomic_cmpxchg_inatomic()
84 #define __futex_atomic_op(insn, ret, oldval, tmp, uaddr, oparg) \ argument
94 : "r" (uaddr), "r" (oparg), "Ir" (-EFAULT) \
100 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
Dtlbflush.h423 __local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) in __local_flush_tlb_page() argument
428 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __local_flush_tlb_page()
432 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", uaddr); in __local_flush_tlb_page()
433 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", uaddr); in __local_flush_tlb_page()
434 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", uaddr); in __local_flush_tlb_page()
439 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", uaddr); in __local_flush_tlb_page()
440 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", uaddr); in __local_flush_tlb_page()
441 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", uaddr); in __local_flush_tlb_page()
445 local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) in local_flush_tlb_page() argument
449 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in local_flush_tlb_page()
[all …]
/arch/xtensa/include/asm/
Dfutex.h24 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
44 : "r" (uaddr), "I" (-EFAULT), "r" (oparg) \
48 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
60 __futex_atomic_op("mov %1, %4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
63 __futex_atomic_op("add %1, %0, %4", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
67 __futex_atomic_op("or %1, %0, %4", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
71 __futex_atomic_op("and %1, %0, %4", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
75 __futex_atomic_op("xor %1, %0, %4", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
91 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
96 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
[all …]
/arch/arc/include/asm/
Dfutex.h21 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\ argument
43 : "r" (uaddr), "r" (oparg), "ir" (-EFAULT) \
49 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\ argument
70 : "r" (uaddr), "r" (oparg), "ir" (-EFAULT) \
77 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
88 __futex_atomic_op("mov %0, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
92 __futex_atomic_op("add %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
95 __futex_atomic_op("or %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
98 __futex_atomic_op("bic %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
101 __futex_atomic_op("xor %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/ia64/include/asm/
Dfutex.h9 #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ argument
18 : "r" (uaddr), "r" (oparg) \
23 #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ argument
40 : "r" (uaddr), "r" (oparg) \
49 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
57 __futex_atomic_op1("xchg4 %1=[%2],%3", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
61 __futex_atomic_op2("add %3=%3,%5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
64 __futex_atomic_op2("or %3=%3,%5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
67 __futex_atomic_op2("and %3=%3,%5", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
71 __futex_atomic_op2("xor %3=%3,%5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/x86/include/asm/
Dfutex.h15 #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ argument
24 : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \
27 #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ argument
42 "+m" (*uaddr), "=&r" (tem) \
46 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
54 __futex_atomic_op1("xchgl %0, %2", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
58 uaddr, oparg); in arch_futex_atomic_op_inuser()
61 __futex_atomic_op2("orl %4, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
64 __futex_atomic_op2("andl %4, %3", ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
67 __futex_atomic_op2("xorl %4, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/openrisc/include/asm/
Dfutex.h11 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
27 : "=&r" (oldval), "=&r" (ret), "+m" (*uaddr) \
34 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
42 __futex_atomic_op("l.or %1,%4,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
45 __futex_atomic_op("l.add %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 __futex_atomic_op("l.or %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 __futex_atomic_op("l.and %1,%0,%4", ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
54 __futex_atomic_op("l.xor %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
69 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
75 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
[all …]
/arch/sparc/include/asm/
Dfutex_64.h9 #define __futex_cas_op(insn, ret, oldval, uaddr, oparg) \ argument
30 : "r" (uaddr), "r" (oparg), "i" (-EFAULT) \
34 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
38 if (unlikely((((unsigned long) uaddr) & 0x3UL))) in arch_futex_atomic_op_inuser()
45 __futex_cas_op("mov\t%4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 __futex_cas_op("add\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 __futex_cas_op("or\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
54 __futex_cas_op("andn\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
57 __futex_cas_op("xor\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
72 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/powerpc/include/asm/
Dfutex.h13 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
30 : "b" (uaddr), "i" (-EFAULT), "r" (oparg) \
34 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
42 __futex_atomic_op("mr %1,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
45 __futex_atomic_op("add %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 __futex_atomic_op("or %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 __futex_atomic_op("andc %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
54 __futex_atomic_op("xor %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
68 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
74 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
[all …]
/arch/microblaze/include/asm/
Dfutex.h11 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
28 : "r" (uaddr), "i" (-EFAULT), "r" (oparg) \
33 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
41 __futex_atomic_op("or %1,%4,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
44 __futex_atomic_op("add %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
47 __futex_atomic_op("or %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
50 __futex_atomic_op("andn %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
53 __futex_atomic_op("xor %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
68 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
74 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
[all …]
/arch/alpha/include/asm/
Dfutex.h12 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
26 : "r" (uaddr), "r"(oparg) \
30 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
38 __futex_atomic_op("mov %3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
41 __futex_atomic_op("addl %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
44 __futex_atomic_op("or %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
47 __futex_atomic_op("andnot %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
50 __futex_atomic_op("xor %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
65 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
71 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
[all …]
/arch/arm64/include/asm/
Dfutex.h28 #define __futex_atomic_op(insn, ret, oldval, uaddr, tmp, oparg) \ argument
51 : "=&r" (ret), "=&r" (oldval), "+Q" (*uaddr), "=&r" (tmp), \
62 u32 __user *uaddr = __uaccess_mask_ptr(_uaddr); in arch_futex_atomic_op_inuser() local
69 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
73 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
77 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
81 ret, oldval, uaddr, tmp, ~oparg); in arch_futex_atomic_op_inuser()
85 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
106 u32 __user *uaddr; in futex_atomic_cmpxchg_inatomic() local
111 uaddr = __uaccess_mask_ptr(_uaddr); in futex_atomic_cmpxchg_inatomic()
[all …]
/arch/tile/include/asm/
Dfutex.h50 : "=r" (ret), "=r" (val), "+m" (*(uaddr)) \
51 : "r" (uaddr), "r" (oparg), "i" (-EFAULT))
63 if ((ret = __get_user(oldval, uaddr)) == 0) { \
78 struct __get_user gu = FN((u32 __force *)uaddr, lock, oparg); \
91 struct __get_user gu = __atomic32_cmpxchg((u32 __force *)uaddr, \
105 __insn_prefetch(uaddr); \
106 lock = __atomic_hashed_lock((int __force *)uaddr)
110 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
148 static inline int futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
155 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
/arch/sh/include/asm/
Dfutex.h22 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
25 if (!access_ok(VERIFY_WRITE, uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
28 return atomic_futex_op_cmpxchg_inatomic(uval, uaddr, oldval, newval); in futex_atomic_cmpxchg_inatomic()
32 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
43 ret = get_user(oldval, uaddr); in arch_futex_atomic_op_inuser()
70 ret = futex_atomic_cmpxchg_inatomic(&prev, uaddr, oldval, newval); in arch_futex_atomic_op_inuser()
Dfutex-irq.h6 u32 __user *uaddr, in atomic_futex_op_cmpxchg_inatomic() argument
15 ret = get_user(prev, uaddr); in atomic_futex_op_cmpxchg_inatomic()
17 ret = put_user(newval, uaddr); in atomic_futex_op_cmpxchg_inatomic()
/arch/arm/kvm/
Dcoproc.c781 static int reg_from_user(void *val, const void __user *uaddr, u64 id) in reg_from_user() argument
783 if (copy_from_user(val, uaddr, KVM_REG_SIZE(id)) != 0) in reg_from_user()
792 static int reg_to_user(void __user *uaddr, const void *val, u64 id) in reg_to_user() argument
794 if (copy_to_user(uaddr, val, KVM_REG_SIZE(id)) != 0) in reg_to_user()
799 static int get_invariant_cp15(u64 id, void __user *uaddr) in get_invariant_cp15() argument
816 ret = reg_to_user(uaddr, &val, id); in get_invariant_cp15()
818 ret = reg_to_user(uaddr, &r->val, id); in get_invariant_cp15()
823 static int set_invariant_cp15(u64 id, void __user *uaddr) in set_invariant_cp15() argument
840 err = reg_from_user(&val32, uaddr, id); in set_invariant_cp15()
844 err = reg_from_user(&val, uaddr, id); in set_invariant_cp15()
[all …]
/arch/unicore32/include/asm/
Dtlbflush.h88 local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) in local_flush_tlb_page() argument
94 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page()
97 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page()
101 : : "r" (uaddr & PAGE_MASK) : "cc"); in local_flush_tlb_page()
/arch/arc/kernel/
Dprocess.c47 SYSCALL_DEFINE3(arc_usr_cmpxchg, int *, uaddr, int, expected, int, new) in SYSCALL_DEFINE3() argument
64 ret = access_ok(VERIFY_WRITE, uaddr, sizeof(*uaddr)); in SYSCALL_DEFINE3()
71 ret = __get_user(uval, uaddr); in SYSCALL_DEFINE3()
78 ret = __put_user(new, uaddr); in SYSCALL_DEFINE3()
95 ret = fixup_user_fault(current, current->mm, (unsigned long) uaddr, in SYSCALL_DEFINE3()
/arch/arm/mm/
Dflush.c131 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, in __flush_ptrace_access() argument
143 flush_pfn_alias(page_to_pfn(page), uaddr); in __flush_ptrace_access()
152 flush_icache_alias(page_to_pfn(page), uaddr, len); in __flush_ptrace_access()
163 unsigned long uaddr, void *kaddr, unsigned long len) in flush_ptrace_access() argument
170 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_ptrace_access()
173 void flush_uprobe_xol_access(struct page *page, unsigned long uaddr, in flush_uprobe_xol_access() argument
178 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_uprobe_xol_access()
189 unsigned long uaddr, void *dst, const void *src, in copy_to_user_page() argument
196 flush_ptrace_access(vma, page, uaddr, dst, len); in copy_to_user_page()
/arch/unicore32/mm/
Dflush.c36 unsigned long uaddr, void *kaddr, unsigned long len) in flush_ptrace_access() argument
54 unsigned long uaddr, void *dst, const void *src, in copy_to_user_page() argument
58 flush_ptrace_access(vma, page, uaddr, dst, len); in copy_to_user_page()

123