Home
last modified time | relevance | path

Searched refs:uaddr (Results 1 – 25 of 72) sorted by relevance

123

/arch/mips/include/asm/
Dfutex.h25 #define __futex_atomic_op(op, insn, ret, oldval, uaddr, oparg) \ argument
52 "=" GCC_OFF_SMALL_ASM() (*uaddr) \
53 : "0" (0), GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oparg), \
82 "=" GCC_OFF_SMALL_ASM() (*uaddr) \
83 : "0" (0), GCC_OFF_SMALL_ASM() (*uaddr), "Jr" (oparg), \
88 ret = futex_atomic_op_inuser_local(op, oparg, oval, uaddr); \
93 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
97 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
102 __futex_atomic_op(op, "move $1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
107 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/hexagon/include/asm/
Dfutex.h13 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
29 : "=&r" (oldval), "=&r" (ret), "+m" (*uaddr) \
30 : "r" (uaddr), "r" (oparg), "i" (-EFAULT) \
35 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
39 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
44 __futex_atomic_op("%1 = %4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
47 __futex_atomic_op("%1 = add(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
51 __futex_atomic_op("%1 = or(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
56 oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
59 __futex_atomic_op("%1 = xor(%0,%4)\n", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
[all …]
/arch/arm/include/asm/
Dfutex.h25 #define __futex_atomic_op(insn, ret, oldval, tmp, uaddr, oparg) \ argument
29 prefetchw(uaddr); \
40 : "r" (uaddr), "r" (oparg), "Ir" (-EFAULT) \
46 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
53 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
58 prefetchw(uaddr); in futex_atomic_cmpxchg_inatomic()
70 : "r" (oldval), "r" (newval), "r" (uaddr), "Ir" (-EFAULT) in futex_atomic_cmpxchg_inatomic()
84 #define __futex_atomic_op(insn, ret, oldval, tmp, uaddr, oparg) \ argument
94 : "r" (uaddr), "r" (oparg), "Ir" (-EFAULT) \
100 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
Dtlbflush.h413 __local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) in __local_flush_tlb_page() argument
418 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in __local_flush_tlb_page()
422 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", uaddr); in __local_flush_tlb_page()
423 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", uaddr); in __local_flush_tlb_page()
424 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", uaddr); in __local_flush_tlb_page()
429 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", uaddr); in __local_flush_tlb_page()
430 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", uaddr); in __local_flush_tlb_page()
431 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", uaddr); in __local_flush_tlb_page()
435 local_flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) in local_flush_tlb_page() argument
439 uaddr = (uaddr & PAGE_MASK) | ASID(vma->vm_mm); in local_flush_tlb_page()
[all …]
/arch/s390/include/asm/
Dfutex.h11 #define __futex_atomic_op(insn, ret, oldval, newval, uaddr, oparg) \ argument
23 "=m" (*uaddr) \
24 : "0" (-EFAULT), "d" (oparg), "a" (uaddr), \
25 "m" (*uaddr) : "cc");
28 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
35 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
39 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
43 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
47 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 ret, oldval, newval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/xtensa/include/asm/
Dfutex.h24 #define __futex_atomic_op(insn, ret, old, uaddr, arg) \ argument
44 : [addr] "r" (uaddr), [oparg] "r" (arg), \
48 #define __futex_atomic_op(insn, ret, old, uaddr, arg) \ argument
68 [mem] "+m" (*(uaddr)) \
74 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
79 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
85 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
89 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
93 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
97 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/ia64/include/asm/
Dfutex.h9 #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ argument
18 : "r" (uaddr), "r" (oparg) \
23 #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ argument
40 : "r" (uaddr), "r" (oparg) \
49 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
53 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
58 __futex_atomic_op1("xchg4 %1=[%2],%3", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
62 __futex_atomic_op2("add %3=%3,%5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
65 __futex_atomic_op2("or %3=%3,%5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
68 __futex_atomic_op2("and %3=%3,%5", ret, oldval, uaddr, in arch_futex_atomic_op_inuser()
[all …]
/arch/arc/include/asm/
Dfutex.h18 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\ argument
40 : "r" (uaddr), "r" (oparg), "ir" (-EFAULT) \
46 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg)\ argument
67 : "r" (uaddr), "r" (oparg), "ir" (-EFAULT) \
74 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
78 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
87 __futex_atomic_op("mov %0, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
91 __futex_atomic_op("add %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
94 __futex_atomic_op("or %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
97 __futex_atomic_op("bic %0, %1, %3", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
[all …]
/arch/loongarch/include/asm/
Dfutex.h14 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
25 "=ZC" (*uaddr) \
26 : "0" (0), "ZC" (*uaddr), "Jr" (oparg) \
31 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
39 __futex_atomic_op("move $t0, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
42 __futex_atomic_op("add.w $t0, %1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
45 __futex_atomic_op("or $t0, %1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 __futex_atomic_op("and $t0, %1, %z5", ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
51 __futex_atomic_op("xor $t0, %1, %z5", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
66 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, u32 oldval, u32 newval) in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/openrisc/include/asm/
Dfutex.h11 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
27 : "=&r" (oldval), "=&r" (ret), "+m" (*uaddr) \
34 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
38 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
43 __futex_atomic_op("l.or %1,%4,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
46 __futex_atomic_op("l.add %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
49 __futex_atomic_op("l.or %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
52 __futex_atomic_op("l.and %1,%0,%4", ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
55 __futex_atomic_op("l.xor %1,%0,%4", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
68 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/x86/include/asm/
Dfutex.h15 #define unsafe_atomic_op1(insn, oval, uaddr, oparg, label) \ argument
21 : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \
29 #define unsafe_atomic_op2(insn, oval, uaddr, oparg, label) \ argument
41 "+m" (*uaddr), "=&r" (tem) \
49 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
51 if (!user_access_begin(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
56 unsafe_atomic_op1("xchgl %0, %2", oval, uaddr, oparg, Efault); in arch_futex_atomic_op_inuser()
60 uaddr, oparg, Efault); in arch_futex_atomic_op_inuser()
63 unsafe_atomic_op2("orl %4, %3", oval, uaddr, oparg, Efault); in arch_futex_atomic_op_inuser()
66 unsafe_atomic_op2("andl %4, %3", oval, uaddr, ~oparg, Efault); in arch_futex_atomic_op_inuser()
[all …]
/arch/powerpc/include/asm/
Dfutex.h12 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
28 : "b" (uaddr), "i" (-EFAULT), "r" (oparg) \
32 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
36 if (!user_access_begin(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
41 __futex_atomic_op("mr %1,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
44 __futex_atomic_op("add %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
47 __futex_atomic_op("or %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
50 __futex_atomic_op("andc %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
53 __futex_atomic_op("xor %1,%0,%4\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
66 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/microblaze/include/asm/
Dfutex.h11 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
28 : "r" (uaddr), "i" (-EFAULT), "r" (oparg) \
33 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
37 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
42 __futex_atomic_op("or %1,%4,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
45 __futex_atomic_op("add %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 __futex_atomic_op("or %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 __futex_atomic_op("andn %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
54 __futex_atomic_op("xor %1,%0,%4;", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
67 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/alpha/include/asm/
Dfutex.h12 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
26 : "r" (uaddr), "r"(oparg) \
30 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
34 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
39 __futex_atomic_op("mov %3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
42 __futex_atomic_op("addl %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
45 __futex_atomic_op("or %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
48 __futex_atomic_op("andnot %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 __futex_atomic_op("xor %0,%3,%1\n", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
64 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/sparc/include/asm/
Dfutex_64.h9 #define __futex_cas_op(insn, ret, oldval, uaddr, oparg) \ argument
30 : "r" (uaddr), "r" (oparg), "i" (-EFAULT) \
34 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
38 if (unlikely((((unsigned long) uaddr) & 0x3UL))) in arch_futex_atomic_op_inuser()
43 __futex_cas_op("mov\t%4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
46 __futex_cas_op("add\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
49 __futex_cas_op("or\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
52 __futex_cas_op("andn\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
55 __futex_cas_op("xor\t%2, %4, %1", ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
68 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/csky/include/asm/
Dfutex.h14 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
34 [u] "+m" (*uaddr), [t] "=&r" (tmp) \
42 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
46 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
52 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
56 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
60 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
64 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
68 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
83 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/riscv/include/asm/
Dfutex.h22 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument
30 [u] "+m" (*uaddr) \
37 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
41 if (!access_ok(uaddr, sizeof(u32))) in arch_futex_atomic_op_inuser()
47 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
51 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
55 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
59 ret, oldval, uaddr, ~oparg); in arch_futex_atomic_op_inuser()
63 ret, oldval, uaddr, oparg); in arch_futex_atomic_op_inuser()
76 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
[all …]
/arch/arm64/include/asm/
Dfutex.h15 #define __futex_atomic_op(insn, ret, oldval, uaddr, tmp, oparg) \ argument
33 : "=&r" (ret), "=&r" (oldval), "+Q" (*uaddr), "=&r" (tmp), \
44 u32 __user *uaddr = __uaccess_mask_ptr(_uaddr); in arch_futex_atomic_op_inuser() local
52 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
56 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
60 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
64 ret, oldval, uaddr, tmp, ~oparg); in arch_futex_atomic_op_inuser()
68 ret, oldval, uaddr, tmp, oparg); in arch_futex_atomic_op_inuser()
87 u32 __user *uaddr; in futex_atomic_cmpxchg_inatomic() local
92 uaddr = __uaccess_mask_ptr(_uaddr); in futex_atomic_cmpxchg_inatomic()
[all …]
Dtlbflush.h266 unsigned long uaddr) in __flush_tlb_page_nosync() argument
271 addr = __TLBI_VADDR(uaddr, ASID(mm)); in __flush_tlb_page_nosync()
274 mmu_notifier_arch_invalidate_secondary_tlbs(mm, uaddr & PAGE_MASK, in __flush_tlb_page_nosync()
275 (uaddr & PAGE_MASK) + PAGE_SIZE); in __flush_tlb_page_nosync()
279 unsigned long uaddr) in flush_tlb_page_nosync() argument
281 return __flush_tlb_page_nosync(vma->vm_mm, uaddr); in flush_tlb_page_nosync()
285 unsigned long uaddr) in flush_tlb_page() argument
287 flush_tlb_page_nosync(vma, uaddr); in flush_tlb_page()
308 unsigned long uaddr) in arch_tlbbatch_add_pending() argument
310 __flush_tlb_page_nosync(mm, uaddr); in arch_tlbbatch_add_pending()
/arch/um/kernel/skas/
Duaccess.c249 int arch_futex_atomic_op_inuser(int op, u32 oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
253 unsigned long addr = (unsigned long) uaddr; in arch_futex_atomic_op_inuser()
257 if (!access_ok(uaddr, sizeof(*uaddr))) in arch_futex_atomic_op_inuser()
273 uaddr = (u32 *) addr; in arch_futex_atomic_op_inuser()
274 oldval = *uaddr; in arch_futex_atomic_op_inuser()
280 *uaddr = oparg; in arch_futex_atomic_op_inuser()
283 *uaddr += oparg; in arch_futex_atomic_op_inuser()
286 *uaddr |= oparg; in arch_futex_atomic_op_inuser()
289 *uaddr &= ~oparg; in arch_futex_atomic_op_inuser()
292 *uaddr ^= oparg; in arch_futex_atomic_op_inuser()
[all …]
/arch/parisc/include/asm/
Dfutex.h35 arch_futex_atomic_op_inuser(int op, int oparg, int *oval, u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
38 unsigned long ua = (unsigned long)uaddr; in arch_futex_atomic_op_inuser()
48 if (unlikely(get_user(oldval, uaddr) != 0)) { in arch_futex_atomic_op_inuser()
77 if (unlikely(put_user(tmp, uaddr) != 0)) in arch_futex_atomic_op_inuser()
90 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
94 unsigned long ua = (unsigned long)uaddr; in futex_atomic_cmpxchg_inatomic()
99 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
111 if (unlikely(get_user(val, uaddr) != 0)) { in futex_atomic_cmpxchg_inatomic()
116 if (val == oldval && unlikely(put_user(newval, uaddr) != 0)) { in futex_atomic_cmpxchg_inatomic()
/arch/sh/include/asm/
Dfutex.h20 futex_atomic_cmpxchg_inatomic(u32 *uval, u32 __user *uaddr, in futex_atomic_cmpxchg_inatomic() argument
23 if (!access_ok(uaddr, sizeof(u32))) in futex_atomic_cmpxchg_inatomic()
26 return atomic_futex_op_cmpxchg_inatomic(uval, uaddr, oldval, newval); in futex_atomic_cmpxchg_inatomic()
30 u32 __user *uaddr) in arch_futex_atomic_op_inuser() argument
36 ret = get_user(oldval, uaddr); in arch_futex_atomic_op_inuser()
63 ret = futex_atomic_cmpxchg_inatomic(&prev, uaddr, oldval, newval); in arch_futex_atomic_op_inuser()
Dfutex-irq.h6 u32 __user *uaddr, in atomic_futex_op_cmpxchg_inatomic() argument
15 ret = get_user(prev, uaddr); in atomic_futex_op_cmpxchg_inatomic()
17 ret = put_user(newval, uaddr); in atomic_futex_op_cmpxchg_inatomic()
/arch/arc/kernel/
Dprocess.c46 SYSCALL_DEFINE3(arc_usr_cmpxchg, int __user *, uaddr, int, expected, int, new) in SYSCALL_DEFINE3() argument
63 ret = access_ok(uaddr, sizeof(*uaddr)); in SYSCALL_DEFINE3()
70 ret = __get_user(uval, uaddr); in SYSCALL_DEFINE3()
77 ret = __put_user(new, uaddr); in SYSCALL_DEFINE3()
94 ret = fixup_user_fault(current->mm, (unsigned long) uaddr, in SYSCALL_DEFINE3()
/arch/arm/mm/
Dflush.c128 void __flush_ptrace_access(struct page *page, unsigned long uaddr, void *kaddr, in __flush_ptrace_access() argument
140 flush_pfn_alias(page_to_pfn(page), uaddr); in __flush_ptrace_access()
149 flush_icache_alias(page_to_pfn(page), uaddr, len); in __flush_ptrace_access()
160 unsigned long uaddr, void *kaddr, unsigned long len) in flush_ptrace_access() argument
167 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_ptrace_access()
170 void flush_uprobe_xol_access(struct page *page, unsigned long uaddr, in flush_uprobe_xol_access() argument
175 __flush_ptrace_access(page, uaddr, kaddr, len, flags); in flush_uprobe_xol_access()
186 unsigned long uaddr, void *dst, const void *src, in copy_to_user_page() argument
193 flush_ptrace_access(vma, page, uaddr, dst, len); in copy_to_user_page()

123