/arch/frv/kernel/ |
D | futex.c | 23 int oldval, ret; in atomic_futex_op_xchg_set() local 44 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_set() 49 *_oldval = oldval; in atomic_futex_op_xchg_set() 55 int oldval, ret; in atomic_futex_op_xchg_add() local 77 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_add() 82 *_oldval = oldval; in atomic_futex_op_xchg_add() 88 int oldval, ret; in atomic_futex_op_xchg_or() local 110 : "+U"(*uaddr), "=&r"(oldval), "=&r"(ret), "=r"(oparg) in atomic_futex_op_xchg_or() 115 *_oldval = oldval; in atomic_futex_op_xchg_or() 121 int oldval, ret; in atomic_futex_op_xchg_and() local [all …]
|
D | pm.c | 215 void __user *oldval, size_t __user *oldlenp, in cmode_sysctl() argument 218 if (oldval && oldlenp) { in cmode_sysctl() 227 if (put_user(clock_cmode_current, (unsigned __user *)oldval) || in cmode_sysctl() 318 void __user *oldval, size_t __user *oldlenp, in p0_sysctl() argument 321 if (oldval && oldlenp) { in p0_sysctl() 330 if (put_user(clock_p0_current, (unsigned __user *)oldval) || in p0_sysctl() 362 void __user *oldval, size_t __user *oldlenp, in cm_sysctl() argument 365 if (oldval && oldlenp) { in cm_sysctl() 374 if (put_user(clock_cm_current, (unsigned __user *)oldval) || in cm_sysctl()
|
/arch/x86/include/asm/ |
D | futex.h | 14 #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ argument 21 : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \ 24 #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ argument 36 : "=&a" (oldval), "=&r" (ret), \ 46 int oldval = 0, ret, tem; in futex_atomic_op_inuser() local 64 __futex_atomic_op1("xchgl %0, %2", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 67 __futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval, in futex_atomic_op_inuser() 71 __futex_atomic_op2("orl %4, %3", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 74 __futex_atomic_op2("andl %4, %3", ret, oldval, uaddr, ~oparg); in futex_atomic_op_inuser() 77 __futex_atomic_op2("xorl %4, %3", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() [all …]
|
/arch/ia64/include/asm/ |
D | futex.h | 9 #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ argument 17 : "+r" (r8), "=r" (oldval) \ 23 #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ argument 38 : "+r" (r8), "=r" (val), "=&r" (oldval), \ 44 } while (unlikely (val != oldval)); \ 55 int oldval = 0, ret; in futex_atomic_op_inuser() local 66 __futex_atomic_op1("xchg4 %1=[%2],%3", ret, oldval, uaddr, in futex_atomic_op_inuser() 70 __futex_atomic_op2("add %3=%3,%5", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 73 __futex_atomic_op2("or %3=%3,%5", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 76 __futex_atomic_op2("and %3=%3,%5", ret, oldval, uaddr, in futex_atomic_op_inuser() [all …]
|
/arch/arm/include/asm/ |
D | futex.h | 17 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 32 : "=&r" (ret), "=&r" (oldval) \ 43 int oldval = 0, ret; in futex_atomic_op_inuser() local 55 __futex_atomic_op("mov %0, %3", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 58 __futex_atomic_op("add %0, %1, %3", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 61 __futex_atomic_op("orr %0, %1, %3", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 64 __futex_atomic_op("and %0, %1, %3", ret, oldval, uaddr, ~oparg); in futex_atomic_op_inuser() 67 __futex_atomic_op("eor %0, %1, %3", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 77 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() 78 case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; in futex_atomic_op_inuser() [all …]
|
D | atomic.h | 85 unsigned long oldval, res; in atomic_cmpxchg() local 93 : "=&r" (res), "=&r" (oldval) in atomic_cmpxchg() 98 return oldval; in atomic_cmpxchg()
|
/arch/sparc/include/asm/ |
D | futex_64.h | 9 #define __futex_cas_op(insn, ret, oldval, uaddr, oparg) \ argument 29 : "=&r" (ret), "=&r" (oldval), "=&r" (tem) \ 39 int oldval = 0, ret, tem; in futex_atomic_op_inuser() local 53 __futex_cas_op("mov\t%4, %1", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 56 __futex_cas_op("add\t%2, %4, %1", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 59 __futex_cas_op("or\t%2, %4, %1", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 62 __futex_cas_op("andn\t%2, %4, %1", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 65 __futex_cas_op("xor\t%2, %4, %1", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 75 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() 76 case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; in futex_atomic_op_inuser() [all …]
|
/arch/mips/include/asm/ |
D | futex.h | 19 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 44 : "=r" (ret), "=&r" (oldval), "=R" (*uaddr) \ 70 : "=r" (ret), "=&r" (oldval), "=R" (*uaddr) \ 84 int oldval = 0, ret; in futex_atomic_op_inuser() local 95 __futex_atomic_op("move $1, %z5", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 100 ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 104 ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 108 ret, oldval, uaddr, ~oparg); in futex_atomic_op_inuser() 112 ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 122 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() [all …]
|
/arch/sh/include/asm/ |
D | futex.h | 19 int oldval = 0, ret; in futex_atomic_op_inuser() local 31 ret = atomic_futex_op_xchg_set(oparg, uaddr, &oldval); in futex_atomic_op_inuser() 34 ret = atomic_futex_op_xchg_add(oparg, uaddr, &oldval); in futex_atomic_op_inuser() 37 ret = atomic_futex_op_xchg_or(oparg, uaddr, &oldval); in futex_atomic_op_inuser() 40 ret = atomic_futex_op_xchg_and(~oparg, uaddr, &oldval); in futex_atomic_op_inuser() 43 ret = atomic_futex_op_xchg_xor(oparg, uaddr, &oldval); in futex_atomic_op_inuser() 54 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() 55 case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; in futex_atomic_op_inuser() 56 case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; in futex_atomic_op_inuser() 57 case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; in futex_atomic_op_inuser() [all …]
|
D | futex-irq.h | 7 int *oldval) in atomic_futex_op_xchg_set() argument 14 ret = get_user(*oldval, uaddr); in atomic_futex_op_xchg_set() 24 int *oldval) in atomic_futex_op_xchg_add() argument 31 ret = get_user(*oldval, uaddr); in atomic_futex_op_xchg_add() 33 ret = put_user(*oldval + oparg, uaddr); in atomic_futex_op_xchg_add() 41 int *oldval) in atomic_futex_op_xchg_or() argument 48 ret = get_user(*oldval, uaddr); in atomic_futex_op_xchg_or() 50 ret = put_user(*oldval | oparg, uaddr); in atomic_futex_op_xchg_or() 58 int *oldval) in atomic_futex_op_xchg_and() argument 65 ret = get_user(*oldval, uaddr); in atomic_futex_op_xchg_and() [all …]
|
D | spinlock.h | 40 unsigned long oldval; in __raw_spin_lock() local 51 : "=&z" (tmp), "=&r" (oldval) in __raw_spin_lock() 72 unsigned long tmp, oldval; in __raw_spin_trylock() local 82 : "=&z" (tmp), "=&r" (oldval) in __raw_spin_trylock() 87 return oldval; in __raw_spin_trylock() 175 unsigned long tmp, oldval; in __raw_read_trylock() local 188 : "=&z" (tmp), "=&r" (oldval) in __raw_read_trylock() 193 return (oldval > 0); in __raw_read_trylock() 198 unsigned long tmp, oldval; in __raw_write_trylock() local 211 : "=&z" (tmp), "=&r" (oldval) in __raw_write_trylock() [all …]
|
/arch/powerpc/include/asm/ |
D | futex.h | 12 #define __futex_atomic_op(insn, ret, oldval, uaddr, oparg) \ argument 29 : "=&r" (oldval), "=&r" (ret) \ 39 int oldval = 0, ret; in futex_atomic_op_inuser() local 50 __futex_atomic_op("", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 53 __futex_atomic_op("add %1,%0,%1\n", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 56 __futex_atomic_op("or %1,%0,%1\n", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 59 __futex_atomic_op("andc %1,%0,%1\n", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 62 __futex_atomic_op("xor %1,%0,%1\n", ret, oldval, uaddr, oparg); in futex_atomic_op_inuser() 72 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() 73 case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; in futex_atomic_op_inuser() [all …]
|
/arch/s390/include/asm/ |
D | futex.h | 16 int oldval, ret; in futex_atomic_op_inuser() local 25 ret = uaccess.futex_atomic_op(op, uaddr, oparg, &oldval); in futex_atomic_op_inuser() 30 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() 31 case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; in futex_atomic_op_inuser() 32 case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; in futex_atomic_op_inuser() 33 case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; in futex_atomic_op_inuser() 34 case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break; in futex_atomic_op_inuser() 35 case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break; in futex_atomic_op_inuser() 43 int oldval, int newval) in futex_atomic_cmpxchg_inatomic() argument 48 return uaccess.futex_atomic_cmpxchg(uaddr, oldval, newval); in futex_atomic_cmpxchg_inatomic()
|
/arch/parisc/include/asm/ |
D | futex.h | 17 int oldval = 0, ret; in futex_atomic_op_inuser() local 40 case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; in futex_atomic_op_inuser() 41 case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; in futex_atomic_op_inuser() 42 case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; in futex_atomic_op_inuser() 43 case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; in futex_atomic_op_inuser() 44 case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break; in futex_atomic_op_inuser() 45 case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break; in futex_atomic_op_inuser() 54 futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, int newval) in futex_atomic_cmpxchg_inatomic() argument 70 if (uval == oldval) in futex_atomic_cmpxchg_inatomic()
|
/arch/s390/lib/ |
D | uaccess_std.c | 243 #define __futex_atomic_op(insn, ret, oldval, newval, uaddr, oparg) \ argument 253 : "=d" (ret), "=&d" (oldval), "=&d" (newval), \ 260 int oldval = 0, newval, ret; in futex_atomic_op_std() local 265 ret, oldval, newval, uaddr, oparg); in futex_atomic_op_std() 269 ret, oldval, newval, uaddr, oparg); in futex_atomic_op_std() 273 ret, oldval, newval, uaddr, oparg); in futex_atomic_op_std() 277 ret, oldval, newval, uaddr, oparg); in futex_atomic_op_std() 281 ret, oldval, newval, uaddr, oparg); in futex_atomic_op_std() 286 *old = oldval; in futex_atomic_op_std() 290 int futex_atomic_cmpxchg_std(int __user *uaddr, int oldval, int newval) in futex_atomic_cmpxchg_std() argument [all …]
|
D | uaccess_pt.c | 337 #define __futex_atomic_op(insn, ret, oldval, newval, uaddr, oparg) \ argument 345 : "=d" (ret), "=&d" (oldval), "=&d" (newval), \ 352 int oldval = 0, newval, ret; in __futex_atomic_op_pt() local 357 ret, oldval, newval, uaddr, oparg); in __futex_atomic_op_pt() 361 ret, oldval, newval, uaddr, oparg); in __futex_atomic_op_pt() 365 ret, oldval, newval, uaddr, oparg); in __futex_atomic_op_pt() 369 ret, oldval, newval, uaddr, oparg); in __futex_atomic_op_pt() 373 ret, oldval, newval, uaddr, oparg); in __futex_atomic_op_pt() 379 *old = oldval; in __futex_atomic_op_pt() 402 static int __futex_atomic_cmpxchg_pt(int __user *uaddr, int oldval, int newval) in __futex_atomic_cmpxchg_pt() argument [all …]
|
/arch/mips/lasat/ |
D | sysctl.c | 42 void *oldval, size_t *oldlenp, in sysctl_lasatstring() argument 47 r = sysctl_string(table, oldval, oldlenp, newval, newlen); in sysctl_lasatstring() 116 void *oldval, size_t *oldlenp, in sysctl_lasat_intvec() argument 121 r = sysctl_intvec(table, oldval, oldlenp, newval, newlen); in sysctl_lasat_intvec() 134 void *oldval, size_t *oldlenp, in sysctl_lasat_rtc() argument 142 r = sysctl_intvec(table, oldval, oldlenp, newval, newlen); in sysctl_lasat_rtc() 214 void *oldval, size_t *oldlenp, in sysctl_lasat_prid() argument 219 r = sysctl_intvec(table, oldval, oldlenp, newval, newlen); in sysctl_lasat_prid()
|
/arch/m32r/kernel/ |
D | sys_m32r.c | 36 int oldval; in sys_tas() local 68 : "=&r" (oldval) in sys_tas() 76 return oldval; in sys_tas()
|
/arch/x86/xen/ |
D | spinlock.c | 258 u8 oldval; in __xen_spin_lock() local 279 : "+m" (xl->lock), "=q" (oldval), "+r" (timeout) in __xen_spin_lock() 285 } while (unlikely(oldval != 0 && in __xen_spin_lock()
|
/arch/parisc/kernel/ |
D | sys_parisc32.c | 104 u32 oldval; member 127 if (tmp.oldval && tmp.oldlenp) { in sys32_sysctl() 148 (void __user *)(u64)tmp.oldval, oldlenp, in sys32_sysctl()
|
/arch/mips/kernel/ |
D | linux32.c | 273 compat_caddr_t oldval; member 293 if (tmp.oldval && tmp.oldlenp) { 307 error = do_sysctl((int __user *)A(tmp.name), tmp.nlen, (void __user *)A(tmp.oldval),
|
/arch/powerpc/kernel/ |
D | sys_ppc32.c | 516 u32 oldval; member 534 if (tmp.oldval && tmp.oldlenp) { in compat_sys_sysctl() 549 compat_ptr(tmp.oldval), oldlenp, in compat_sys_sysctl()
|
/arch/sparc/kernel/ |
D | sys_sparc32.c | 607 u32 oldval; member 627 if (tmp.oldval && tmp.oldlenp) { in sys32_sysctl() 643 (void __user *)(unsigned long) tmp.oldval, in sys32_sysctl()
|
/arch/s390/kernel/ |
D | compat_linux.c | 576 u32 oldval; member 594 if (tmp.oldval && tmp.oldlenp) { in sys32_sysctl() 608 error = do_sysctl(compat_ptr(tmp.name), tmp.nlen, compat_ptr(tmp.oldval), in sys32_sysctl()
|
/arch/x86/ia32/ |
D | sys_ia32.c | 464 unsigned int oldval; member 493 oldvalp = compat_ptr(a32.oldval); in sys32_sysctl()
|