/arch/mips/include/asm/ |
D | atomic.h | 32 #define atomic_read(v) (*(volatile int *)&(v)->counter) 41 #define atomic_set(v, i) ((v)->counter = (i)) 62 : "=&r" (temp), "=m" (v->counter) in atomic_add() 63 : "Ir" (i), "m" (v->counter)); in atomic_add() 74 : "=&r" (temp), "=m" (v->counter) in atomic_add() 75 : "Ir" (i), "m" (v->counter)); in atomic_add() 81 v->counter += i; in atomic_add() 105 : "=&r" (temp), "=m" (v->counter) in atomic_sub() 106 : "Ir" (i), "m" (v->counter)); in atomic_sub() 117 : "=&r" (temp), "=m" (v->counter) in atomic_sub() [all …]
|
D | local.h | 43 : "=&r" (result), "=&r" (temp), "=m" (l->a.counter) in local_add_return() 44 : "Ir" (i), "m" (l->a.counter) in local_add_return() 57 : "=&r" (result), "=&r" (temp), "=m" (l->a.counter) in local_add_return() 58 : "Ir" (i), "m" (l->a.counter) in local_add_return() 64 result = l->a.counter; in local_add_return() 66 l->a.counter = result; in local_add_return() 88 : "=&r" (result), "=&r" (temp), "=m" (l->a.counter) in local_sub_return() 89 : "Ir" (i), "m" (l->a.counter) in local_sub_return() 102 : "=&r" (result), "=&r" (temp), "=m" (l->a.counter) in local_sub_return() 103 : "Ir" (i), "m" (l->a.counter) in local_sub_return() [all …]
|
/arch/x86/kernel/cpu/ |
D | perfctr-watchdog.c | 94 int avail_to_resrv_perfctr_nmi_bit(unsigned int counter) in avail_to_resrv_perfctr_nmi_bit() argument 96 BUG_ON(counter > NMI_MAX_COUNTER_BITS); in avail_to_resrv_perfctr_nmi_bit() 98 return !test_bit(counter, perfctr_nmi_owner); in avail_to_resrv_perfctr_nmi_bit() 104 unsigned int counter; in reserve_perfctr_nmi() local 106 counter = nmi_perfctr_msr_to_bit(msr); in reserve_perfctr_nmi() 108 if (counter > NMI_MAX_COUNTER_BITS) in reserve_perfctr_nmi() 111 if (!test_and_set_bit(counter, perfctr_nmi_owner)) in reserve_perfctr_nmi() 119 unsigned int counter; in release_perfctr_nmi() local 121 counter = nmi_perfctr_msr_to_bit(msr); in release_perfctr_nmi() 123 if (counter > NMI_MAX_COUNTER_BITS) in release_perfctr_nmi() [all …]
|
/arch/x86/include/asm/ |
D | atomic64_64.h | 21 return (*(volatile long *)&(v)->counter); in atomic64_read() 33 v->counter = i; in atomic64_set() 46 : "=m" (v->counter) in atomic64_add() 47 : "er" (i), "m" (v->counter)); in atomic64_add() 60 : "=m" (v->counter) in atomic64_sub() 61 : "er" (i), "m" (v->counter)); in atomic64_sub() 78 : "=m" (v->counter), "=qm" (c) in atomic64_sub_and_test() 79 : "er" (i), "m" (v->counter) : "memory"); in atomic64_sub_and_test() 92 : "=m" (v->counter) in atomic64_inc() 93 : "m" (v->counter)); in atomic64_inc() [all …]
|
D | atomic.h | 25 return (*(volatile int *)&(v)->counter); in atomic_read() 37 v->counter = i; in atomic_set() 50 : "+m" (v->counter) in atomic_add() 64 : "+m" (v->counter) in atomic_sub() 82 : "+m" (v->counter), "=qm" (c) in atomic_sub_and_test() 96 : "+m" (v->counter)); in atomic_inc() 108 : "+m" (v->counter)); in atomic_dec() 124 : "+m" (v->counter), "=qm" (c) in atomic_dec_and_test() 142 : "+m" (v->counter), "=qm" (c) in atomic_inc_and_test() 161 : "+m" (v->counter), "=qm" (c) in atomic_add_negative() [all …]
|
D | local.h | 21 : "+m" (l->a.counter)); in local_inc() 27 : "+m" (l->a.counter)); in local_dec() 33 : "+m" (l->a.counter) in local_add() 40 : "+m" (l->a.counter) in local_sub() 58 : "+m" (l->a.counter), "=qm" (c) in local_sub_and_test() 76 : "+m" (l->a.counter), "=qm" (c) in local_dec_and_test() 94 : "+m" (l->a.counter), "=qm" (c) in local_inc_and_test() 113 : "+m" (l->a.counter), "=qm" (c) in local_add_negative() 136 : "+r" (i), "+m" (l->a.counter) in local_add_return() 159 (cmpxchg_local(&((l)->a.counter), (o), (n))) [all …]
|
/arch/powerpc/include/asm/ |
D | atomic.h | 18 __asm__ __volatile__("lwz%U1%X1 %0,%1" : "=r"(t) : "m"(v->counter)); in atomic_read() 25 __asm__ __volatile__("stw%U0%X0 %1,%0" : "=m"(v->counter) : "r"(i)); in atomic_set() 38 : "=&r" (t), "+m" (v->counter) in atomic_add() 39 : "r" (a), "r" (&v->counter) in atomic_add() 56 : "r" (a), "r" (&v->counter) in atomic_add_return() 74 : "=&r" (t), "+m" (v->counter) in atomic_sub() 75 : "r" (a), "r" (&v->counter) in atomic_sub() 92 : "r" (a), "r" (&v->counter) in atomic_sub_return() 108 : "=&r" (t), "+m" (v->counter) in atomic_inc() 109 : "r" (&v->counter) in atomic_inc() [all …]
|
D | local.h | 33 : "r" (a), "r" (&(l->a.counter)) in local_add_return() 52 : "r" (a), "r" (&(l->a.counter)) in local_sub_return() 69 : "r" (&(l->a.counter)) in local_inc_return() 96 : "r" (&(l->a.counter)) in local_dec_return() 103 (cmpxchg_local(&((l)->a.counter), (o), (n))) 104 #define local_xchg(l, n) (xchg_local(&((l)->a.counter), (n))) 130 : "r" (&(l->a.counter)), "r" (a), "r" (u) in local_add_unless() 159 : "r" (&(l->a.counter)) in local_dec_if_positive() 170 #define __local_inc(l) ((l)->a.counter++) 171 #define __local_dec(l) ((l)->a.counter++) [all …]
|
/arch/arm/include/asm/ |
D | atomic.h | 29 #define atomic_read(v) (*(volatile int *)&(v)->counter) 30 #define atomic_set(v,i) (((v)->counter) = (i)) 50 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_add() 51 : "r" (&v->counter), "Ir" (i) in atomic_add() 68 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_add_return() 69 : "r" (&v->counter), "Ir" (i) in atomic_add_return() 88 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_sub() 89 : "r" (&v->counter), "Ir" (i) in atomic_sub() 106 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_sub_return() 107 : "r" (&v->counter), "Ir" (i) in atomic_sub_return() [all …]
|
/arch/alpha/include/asm/ |
D | atomic.h | 20 #define atomic_read(v) (*(volatile int *)&(v)->counter) 21 #define atomic64_read(v) (*(volatile long *)&(v)->counter) 23 #define atomic_set(v,i) ((v)->counter = (i)) 24 #define atomic64_set(v,i) ((v)->counter = (i)) 43 :"=&r" (temp), "=m" (v->counter) in atomic_add() 44 :"Ir" (i), "m" (v->counter)); in atomic_add() 58 :"=&r" (temp), "=m" (v->counter) in atomic64_add() 59 :"Ir" (i), "m" (v->counter)); in atomic64_add() 73 :"=&r" (temp), "=m" (v->counter) in atomic_sub() 74 :"Ir" (i), "m" (v->counter)); in atomic_sub() [all …]
|
D | local.h | 32 :"=&r" (temp), "=m" (l->a.counter), "=&r" (result) in local_add_return() 33 :"Ir" (i), "m" (l->a.counter) : "memory"); in local_add_return() 49 :"=&r" (temp), "=m" (l->a.counter), "=&r" (result) in local_sub_return() 50 :"Ir" (i), "m" (l->a.counter) : "memory"); in local_sub_return() 55 (cmpxchg_local(&((l)->a.counter), (o), (n))) 56 #define local_xchg(l, n) (xchg_local(&((l)->a.counter), (n))) 96 #define __local_inc(l) ((l)->a.counter++) 97 #define __local_dec(l) ((l)->a.counter++) 98 #define __local_add(i,l) ((l)->a.counter+=(i)) 99 #define __local_sub(i,l) ((l)->a.counter-=(i))
|
/arch/cris/include/asm/ |
D | atomic.h | 18 #define atomic_read(v) (*(volatile int *)&(v)->counter) 19 #define atomic_set(v,i) (((v)->counter) = (i)) 27 v->counter += i; in atomic_add() 35 v->counter -= i; in atomic_sub() 44 retval = (v->counter += i); in atomic_add_return() 56 retval = (v->counter -= i); in atomic_sub_return() 66 retval = (v->counter -= i) == 0; in atomic_sub_and_test() 75 (v->counter)++; in atomic_inc() 83 (v->counter)--; in atomic_dec() 92 retval = ++(v->counter); in atomic_inc_return() [all …]
|
/arch/avr32/include/asm/ |
D | atomic.h | 22 #define atomic_read(v) (*(volatile int *)&(v)->counter) 23 #define atomic_set(v, i) (((v)->counter) = i) 43 : "=&r"(result), "=o"(v->counter) in atomic_sub_return() 44 : "m"(v->counter), "rKs21"(i) in atomic_sub_return() 71 : "=&r"(result), "=o"(v->counter) in atomic_add_return() 72 : "m"(v->counter), "r"(i) in atomic_add_return() 101 : "=&r"(tmp), "=o"(v->counter) in atomic_sub_unless() 102 : "m"(v->counter), "rKs21"(a), "rKs21"(u) in atomic_sub_unless() 132 : "=&r"(tmp), "=o"(v->counter) in __atomic_add_unless() 133 : "m"(v->counter), "r"(a), "ir"(u) in __atomic_add_unless() [all …]
|
/arch/h8300/include/asm/ |
D | atomic.h | 14 #define atomic_read(v) (*(volatile int *)&(v)->counter) 15 #define atomic_set(v, i) (((v)->counter) = i) 24 ret = v->counter += i; in atomic_add_return() 37 ret = v->counter -= i; in atomic_sub_return() 50 v->counter++; in atomic_inc_return() 51 ret = v->counter; in atomic_inc_return() 73 --v->counter; in atomic_dec_return() 74 ret = v->counter; in atomic_dec_return() 86 --v->counter; in atomic_dec_and_test() 87 ret = v->counter; in atomic_dec_and_test() [all …]
|
/arch/sh/include/asm/ |
D | atomic-irq.h | 16 v->counter += i; in atomic_add() 25 v->counter -= i; in atomic_sub() 34 temp = v->counter; in atomic_add_return() 36 v->counter = temp; in atomic_add_return() 47 temp = v->counter; in atomic_sub_return() 49 v->counter = temp; in atomic_sub_return() 60 v->counter &= ~mask; in atomic_clear_mask() 69 v->counter |= mask; in atomic_set_mask()
|
D | atomic-llsc.h | 19 : "r" (i), "r" (&v->counter) in atomic_add() 33 : "r" (i), "r" (&v->counter) in atomic_sub() 56 : "r" (i), "r" (&v->counter) in atomic_add_return() 73 : "r" (i), "r" (&v->counter) in atomic_sub_return() 89 : "r" (~mask), "r" (&v->counter) in atomic_clear_mask() 103 : "r" (mask), "r" (&v->counter) in atomic_set_mask()
|
/arch/frv/include/asm/ |
D | atomic.h | 39 #define atomic_read(v) (*(volatile int *)&(v)->counter) 40 #define atomic_set(v, i) (((v)->counter) = (i)) 56 : "+U"(v->counter), "=&r"(val) in atomic_add_return() 77 : "+U"(v->counter), "=&r"(val) in atomic_sub_return() 128 volatile long long counter; member 135 long long counter; in atomic64_read() local 138 : "=e"(counter) in atomic64_read() 139 : "m"(v->counter)); in atomic64_read() 140 return counter; in atomic64_read() 146 : "=m"(v->counter) in atomic64_set() [all …]
|
/arch/parisc/include/asm/ |
D | spinlock.h | 78 rw->counter++; in arch_read_lock() 90 rw->counter--; in arch_read_unlock() 103 rw->counter++; in arch_read_trylock() 111 if (rw->counter < 0) in arch_read_trylock() 115 while (arch_spin_is_locked(&rw->lock) && rw->counter >= 0) in arch_read_trylock() 130 if (rw->counter != 0) { in arch_write_lock() 134 while (rw->counter != 0) in arch_write_lock() 140 rw->counter = -1; /* mark as write-locked */ in arch_write_lock() 147 rw->counter = 0; in arch_write_unlock() 160 if (rw->counter == 0) { in arch_write_trylock() [all …]
|
/arch/mn10300/include/asm/ |
D | atomic.h | 37 #define atomic_read(v) (ACCESS_ONCE((v)->counter)) 47 #define atomic_set(v, i) (((v)->counter) = (i)) 72 : "=&r"(status), "=&r"(retval), "=m"(v->counter) in atomic_add_return() 73 : "a"(ATOMIC_OPS_BASE_ADDR), "r"(&v->counter), "r"(i) in atomic_add_return() 80 retval = v->counter; in atomic_add_return() 82 v->counter = retval; in atomic_add_return() 111 : "=&r"(status), "=&r"(retval), "=m"(v->counter) in atomic_sub_return() 112 : "a"(ATOMIC_OPS_BASE_ADDR), "r"(&v->counter), "r"(i) in atomic_sub_return() 118 retval = v->counter; in atomic_sub_return() 120 v->counter = retval; in atomic_sub_return() [all …]
|
/arch/tile/include/asm/ |
D | atomic_64.h | 27 #define atomic_set(v, i) ((v)->counter = (i)) 40 val = __insn_cmpexch4((void *)&v->counter, n); in atomic_cmpxchg() 49 val = __insn_exch4((void *)&v->counter, n); in atomic_xchg() 56 __insn_fetchadd4((void *)&v->counter, i); in atomic_add() 63 val = __insn_fetchadd4((void *)&v->counter, i) + i; in atomic_add_return() 70 int guess, oldval = v->counter; in __atomic_add_unless() 84 #define atomic64_read(v) ((v)->counter) 85 #define atomic64_set(v, i) ((v)->counter = (i)) 92 val = __insn_cmpexch((void *)&v->counter, n); in atomic64_cmpxchg() 101 val = __insn_exch((void *)&v->counter, n); in atomic64_xchg() [all …]
|
/arch/m32r/include/asm/ |
D | local.h | 27 typedef struct { volatile int counter; } local_t; member 37 #define local_read(l) ((l)->counter) 46 #define local_set(l, i) (((l)->counter) = (i)) 68 : "r" (&l->counter), "r" (i) in local_add_return() 99 : "r" (&l->counter), "r" (i) in local_sub_return() 158 : "r" (&l->counter) in local_inc_return() 188 : "r" (&l->counter) in local_dec_return() 246 #define local_cmpxchg(l, o, n) (cmpxchg_local(&((l)->counter), (o), (n))) 247 #define local_xchg(v, new) (xchg_local(&((l)->counter), new)) 330 #define __local_inc(l) ((l)->a.counter++) [all …]
|
/arch/s390/include/asm/ |
D | atomic.h | 31 "=Q" (((atomic_t *)(ptr))->counter) \ 32 : "d" (op_val), "Q" (((atomic_t *)(ptr))->counter) \ 43 : "=d" (c) : "Q" (v->counter)); in atomic_read() 51 : "=Q" (v->counter) : "d" (i)); in atomic_set() 84 #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) 90 : "+d" (old), "=Q" (v->counter) in atomic_cmpxchg() 91 : "d" (new), "Q" (v->counter) in atomic_cmpxchg() 127 "=Q" (((atomic_t *)(ptr))->counter) \ 128 : "d" (op_val), "Q" (((atomic_t *)(ptr))->counter) \ 139 : "=d" (c) : "Q" (v->counter)); in atomic64_read() [all …]
|
/arch/blackfin/include/asm/ |
D | atomic.h | 23 #define atomic_read(v) __raw_uncached_fetch_asm(&(v)->counter) 25 #define atomic_add_return(i, v) __raw_atomic_update_asm(&(v)->counter, i) 26 #define atomic_sub_return(i, v) __raw_atomic_update_asm(&(v)->counter, -(i)) 28 #define atomic_clear_mask(m, v) __raw_atomic_clear_asm(&(v)->counter, m) 29 #define atomic_set_mask(m, v) __raw_atomic_set_asm(&(v)->counter, m)
|
/arch/hexagon/include/asm/ |
D | atomic.h | 29 #define atomic_set(v, i) ((v)->counter = (i)) 37 #define atomic_read(v) ((v)->counter) 44 #define atomic_xchg(v, new) (xchg(&((v)->counter), (new))) 76 : "r" (&v->counter), "r" (old), "r" (new) in atomic_cmpxchg() 93 : "r" (&v->counter), "r" (i) in atomic_add_return() 111 : "r" (&v->counter), "r" (i) in atomic_sub_return()
|
/arch/sparc/include/asm/ |
D | atomic_64.h | 16 #define atomic_read(v) (*(volatile int *)&(v)->counter) 17 #define atomic64_read(v) (*(volatile long *)&(v)->counter) 19 #define atomic_set(v, i) (((v)->counter) = i) 20 #define atomic64_set(v, i) (((v)->counter) = i) 70 #define atomic_cmpxchg(v, o, n) (cmpxchg(&((v)->counter), (o), (n))) 71 #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) 89 ((__typeof__((v)->counter))cmpxchg(&((v)->counter), (o), (n))) 90 #define atomic64_xchg(v, new) (xchg(&((v)->counter), new))
|