/arch/sh/mm/ |
D | flush-sh4.c | 15 reg_size_t aligned_start, v, cnt, end; in sh4__flush_wback_region() local 18 v = aligned_start & ~(L1_CACHE_BYTES-1); in sh4__flush_wback_region() 21 cnt = (end - v) / L1_CACHE_BYTES; in sh4__flush_wback_region() 24 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 25 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 26 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 27 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 28 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 29 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() 30 __ocbwb(v); v += L1_CACHE_BYTES; in sh4__flush_wback_region() [all …]
|
/arch/frv/include/asm/ |
D | atomic.h | 39 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 40 #define atomic_set(v, i) (((v)->counter) = (i)) argument 43 static inline int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 56 : "+U"(v->counter), "=&r"(val) in atomic_add_return() 64 static inline int atomic_sub_return(int i, atomic_t *v) in atomic_sub_return() argument 77 : "+U"(v->counter), "=&r"(val) in atomic_sub_return() 87 extern int atomic_add_return(int i, atomic_t *v); 88 extern int atomic_sub_return(int i, atomic_t *v); 92 static inline int atomic_add_negative(int i, atomic_t *v) in atomic_add_negative() argument 94 return atomic_add_return(i, v) < 0; in atomic_add_negative() [all …]
|
/arch/sparc/include/asm/ |
D | atomic_64.h | 16 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 17 #define atomic64_read(v) (*(volatile long *)&(v)->counter) argument 19 #define atomic_set(v, i) (((v)->counter) = i) argument 20 #define atomic64_set(v, i) (((v)->counter) = i) argument 32 #define atomic_dec_return(v) atomic_sub_ret(1, v) argument 33 #define atomic64_dec_return(v) atomic64_sub_ret(1, v) argument 35 #define atomic_inc_return(v) atomic_add_ret(1, v) argument 36 #define atomic64_inc_return(v) atomic64_add_ret(1, v) argument 38 #define atomic_sub_return(i, v) atomic_sub_ret(i, v) argument 39 #define atomic64_sub_return(i, v) atomic64_sub_ret(i, v) argument [all …]
|
D | atomic_32.h | 24 #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) argument 28 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 30 #define atomic_add(i, v) ((void)__atomic_add_return( (int)(i), (v))) argument 31 #define atomic_sub(i, v) ((void)__atomic_add_return(-(int)(i), (v))) argument 32 #define atomic_inc(v) ((void)__atomic_add_return( 1, (v))) argument 33 #define atomic_dec(v) ((void)__atomic_add_return( -1, (v))) argument 35 #define atomic_add_return(i, v) (__atomic_add_return( (int)(i), (v))) argument 36 #define atomic_sub_return(i, v) (__atomic_add_return(-(int)(i), (v))) argument 37 #define atomic_inc_return(v) (__atomic_add_return( 1, (v))) argument 38 #define atomic_dec_return(v) (__atomic_add_return( -1, (v))) argument [all …]
|
/arch/alpha/include/asm/ |
D | atomic.h | 20 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 21 #define atomic64_read(v) (*(volatile long *)&(v)->counter) argument 23 #define atomic_set(v,i) ((v)->counter = (i)) argument 24 #define atomic64_set(v,i) ((v)->counter = (i)) argument 32 static __inline__ void atomic_add(int i, atomic_t * v) in atomic_add() argument 43 :"=&r" (temp), "=m" (v->counter) in atomic_add() 44 :"Ir" (i), "m" (v->counter)); in atomic_add() 47 static __inline__ void atomic64_add(long i, atomic64_t * v) in atomic64_add() argument 58 :"=&r" (temp), "=m" (v->counter) in atomic64_add() 59 :"Ir" (i), "m" (v->counter)); in atomic64_add() [all …]
|
/arch/parisc/include/asm/ |
D | atomic.h | 62 static __inline__ int __atomic_add_return(int i, atomic_t *v) in __atomic_add_return() argument 66 _atomic_spin_lock_irqsave(v, flags); in __atomic_add_return() 68 ret = (v->counter += i); in __atomic_add_return() 70 _atomic_spin_unlock_irqrestore(v, flags); in __atomic_add_return() 74 static __inline__ void atomic_set(atomic_t *v, int i) in atomic_set() argument 77 _atomic_spin_lock_irqsave(v, flags); in atomic_set() 79 v->counter = i; in atomic_set() 81 _atomic_spin_unlock_irqrestore(v, flags); in atomic_set() 84 static __inline__ int atomic_read(const atomic_t *v) in atomic_read() argument 86 return (*(volatile int *)&(v)->counter); in atomic_read() [all …]
|
/arch/cris/include/asm/ |
D | atomic.h | 18 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 19 #define atomic_set(v,i) (((v)->counter) = (i)) argument 23 static inline void atomic_add(int i, volatile atomic_t *v) in atomic_add() argument 26 cris_atomic_save(v, flags); in atomic_add() 27 v->counter += i; in atomic_add() 28 cris_atomic_restore(v, flags); in atomic_add() 31 static inline void atomic_sub(int i, volatile atomic_t *v) in atomic_sub() argument 34 cris_atomic_save(v, flags); in atomic_sub() 35 v->counter -= i; in atomic_sub() 36 cris_atomic_restore(v, flags); in atomic_sub() [all …]
|
/arch/ia64/include/asm/ |
D | atomic.h | 23 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 24 #define atomic64_read(v) (*(volatile long *)&(v)->counter) argument 26 #define atomic_set(v,i) (((v)->counter) = (i)) argument 27 #define atomic64_set(v,i) (((v)->counter) = (i)) argument 30 ia64_atomic_add (int i, atomic_t *v) in ia64_atomic_add() argument 36 CMPXCHG_BUGCHECK(v); in ia64_atomic_add() 37 old = atomic_read(v); in ia64_atomic_add() 39 } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic_t)) != old); in ia64_atomic_add() 44 ia64_atomic64_add (__s64 i, atomic64_t *v) in ia64_atomic64_add() argument 50 CMPXCHG_BUGCHECK(v); in ia64_atomic64_add() [all …]
|
/arch/mips/include/asm/ |
D | atomic.h | 32 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 41 #define atomic_set(v, i) ((v)->counter = (i)) argument 50 static __inline__ void atomic_add(int i, atomic_t * v) in atomic_add() argument 62 : "=&r" (temp), "=m" (v->counter) in atomic_add() 63 : "Ir" (i), "m" (v->counter)); in atomic_add() 74 : "=&r" (temp), "=m" (v->counter) in atomic_add() 75 : "Ir" (i), "m" (v->counter)); in atomic_add() 81 v->counter += i; in atomic_add() 93 static __inline__ void atomic_sub(int i, atomic_t * v) in atomic_sub() argument 105 : "=&r" (temp), "=m" (v->counter) in atomic_sub() [all …]
|
/arch/avr32/include/asm/ |
D | atomic.h | 22 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 23 #define atomic_set(v, i) (((v)->counter) = i) argument 32 static inline int atomic_sub_return(int i, atomic_t *v) in atomic_sub_return() argument 43 : "=&r"(result), "=o"(v->counter) in atomic_sub_return() 44 : "m"(v->counter), "rKs21"(i) in atomic_sub_return() 57 static inline int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 62 result = atomic_sub_return(-i, v); in atomic_add_return() 71 : "=&r"(result), "=o"(v->counter) in atomic_add_return() 72 : "m"(v->counter), "r"(i) in atomic_add_return() 87 static inline void atomic_sub_unless(atomic_t *v, int a, int u) in atomic_sub_unless() argument [all …]
|
/arch/x86/lib/ |
D | atomic64_386_32.S | 34 LOCK v; 39 UNLOCK v; \ 46 #define v %ecx macro 48 movl (v), %eax 49 movl 4(v), %edx 51 #undef v 53 #define v %esi macro 55 movl %ebx, (v) 56 movl %ecx, 4(v) 58 #undef v [all …]
|
/arch/x86/include/asm/ |
D | atomic64_64.h | 19 static inline long atomic64_read(const atomic64_t *v) in atomic64_read() argument 21 return (*(volatile long *)&(v)->counter); in atomic64_read() 31 static inline void atomic64_set(atomic64_t *v, long i) in atomic64_set() argument 33 v->counter = i; in atomic64_set() 43 static inline void atomic64_add(long i, atomic64_t *v) in atomic64_add() argument 46 : "=m" (v->counter) in atomic64_add() 47 : "er" (i), "m" (v->counter)); in atomic64_add() 57 static inline void atomic64_sub(long i, atomic64_t *v) in atomic64_sub() argument 60 : "=m" (v->counter) in atomic64_sub() 61 : "er" (i), "m" (v->counter)); in atomic64_sub() [all …]
|
D | atomic.h | 23 static inline int atomic_read(const atomic_t *v) in atomic_read() argument 25 return (*(volatile int *)&(v)->counter); in atomic_read() 35 static inline void atomic_set(atomic_t *v, int i) in atomic_set() argument 37 v->counter = i; in atomic_set() 47 static inline void atomic_add(int i, atomic_t *v) in atomic_add() argument 50 : "+m" (v->counter) in atomic_add() 61 static inline void atomic_sub(int i, atomic_t *v) in atomic_sub() argument 64 : "+m" (v->counter) in atomic_sub() 77 static inline int atomic_sub_and_test(int i, atomic_t *v) in atomic_sub_and_test() argument 82 : "+m" (v->counter), "=qm" (c) in atomic_sub_and_test() [all …]
|
/arch/hexagon/include/asm/ |
D | atomic.h | 29 #define atomic_set(v, i) ((v)->counter = (i)) argument 37 #define atomic_read(v) ((v)->counter) argument 44 #define atomic_xchg(v, new) (xchg(&((v)->counter), (new))) argument 64 static inline int atomic_cmpxchg(atomic_t *v, int old, int new) in atomic_cmpxchg() argument 76 : "r" (&v->counter), "r" (old), "r" (new) in atomic_cmpxchg() 83 static inline int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 93 : "r" (&v->counter), "r" (i) in atomic_add_return() 100 #define atomic_add(i, v) atomic_add_return(i, (v)) argument 102 static inline int atomic_sub_return(int i, atomic_t *v) in atomic_sub_return() argument 111 : "r" (&v->counter), "r" (i) in atomic_sub_return() [all …]
|
/arch/powerpc/include/asm/ |
D | atomic.h | 14 static __inline__ int atomic_read(const atomic_t *v) in atomic_read() argument 18 __asm__ __volatile__("lwz%U1%X1 %0,%1" : "=r"(t) : "m"(v->counter)); in atomic_read() 23 static __inline__ void atomic_set(atomic_t *v, int i) in atomic_set() argument 25 __asm__ __volatile__("stw%U0%X0 %1,%0" : "=m"(v->counter) : "r"(i)); in atomic_set() 28 static __inline__ void atomic_add(int a, atomic_t *v) in atomic_add() argument 38 : "=&r" (t), "+m" (v->counter) in atomic_add() 39 : "r" (a), "r" (&v->counter) in atomic_add() 43 static __inline__ int atomic_add_return(int a, atomic_t *v) in atomic_add_return() argument 56 : "r" (a), "r" (&v->counter) in atomic_add_return() 62 #define atomic_add_negative(a, v) (atomic_add_return((a), (v)) < 0) argument [all …]
|
/arch/arm/include/asm/ |
D | atomic.h | 29 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 30 #define atomic_set(v,i) (((v)->counter) = (i)) argument 39 static inline void atomic_add(int i, atomic_t *v) in atomic_add() argument 50 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_add() 51 : "r" (&v->counter), "Ir" (i) in atomic_add() 55 static inline int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 68 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_add_return() 69 : "r" (&v->counter), "Ir" (i) in atomic_add_return() 77 static inline void atomic_sub(int i, atomic_t *v) in atomic_sub() argument 88 : "=&r" (result), "=&r" (tmp), "+Qo" (v->counter) in atomic_sub() [all …]
|
/arch/tile/include/asm/ |
D | atomic_64.h | 27 #define atomic_set(v, i) ((v)->counter = (i)) argument 35 static inline int atomic_cmpxchg(atomic_t *v, int o, int n) in atomic_cmpxchg() argument 40 val = __insn_cmpexch4((void *)&v->counter, n); in atomic_cmpxchg() 45 static inline int atomic_xchg(atomic_t *v, int n) in atomic_xchg() argument 49 val = __insn_exch4((void *)&v->counter, n); in atomic_xchg() 54 static inline void atomic_add(int i, atomic_t *v) in atomic_add() argument 56 __insn_fetchadd4((void *)&v->counter, i); in atomic_add() 59 static inline int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 63 val = __insn_fetchadd4((void *)&v->counter, i) + i; in atomic_add_return() 68 static inline int __atomic_add_unless(atomic_t *v, int a, int u) in __atomic_add_unless() argument [all …]
|
D | atomic.h | 35 static inline int atomic_read(const atomic_t *v) in atomic_read() argument 37 return ACCESS_ONCE(v->counter); in atomic_read() 47 #define atomic_sub_return(i, v) atomic_add_return((int)(-(i)), (v)) argument 56 #define atomic_sub(i, v) atomic_add((int)(-(i)), (v)) argument 66 #define atomic_sub_and_test(i, v) (atomic_sub_return((i), (v)) == 0) argument 74 #define atomic_inc_return(v) atomic_add_return(1, (v)) argument 82 #define atomic_dec_return(v) atomic_sub_return(1, (v)) argument 90 #define atomic_inc(v) atomic_add(1, (v)) argument 98 #define atomic_dec(v) atomic_sub(1, (v)) argument 106 #define atomic_dec_and_test(v) (atomic_dec_return(v) == 0) argument [all …]
|
D | atomic_32.h | 26 int _atomic_xchg(atomic_t *v, int n); 27 int _atomic_xchg_add(atomic_t *v, int i); 28 int _atomic_xchg_add_unless(atomic_t *v, int a, int u); 29 int _atomic_cmpxchg(atomic_t *v, int o, int n); 38 static inline int atomic_xchg(atomic_t *v, int n) in atomic_xchg() argument 41 return _atomic_xchg(v, n); in atomic_xchg() 53 static inline int atomic_cmpxchg(atomic_t *v, int o, int n) in atomic_cmpxchg() argument 56 return _atomic_cmpxchg(v, o, n); in atomic_cmpxchg() 66 static inline void atomic_add(int i, atomic_t *v) in atomic_add() argument 68 _atomic_xchg_add(v, i); in atomic_add() [all …]
|
/arch/m68k/include/asm/ |
D | atomic.h | 19 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 20 #define atomic_set(v, i) (((v)->counter) = i) argument 32 static inline void atomic_add(int i, atomic_t *v) in atomic_add() argument 34 __asm__ __volatile__("addl %1,%0" : "+m" (*v) : ASM_DI (i)); in atomic_add() 37 static inline void atomic_sub(int i, atomic_t *v) in atomic_sub() argument 39 __asm__ __volatile__("subl %1,%0" : "+m" (*v) : ASM_DI (i)); in atomic_sub() 42 static inline void atomic_inc(atomic_t *v) in atomic_inc() argument 44 __asm__ __volatile__("addql #1,%0" : "+m" (*v)); in atomic_inc() 47 static inline void atomic_dec(atomic_t *v) in atomic_dec() argument 49 __asm__ __volatile__("subql #1,%0" : "+m" (*v)); in atomic_dec() [all …]
|
/arch/arm/mach-mxs/ |
D | regs-clkctrl-mx28.h | 36 #define BF_CLKCTRL_PLL0CTRL0_LFR_SEL(v) \ argument 37 (((v) << 28) & BM_CLKCTRL_PLL0CTRL0_LFR_SEL) 44 #define BF_CLKCTRL_PLL0CTRL0_CP_SEL(v) \ argument 45 (((v) << 24) & BM_CLKCTRL_PLL0CTRL0_CP_SEL) 52 #define BF_CLKCTRL_PLL0CTRL0_DIV_SEL(v) \ argument 53 (((v) << 20) & BM_CLKCTRL_PLL0CTRL0_DIV_SEL) 67 #define BF_CLKCTRL_PLL0CTRL1_LOCK_COUNT(v) \ argument 68 (((v) << 0) & BM_CLKCTRL_PLL0CTRL1_LOCK_COUNT) 78 #define BF_CLKCTRL_PLL1CTRL0_LFR_SEL(v) \ argument 79 (((v) << 28) & BM_CLKCTRL_PLL1CTRL0_LFR_SEL) [all …]
|
/arch/h8300/include/asm/ |
D | atomic.h | 14 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 15 #define atomic_set(v, i) (((v)->counter) = i) argument 19 static __inline__ int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 24 ret = v->counter += i; in atomic_add_return() 29 #define atomic_add(i, v) atomic_add_return(i, v) argument 30 #define atomic_add_negative(a, v) (atomic_add_return((a), (v)) < 0) argument 32 static __inline__ int atomic_sub_return(int i, atomic_t *v) in atomic_sub_return() argument 37 ret = v->counter -= i; in atomic_sub_return() 42 #define atomic_sub(i, v) atomic_sub_return(i, v) argument 43 #define atomic_sub_and_test(i,v) (atomic_sub_return(i, v) == 0) argument [all …]
|
/arch/sh/include/asm/ |
D | atomic.h | 16 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 17 #define atomic_set(v,i) ((v)->counter = (i)) argument 27 #define atomic_add_negative(a, v) (atomic_add_return((a), (v)) < 0) argument 28 #define atomic_dec_return(v) atomic_sub_return(1, (v)) argument 29 #define atomic_inc_return(v) atomic_add_return(1, (v)) argument 30 #define atomic_inc_and_test(v) (atomic_inc_return(v) == 0) argument 31 #define atomic_sub_and_test(i,v) (atomic_sub_return((i), (v)) == 0) argument 32 #define atomic_dec_and_test(v) (atomic_sub_return(1, (v)) == 0) argument 34 #define atomic_inc(v) atomic_add(1, (v)) argument 35 #define atomic_dec(v) atomic_sub(1, (v)) argument [all …]
|
/arch/xtensa/include/asm/ |
D | atomic.h | 49 #define atomic_read(v) (*(volatile int *)&(v)->counter) argument 58 #define atomic_set(v,i) ((v)->counter = (i)) argument 67 static inline void atomic_add(int i, atomic_t * v) in atomic_add() argument 79 : "a" (i), "a" (v) in atomic_add() 91 static inline void atomic_sub(int i, atomic_t *v) in atomic_sub() argument 103 : "a" (i), "a" (v) in atomic_sub() 112 static inline int atomic_add_return(int i, atomic_t * v) in atomic_add_return() argument 124 : "a" (i), "a" (v) in atomic_add_return() 131 static inline int atomic_sub_return(int i, atomic_t * v) in atomic_sub_return() argument 143 : "a" (i), "a" (v) in atomic_sub_return() [all …]
|
/arch/mn10300/include/asm/ |
D | atomic.h | 37 #define atomic_read(v) (ACCESS_ONCE((v)->counter)) argument 47 #define atomic_set(v, i) (((v)->counter) = (i)) argument 57 static inline int atomic_add_return(int i, atomic_t *v) in atomic_add_return() argument 72 : "=&r"(status), "=&r"(retval), "=m"(v->counter) in atomic_add_return() 73 : "a"(ATOMIC_OPS_BASE_ADDR), "r"(&v->counter), "r"(i) in atomic_add_return() 80 retval = v->counter; in atomic_add_return() 82 v->counter = retval; in atomic_add_return() 96 static inline int atomic_sub_return(int i, atomic_t *v) in atomic_sub_return() argument 111 : "=&r"(status), "=&r"(retval), "=m"(v->counter) in atomic_sub_return() 112 : "a"(ATOMIC_OPS_BASE_ADDR), "r"(&v->counter), "r"(i) in atomic_sub_return() [all …]
|