/arch/parisc/include/asm/ |
D | spinlock.h | 9 static inline int arch_spin_is_locked(arch_spinlock_t *x) in arch_spin_is_locked() function 17 do { cpu_relax(); } while (arch_spin_is_locked(x)) 111 while (arch_spin_is_locked(&rw->lock) && rw->counter >= 0) in arch_read_trylock()
|
/arch/hexagon/include/asm/ |
D | spinlock.h | 180 do {while (arch_spin_is_locked(lock)) cpu_relax(); } while (0) 181 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
|
/arch/xtensa/include/asm/ |
D | spinlock.h | 31 #define arch_spin_is_locked(x) ((x)->slock != 0) macro 33 do { while (arch_spin_is_locked(lock)) cpu_relax(); } while (0)
|
/arch/sh/include/asm/ |
D | spinlock.h | 26 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro 29 do { while (arch_spin_is_locked(x)) cpu_relax(); } while (0)
|
/arch/blackfin/include/asm/ |
D | spinlock.h | 27 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function 51 while (arch_spin_is_locked(lock)) in arch_spin_unlock_wait()
|
/arch/mn10300/include/asm/ |
D | spinlock.h | 25 #define arch_spin_is_locked(x) (*(volatile signed char *)(&(x)->slock) != 0) macro 26 #define arch_spin_unlock_wait(x) do { barrier(); } while (arch_spin_is_locked(x))
|
/arch/sparc/include/asm/ |
D | spinlock_32.h | 14 #define arch_spin_is_locked(lock) (*((volatile unsigned char *)(lock)) != 0) macro 17 do { while (arch_spin_is_locked(lock)) cpu_relax(); } while (0)
|
D | spinlock_64.h | 24 #define arch_spin_is_locked(lp) ((lp)->lock != 0) macro
|
/arch/m32r/include/asm/ |
D | spinlock.h | 28 #define arch_spin_is_locked(x) (*(volatile int *)(&(x)->slock) <= 0) macro 31 do { cpu_relax(); } while (arch_spin_is_locked(x))
|
/arch/metag/include/asm/ |
D | spinlock.h | 11 do { while (arch_spin_is_locked(lock)) cpu_relax(); } while (0)
|
D | spinlock_lnkget.h | 10 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
D | spinlock_lock1.h | 7 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/arch/powerpc/include/asm/ |
D | spinlock.h | 60 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function 169 do { while (arch_spin_is_locked(lock)) cpu_relax(); } while (0)
|
/arch/arm/include/asm/ |
D | spinlock.h | 54 do { while (arch_spin_is_locked(lock)) cpu_relax(); } while (0) 121 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/arch/tile/include/asm/ |
D | spinlock_32.h | 36 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
D | spinlock_64.h | 47 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/arch/mips/include/asm/ |
D | spinlock.h | 38 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function 52 while (arch_spin_is_locked(x)) { cpu_relax(); }
|
/arch/s390/include/asm/ |
D | spinlock.h | 54 static inline int arch_spin_is_locked(arch_spinlock_t *lp) in arch_spin_is_locked() function 98 while (arch_spin_is_locked(lock)) in arch_spin_unlock_wait()
|
/arch/alpha/include/asm/ |
D | spinlock.h | 15 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
|
/arch/arm64/include/asm/ |
D | spinlock.h | 153 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/arch/arc/include/asm/ |
D | spinlock.h | 16 #define arch_spin_is_locked(x) ((x)->slock != __ARCH_SPIN_LOCK_UNLOCKED__) macro 19 do { while (arch_spin_is_locked(x)) cpu_relax(); } while (0)
|
/arch/x86/include/asm/ |
D | spinlock.h | 164 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|
/arch/ia64/include/asm/ |
D | spinlock.h | 110 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
|