Home
last modified time | relevance | path

Searched refs:arch_spin_is_locked (Results 1 – 23 of 23) sorted by relevance

/arch/parisc/include/asm/
Dspinlock.h9 static inline int arch_spin_is_locked(arch_spinlock_t *x) in arch_spin_is_locked() function
120 while (arch_spin_is_locked(&rw->lock) && rw->counter >= 0) in arch_read_trylock()
/arch/tile/include/asm/
Dspinlock_32.h36 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
Dspinlock_64.h47 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/s390/include/asm/
Dspinlock.h56 static inline int arch_spin_is_locked(arch_spinlock_t *lp) in arch_spin_is_locked() function
100 while (arch_spin_is_locked(lock)) in arch_spin_unlock_wait()
/arch/hexagon/include/asm/
Dspinlock.h187 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
/arch/metag/include/asm/
Dspinlock_lnkget.h10 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
Dspinlock_lock1.h7 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/xtensa/include/asm/
Dspinlock.h34 #define arch_spin_is_locked(x) ((x)->slock != 0) macro
/arch/alpha/include/asm/
Dspinlock.h17 #define arch_spin_is_locked(x) ((x)->lock != 0) macro
/arch/sh/include/asm/
Dspinlock-cas.h29 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro
Dspinlock-llsc.h21 #define arch_spin_is_locked(x) ((x)->lock <= 0) macro
/arch/blackfin/include/asm/
Dspinlock.h29 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/sparc/include/asm/
Dspinlock_64.h27 #define arch_spin_is_locked(lp) ((lp)->lock != 0) macro
Dspinlock_32.h15 #define arch_spin_is_locked(lock) (*((volatile unsigned char *)(lock)) != 0) macro
/arch/mn10300/include/asm/
Dspinlock.h27 #define arch_spin_is_locked(x) (*(volatile signed char *)(&(x)->slock) != 0) macro
/arch/m32r/include/asm/
Dspinlock.h30 #define arch_spin_is_locked(x) (*(volatile int *)(&(x)->slock) <= 0) macro
/arch/arm64/include/asm/
Dspinlock.h177 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/arm/include/asm/
Dspinlock.h136 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/powerpc/include/asm/
Dspinlock.h60 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/arc/include/asm/
Dspinlock.h16 #define arch_spin_is_locked(x) ((x)->slock != __ARCH_SPIN_LOCK_UNLOCKED__) macro
/arch/mips/include/asm/
Dspinlock.h39 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/ia64/include/asm/
Dspinlock.h114 static inline int arch_spin_is_locked(arch_spinlock_t *lock) in arch_spin_is_locked() function
/arch/x86/kernel/
Dhpet.c813 if (arch_spin_is_locked(&old.lock)) in read_hpet()
845 } while ((new.value == old.value) && arch_spin_is_locked(&new.lock)); in read_hpet()