Searched refs:raw_lock (Results 1 – 11 of 11) sorted by relevance
26 lock->raw_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED; in __raw_spin_lock_init()44 lock->raw_lock = (arch_rwlock_t) __ARCH_RW_LOCK_UNLOCKED; in __rwlock_init()113 arch_spin_lock(&lock->raw_lock); in do_raw_spin_lock()120 int ret = arch_spin_trylock(&lock->raw_lock); in do_raw_spin_trylock()139 arch_spin_unlock(&lock->raw_lock); in do_raw_spin_unlock()158 arch_read_lock(&lock->raw_lock); in do_raw_read_lock()163 int ret = arch_read_trylock(&lock->raw_lock); in do_raw_read_trylock()177 arch_read_unlock(&lock->raw_lock); in do_raw_read_unlock()207 arch_write_lock(&lock->raw_lock); in do_raw_write_lock()213 int ret = arch_write_trylock(&lock->raw_lock); in do_raw_write_trylock()[all …]
76 arch_##op##_relax(&lock->raw_lock); \92 arch_##op##_relax(&lock->raw_lock); \
50 # define do_raw_read_lock(rwlock) do {__acquire(lock); arch_read_lock(&(rwlock)->raw_lock); } while…52 do {__acquire(lock); arch_read_lock_flags(&(lock)->raw_lock, *(flags)); } while (0)53 # define do_raw_read_trylock(rwlock) arch_read_trylock(&(rwlock)->raw_lock)54 # define do_raw_read_unlock(rwlock) do {arch_read_unlock(&(rwlock)->raw_lock); __release(lock); } w…55 # define do_raw_write_lock(rwlock) do {__acquire(lock); arch_write_lock(&(rwlock)->raw_lock); } whi…57 do {__acquire(lock); arch_write_lock_flags(&(lock)->raw_lock, *(flags)); } while (0)58 # define do_raw_write_trylock(rwlock) arch_write_trylock(&(rwlock)->raw_lock)59 # define do_raw_write_unlock(rwlock) do {arch_write_unlock(&(rwlock)->raw_lock); __release(lock); }…
12 arch_rwlock_t raw_lock; member36 (rwlock_t) { .raw_lock = __ARCH_RW_LOCK_UNLOCKED, \43 (rwlock_t) { .raw_lock = __ARCH_RW_LOCK_UNLOCKED, \
21 arch_spinlock_t raw_lock; member62 .raw_lock = __ARCH_SPIN_LOCK_UNLOCKED, \87 .raw_lock = __ARCH_SPIN_LOCK_UNLOCKED, \
111 #define raw_spin_is_locked(lock) arch_spin_is_locked(&(lock)->raw_lock)114 #define raw_spin_is_contended(lock) arch_spin_is_contended(&(lock)->raw_lock)183 arch_spin_lock(&lock->raw_lock); in do_raw_spin_lock()195 arch_spin_lock_flags(&lock->raw_lock, *flags); in do_raw_spin_lock_flags()201 int ret = arch_spin_trylock(&(lock)->raw_lock); in do_raw_spin_trylock()212 arch_spin_unlock(&lock->raw_lock); in do_raw_spin_unlock()
172 static DEFINE_RWLOCK(raw_lock);176 write_lock_bh(&raw_lock); in raw_hash()179 write_unlock_bh(&raw_lock); in raw_hash()186 write_lock_bh(&raw_lock); in raw_unhash()189 write_unlock_bh(&raw_lock); in raw_unhash()362 read_lock(&raw_lock); in ieee802154_raw_deliver()375 read_unlock(&raw_lock); in ieee802154_raw_deliver()
252 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()267 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()281 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_enter()936 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()948 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_zero()964 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()979 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_do_h_page_init_copy()
551 arch_spin_lock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()596 arch_spin_unlock(&kvm->mmu_lock.rlock.raw_lock); in kvmppc_rm_h_put_tce_indirect()
16 while (likely(arch_spin_value_unlocked(old.lock.rlock.raw_lock))) { \
84 raw_spinlock_t raw_lock; member142 raw_spin_lock_init(&htab->buckets[i].raw_lock); in htab_init_buckets()154 raw_spin_lock_irqsave(&b->raw_lock, flags); in htab_lock_bucket()165 raw_spin_unlock_irqrestore(&b->raw_lock, flags); in htab_unlock_bucket()