Searched refs:locked (Results 1 – 11 of 11) sorted by relevance
/kernel/locking/ |
D | qspinlock_paravirt.h | 91 (cmpxchg_acquire(&lock->locked, 0, _Q_LOCKED_VAL) == 0)) { in pv_hybrid_queued_unfair_trylock() 121 return !READ_ONCE(lock->locked) && in trylock_clear_pending() 302 if (READ_ONCE(node->locked)) in pv_wait_node() 322 if (!READ_ONCE(node->locked)) { in pv_wait_node() 343 !READ_ONCE(node->locked)); in pv_wait_node() 391 WRITE_ONCE(lock->locked, _Q_SLOW_VAL); in pv_kick_node() 456 if (xchg(&lock->locked, _Q_SLOW_VAL) == 0) { in pv_wait_head_or_lock() 462 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in pv_wait_head_or_lock() 470 pv_wait(&lock->locked, _Q_SLOW_VAL); in pv_wait_head_or_lock() 493 __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked) in __pv_queued_spin_unlock_slowpath() argument [all …]
|
D | osq_lock.c | 97 node->locked = 0; in osq_lock() 137 while (!READ_ONCE(node->locked)) { in osq_lock() 169 if (smp_load_acquire(&node->locked)) in osq_lock() 224 WRITE_ONCE(next->locked, 1); in osq_unlock() 230 WRITE_ONCE(next->locked, 1); in osq_unlock()
|
D | mcs_spinlock.h | 20 int locked; /* 1 if lock acquired */ member 70 node->locked = 0; in mcs_spin_lock() 94 arch_mcs_spin_lock_contended(&node->locked); in mcs_spin_lock() 118 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
|
D | qspinlock.c | 262 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in set_locked() 433 node->locked = 0; in queued_spin_lock_slowpath() 473 arch_mcs_spin_lock_contended(&node->locked); in queued_spin_lock_slowpath() 508 goto locked; in queued_spin_lock_slowpath() 512 locked: in queued_spin_lock_slowpath() 552 arch_mcs_spin_unlock_contended(&next->locked); in queued_spin_lock_slowpath()
|
D | mutex.c | 1412 bool locked; in mutex_trylock() local 1418 locked = __mutex_trylock(lock); in mutex_trylock() 1419 if (locked) in mutex_trylock() 1422 return locked; in mutex_trylock()
|
D | lockdep.c | 5124 int locked; in lockdep_reset_lock_reg() local 5127 locked = graph_lock(); in lockdep_reset_lock_reg() 5128 if (!locked) in lockdep_reset_lock_reg()
|
/kernel/bpf/ |
D | local_storage.c | 50 bool locked) in cgroup_storage_lookup() argument 55 if (!locked) in cgroup_storage_lookup() 72 if (!locked) in cgroup_storage_lookup() 78 if (!locked) in cgroup_storage_lookup()
|
/kernel/ |
D | futex.c | 2625 static int fixup_owner(u32 __user *uaddr, struct futex_q *q, int locked) in fixup_owner() argument 2629 if (locked) { in fixup_owner() 2668 return ret ? ret : locked; in fixup_owner()
|
/kernel/power/ |
D | Kconfig | 203 locked up attempting to suspend/resume a device.
|
/kernel/trace/ |
D | ring_buffer.c | 4062 rb_reader_unlock(struct ring_buffer_per_cpu *cpu_buffer, bool locked) in rb_reader_unlock() argument 4064 if (likely(locked)) in rb_reader_unlock()
|
/kernel/events/ |
D | core.c | 5716 unsigned long locked, lock_limit; in perf_mmap() local 5861 locked = atomic64_read(&vma->vm_mm->pinned_vm) + extra; in perf_mmap() 5863 if ((locked > lock_limit) && perf_is_paranoid() && in perf_mmap()
|