Home
last modified time | relevance | path

Searched refs:locked (Results 1 – 11 of 11) sorted by relevance

/kernel/locking/
Dqspinlock_paravirt.h91 (cmpxchg_acquire(&lock->locked, 0, _Q_LOCKED_VAL) == 0)) { in pv_hybrid_queued_unfair_trylock()
121 return !READ_ONCE(lock->locked) && in trylock_clear_pending()
302 if (READ_ONCE(node->locked)) in pv_wait_node()
322 if (!READ_ONCE(node->locked)) { in pv_wait_node()
343 !READ_ONCE(node->locked)); in pv_wait_node()
391 WRITE_ONCE(lock->locked, _Q_SLOW_VAL); in pv_kick_node()
456 if (xchg(&lock->locked, _Q_SLOW_VAL) == 0) { in pv_wait_head_or_lock()
462 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in pv_wait_head_or_lock()
470 pv_wait(&lock->locked, _Q_SLOW_VAL); in pv_wait_head_or_lock()
493 __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked) in __pv_queued_spin_unlock_slowpath() argument
[all …]
Dosq_lock.c97 node->locked = 0; in osq_lock()
137 while (!READ_ONCE(node->locked)) { in osq_lock()
169 if (smp_load_acquire(&node->locked)) in osq_lock()
224 WRITE_ONCE(next->locked, 1); in osq_unlock()
230 WRITE_ONCE(next->locked, 1); in osq_unlock()
Dmcs_spinlock.h20 int locked; /* 1 if lock acquired */ member
70 node->locked = 0; in mcs_spin_lock()
94 arch_mcs_spin_lock_contended(&node->locked); in mcs_spin_lock()
118 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
Dqspinlock.c262 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in set_locked()
433 node->locked = 0; in queued_spin_lock_slowpath()
473 arch_mcs_spin_lock_contended(&node->locked); in queued_spin_lock_slowpath()
508 goto locked; in queued_spin_lock_slowpath()
512 locked: in queued_spin_lock_slowpath()
552 arch_mcs_spin_unlock_contended(&next->locked); in queued_spin_lock_slowpath()
Dmutex.c1412 bool locked; in mutex_trylock() local
1418 locked = __mutex_trylock(lock); in mutex_trylock()
1419 if (locked) in mutex_trylock()
1422 return locked; in mutex_trylock()
Dlockdep.c5124 int locked; in lockdep_reset_lock_reg() local
5127 locked = graph_lock(); in lockdep_reset_lock_reg()
5128 if (!locked) in lockdep_reset_lock_reg()
/kernel/bpf/
Dlocal_storage.c50 bool locked) in cgroup_storage_lookup() argument
55 if (!locked) in cgroup_storage_lookup()
72 if (!locked) in cgroup_storage_lookup()
78 if (!locked) in cgroup_storage_lookup()
/kernel/
Dfutex.c2625 static int fixup_owner(u32 __user *uaddr, struct futex_q *q, int locked) in fixup_owner() argument
2629 if (locked) { in fixup_owner()
2668 return ret ? ret : locked; in fixup_owner()
/kernel/power/
DKconfig203 locked up attempting to suspend/resume a device.
/kernel/trace/
Dring_buffer.c4062 rb_reader_unlock(struct ring_buffer_per_cpu *cpu_buffer, bool locked) in rb_reader_unlock() argument
4064 if (likely(locked)) in rb_reader_unlock()
/kernel/events/
Dcore.c5716 unsigned long locked, lock_limit; in perf_mmap() local
5861 locked = atomic64_read(&vma->vm_mm->pinned_vm) + extra; in perf_mmap()
5863 if ((locked > lock_limit) && perf_is_paranoid() && in perf_mmap()