Lines Matching refs:lock
221 static inline void __bpf_spin_lock(struct bpf_spin_lock *lock) in __bpf_spin_lock() argument
223 arch_spinlock_t *l = (void *)lock; in __bpf_spin_lock()
226 arch_spinlock_t lock; in __bpf_spin_lock() member
227 } u = { .lock = __ARCH_SPIN_LOCK_UNLOCKED }; in __bpf_spin_lock()
231 BUILD_BUG_ON(sizeof(*lock) != sizeof(__u32)); in __bpf_spin_lock()
235 static inline void __bpf_spin_unlock(struct bpf_spin_lock *lock) in __bpf_spin_unlock() argument
237 arch_spinlock_t *l = (void *)lock; in __bpf_spin_unlock()
244 static inline void __bpf_spin_lock(struct bpf_spin_lock *lock) in __bpf_spin_lock() argument
246 atomic_t *l = (void *)lock; in __bpf_spin_lock()
248 BUILD_BUG_ON(sizeof(*l) != sizeof(*lock)); in __bpf_spin_lock()
254 static inline void __bpf_spin_unlock(struct bpf_spin_lock *lock) in __bpf_spin_unlock() argument
256 atomic_t *l = (void *)lock; in __bpf_spin_unlock()
265 notrace BPF_CALL_1(bpf_spin_lock, struct bpf_spin_lock *, lock) in BPF_CALL_1() argument
270 __bpf_spin_lock(lock); in BPF_CALL_1()
282 notrace BPF_CALL_1(bpf_spin_unlock, struct bpf_spin_lock *, lock) in BPF_CALL_1() argument
287 __bpf_spin_unlock(lock); in BPF_CALL_1()
302 struct bpf_spin_lock *lock; in copy_map_value_locked() local
305 lock = src + map->spin_lock_off; in copy_map_value_locked()
307 lock = dst + map->spin_lock_off; in copy_map_value_locked()
309 ____bpf_spin_lock(lock); in copy_map_value_locked()
311 ____bpf_spin_unlock(lock); in copy_map_value_locked()