Home
last modified time | relevance | path

Searched refs:lock (Results 1 – 17 of 17) sorted by relevance

/lib/
Datomic64.c32 raw_spinlock_t lock; member
36 .lock = __RAW_SPIN_LOCK_UNLOCKED(atomic64_lock.lock),
46 return &atomic64_lock[addr & (NR_LOCKS - 1)].lock; in lock_addr()
52 raw_spinlock_t *lock = lock_addr(v); in atomic64_read() local
55 raw_spin_lock_irqsave(lock, flags); in atomic64_read()
57 raw_spin_unlock_irqrestore(lock, flags); in atomic64_read()
65 raw_spinlock_t *lock = lock_addr(v); in atomic64_set() local
67 raw_spin_lock_irqsave(lock, flags); in atomic64_set()
69 raw_spin_unlock_irqrestore(lock, flags); in atomic64_set()
77 raw_spinlock_t *lock = lock_addr(v); \
[all …]
Dlockref.c14 while (likely(arch_spin_value_unlocked(old.lock.rlock.raw_lock))) { \
48 spin_lock(&lockref->lock); in lockref_get()
50 spin_unlock(&lockref->lock); in lockref_get()
71 spin_lock(&lockref->lock); in lockref_get_not_zero()
77 spin_unlock(&lockref->lock); in lockref_get_not_zero()
98 spin_lock(&lockref->lock); in lockref_get_or_lock()
102 spin_unlock(&lockref->lock); in lockref_get_or_lock()
142 spin_lock(&lockref->lock); in lockref_put_or_lock()
146 spin_unlock(&lockref->lock); in lockref_put_or_lock()
157 assert_spin_locked(&lockref->lock); in lockref_mark_dead()
[all …]
Dpercpu_ida.c35 spinlock_t lock; member
84 spin_lock(&remote->lock); in steal_tags()
95 spin_unlock(&remote->lock); in steal_tags()
118 spin_lock(&tags->lock); in alloc_local_tag()
121 spin_unlock(&tags->lock); in alloc_local_tag()
162 spin_lock(&pool->lock); in percpu_ida_alloc()
186 spin_unlock(&pool->lock); in percpu_ida_alloc()
227 spin_lock(&tags->lock); in percpu_ida_free()
231 spin_unlock(&tags->lock); in percpu_ida_free()
240 spin_lock(&pool->lock); in percpu_ida_free()
[all …]
Dpercpu_counter.c65 raw_spin_lock_irqsave(&fbc->lock, flags); in percpu_counter_set()
71 raw_spin_unlock_irqrestore(&fbc->lock, flags); in percpu_counter_set()
83 raw_spin_lock_irqsave(&fbc->lock, flags); in __percpu_counter_add()
86 raw_spin_unlock_irqrestore(&fbc->lock, flags); in __percpu_counter_add()
104 raw_spin_lock_irqsave(&fbc->lock, flags); in __percpu_counter_sum()
110 raw_spin_unlock_irqrestore(&fbc->lock, flags); in __percpu_counter_sum()
120 raw_spin_lock_init(&fbc->lock); in __percpu_counter_init()
121 lockdep_set_class(&fbc->lock, key); in __percpu_counter_init()
185 raw_spin_lock_irqsave(&fbc->lock, flags); in percpu_counter_hotcpu_callback()
189 raw_spin_unlock_irqrestore(&fbc->lock, flags); in percpu_counter_hotcpu_callback()
Ddebugobjects.c33 raw_spinlock_t lock; member
227 raw_spin_lock_irqsave(&db->lock, flags); in debug_objects_oom()
229 raw_spin_unlock_irqrestore(&db->lock, flags); in debug_objects_oom()
319 raw_spin_lock_irqsave(&db->lock, flags); in __debug_object_init()
326 raw_spin_unlock_irqrestore(&db->lock, flags); in __debug_object_init()
343 raw_spin_unlock_irqrestore(&db->lock, flags); in __debug_object_init()
354 raw_spin_unlock_irqrestore(&db->lock, flags); in __debug_object_init()
406 raw_spin_lock_irqsave(&db->lock, flags); in debug_object_activate()
420 raw_spin_unlock_irqrestore(&db->lock, flags); in debug_object_activate()
432 raw_spin_unlock_irqrestore(&db->lock, flags); in debug_object_activate()
[all …]
Ddec_and_lock.c20 int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock) in _atomic_dec_and_lock() argument
27 spin_lock(lock); in _atomic_dec_and_lock()
30 spin_unlock(lock); in _atomic_dec_and_lock()
Diommu-common.c81 spin_lock_init(&(iommu->pools[i].lock)); in iommu_tbl_pool_init()
90 spin_lock_init(&(p->lock)); in iommu_tbl_pool_init()
133 spin_lock_irqsave(&pool->lock, flags); in iommu_tbl_range_alloc()
160 spin_unlock(&(pool->lock)); in iommu_tbl_range_alloc()
162 spin_lock(&(pool->lock)); in iommu_tbl_range_alloc()
195 spin_unlock(&(pool->lock)); in iommu_tbl_range_alloc()
198 spin_lock(&(pool->lock)); in iommu_tbl_range_alloc()
222 spin_unlock_irqrestore(&(pool->lock), flags); in iommu_tbl_range_alloc()
262 spin_lock_irqsave(&(pool->lock), flags); in iommu_tbl_range_free()
264 spin_unlock_irqrestore(&(pool->lock), flags); in iommu_tbl_range_free()
Dflex_proportions.c97 raw_spin_lock_init(&pl->lock); in fprop_local_init_single()
114 raw_spin_lock_irqsave(&pl->lock, flags); in fprop_reflect_period_single()
117 raw_spin_unlock_irqrestore(&pl->lock, flags); in fprop_reflect_period_single()
126 raw_spin_unlock_irqrestore(&pl->lock, flags); in fprop_reflect_period_single()
179 raw_spin_lock_init(&pl->lock); in fprop_local_init_percpu()
197 raw_spin_lock_irqsave(&pl->lock, flags); in fprop_reflect_period_percpu()
200 raw_spin_unlock_irqrestore(&pl->lock, flags); in fprop_reflect_period_percpu()
215 raw_spin_unlock_irqrestore(&pl->lock, flags); in fprop_reflect_period_percpu()
Drhashtable.c53 spinlock_t *lock = rht_bucket_lock(tbl, hash); in lockdep_rht_bucket_is_held() local
55 return (debug_locks) ? lockdep_is_held(lock) : 1; in lockdep_rht_bucket_is_held()
261 spin_lock(&ht->lock); in rhashtable_rehash_table()
264 spin_unlock(&ht->lock); in rhashtable_rehash_table()
527 spin_lock(&ht->lock); in rhashtable_walk_init()
529 rcu_dereference_protected(ht->tbl, lockdep_is_held(&ht->lock)); in rhashtable_walk_init()
531 spin_unlock(&ht->lock); in rhashtable_walk_init()
545 spin_lock(&iter->ht->lock); in rhashtable_walk_exit()
548 spin_unlock(&iter->ht->lock); in rhashtable_walk_exit()
574 spin_lock(&ht->lock); in rhashtable_walk_start()
[all …]
Dratelimit.c42 if (!raw_spin_trylock_irqsave(&rs->lock, flags)) in ___ratelimit()
63 raw_spin_unlock_irqrestore(&rs->lock, flags); in ___ratelimit()
Dproportions.c193 raw_spin_lock_init(&pl->lock); in prop_local_init_percpu()
229 raw_spin_lock_irqsave(&pl->lock, flags); in prop_norm_percpu()
250 raw_spin_unlock_irqrestore(&pl->lock, flags); in prop_norm_percpu()
327 raw_spin_lock_init(&pl->lock); in prop_local_init_single()
359 raw_spin_lock_irqsave(&pl->lock, flags); in prop_norm_single()
370 raw_spin_unlock_irqrestore(&pl->lock, flags); in prop_norm_single()
Didr.c71 spin_lock_irqsave(&idp->lock, flags); in get_from_free_list()
77 spin_unlock_irqrestore(&idp->lock, flags); in get_from_free_list()
167 spin_lock_irqsave(&idp->lock, flags); in move_to_free_list()
169 spin_unlock_irqrestore(&idp->lock, flags); in move_to_free_list()
327 spin_lock_irqsave(&idp->lock, flags); in idr_get_empty_slot()
335 spin_unlock_irqrestore(&idp->lock, flags); in idr_get_empty_slot()
841 spin_lock_init(&idp->lock); in idr_init()
873 spin_lock_irqsave(&ida->idr.lock, flags); in free_bitmap()
878 spin_unlock_irqrestore(&ida->idr.lock, flags); in free_bitmap()
957 spin_lock_irqsave(&ida->idr.lock, flags); in ida_get_new_above()
[all …]
Ddma-debug.c92 spinlock_t lock; member
260 spin_lock_irqsave(&dma_entry_hash[idx].lock, __flags); in get_hash_bucket()
273 spin_unlock_irqrestore(&bucket->lock, __flags); in put_hash_bucket()
415 spin_lock_irqsave(&bucket->lock, flags); in debug_dma_dump_mappings()
429 spin_unlock_irqrestore(&bucket->lock, flags); in debug_dma_dump_mappings()
939 spin_lock(&dma_entry_hash[i].lock); in device_dma_allocations()
946 spin_unlock(&dma_entry_hash[i].lock); in device_dma_allocations()
1017 spin_lock_init(&dma_entry_hash[i].lock); in dma_debug_init()
Dgenalloc.c160 spin_lock_init(&pool->lock); in gen_pool_create()
201 spin_lock(&pool->lock); in gen_pool_add_virt()
203 spin_unlock(&pool->lock); in gen_pool_add_virt()
DKconfig.debug972 bool "Spinlock and rw-lock debugging: basic checks"
1012 This feature will check whether any held lock (spinlock, rwlock,
1015 vfree(), etc.), whether a live lock is incorrectly reinitialized via
1016 spin_lock_init()/mutex_init()/etc., or whether there is any lock
1079 This feature enables tracking lock contention points
1083 This also enables lock events required by "perf lock",
1085 If you want to use "perf lock", you also need to turn on
1088 CONFIG_LOCK_STAT defines "contended" and "acquired" lock events.
1095 If you say Y here, the lock dependency engine will do
1116 lock debugging then those bugs wont be detected of course.)
[all …]
Dlocking-selftest.c117 init_class_##class(raw_spinlock_t *lock, rwlock_t *rwlock, \
120 raw_spin_lock_init(lock); \
/lib/raid6/
Daltivec.uc22 * bracked this with preempt_disable/enable or in a lock)