Home
last modified time | relevance | path

Searched refs:lock (Results 1 – 25 of 46) sorted by relevance

12

/mm/
Dswap.c54 local_lock_t lock; member
58 .lock = INIT_LOCAL_LOCK(lock),
66 local_lock_t lock; member
76 .lock = INIT_LOCAL_LOCK(lock),
248 local_lock_irqsave(&lru_rotate.lock, flags); in rotate_reclaimable_page()
252 local_unlock_irqrestore(&lru_rotate.lock, flags); in rotate_reclaimable_page()
338 local_lock(&lru_pvecs.lock); in activate_page()
343 local_unlock(&lru_pvecs.lock); in activate_page()
371 local_lock(&lru_pvecs.lock); in __lru_cache_activate_page()
393 local_unlock(&lru_pvecs.lock); in __lru_cache_activate_page()
[all …]
Dz3fold.c94 rwlock_t lock; member
159 spinlock_t lock; member
224 rwlock_init(&slots->lock); in alloc_slots()
270 read_lock(&slots->lock); in get_z3fold_header()
274 read_unlock(&slots->lock); in get_z3fold_header()
307 write_lock(&slots->lock); in free_handle()
311 write_unlock(&slots->lock); in free_handle()
325 write_unlock(&slots->lock); in free_handle()
457 write_lock(&slots->lock); in __encode_handle()
459 write_unlock(&slots->lock); in __encode_handle()
[all …]
Dmmu_notifier.c42 spinlock_t lock; member
89 lockdep_assert_held(&subscriptions->lock); in mn_itree_is_invalidating()
101 spin_lock(&subscriptions->lock); in mn_itree_inv_start_range()
112 spin_unlock(&subscriptions->lock); in mn_itree_inv_start_range()
134 spin_lock(&subscriptions->lock); in mn_itree_inv_end()
137 spin_unlock(&subscriptions->lock); in mn_itree_inv_end()
161 spin_unlock(&subscriptions->lock); in mn_itree_inv_end()
232 spin_lock(&subscriptions->lock); in mmu_interval_read_begin()
236 spin_unlock(&subscriptions->lock); in mmu_interval_read_begin()
321 spin_lock(&subscriptions->lock); in mn_hlist_release()
[all …]
Dlist_lru.c54 lockdep_is_held(&nlru->lock)); in list_lru_from_memcg_idx()
122 spin_lock(&nlru->lock); in list_lru_add()
131 spin_unlock(&nlru->lock); in list_lru_add()
134 spin_unlock(&nlru->lock); in list_lru_add()
145 spin_lock(&nlru->lock); in list_lru_del()
151 spin_unlock(&nlru->lock); in list_lru_del()
154 spin_unlock(&nlru->lock); in list_lru_del()
222 ret = isolate(item, l, &nlru->lock, cb_arg); in __list_lru_walk_one()
225 assert_spin_locked(&nlru->lock); in __list_lru_walk_one()
248 assert_spin_locked(&nlru->lock); in __list_lru_walk_one()
[all …]
Dmempool.c184 spin_lock_init(&pool->lock); in mempool_init_node()
308 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
312 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
314 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
319 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
327 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
330 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
341 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
345 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
349 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
[all …]
Dzbud.c102 spinlock_t lock; member
228 spin_lock_init(&pool->lock); in zbud_create_pool()
281 spin_lock(&pool->lock); in zbud_alloc()
298 spin_unlock(&pool->lock); in zbud_alloc()
302 spin_lock(&pool->lock); in zbud_alloc()
328 spin_unlock(&pool->lock); in zbud_alloc()
348 spin_lock(&pool->lock); in zbud_free()
359 spin_unlock(&pool->lock); in zbud_free()
377 spin_unlock(&pool->lock); in zbud_free()
421 spin_lock(&pool->lock); in zbud_reclaim_page()
[all …]
Dkmemleak.c140 raw_spinlock_t lock; member
596 raw_spin_lock_init(&object->lock); in create_object()
679 raw_spin_lock_irqsave(&object->lock, flags); in __delete_object()
681 raw_spin_unlock_irqrestore(&object->lock, flags); in __delete_object()
751 raw_spin_lock_irqsave(&object->lock, flags); in paint_it()
753 raw_spin_unlock_irqrestore(&object->lock, flags); in paint_it()
815 raw_spin_lock_irqsave(&object->lock, flags); in add_scan_area()
837 raw_spin_unlock_irqrestore(&object->lock, flags); in add_scan_area()
859 raw_spin_lock_irqsave(&object->lock, flags); in object_set_excess_ref()
861 raw_spin_unlock_irqrestore(&object->lock, flags); in object_set_excess_ref()
[all …]
Ddmapool.c45 spinlock_t lock; member
85 spin_lock_irq(&pool->lock); in pools_show()
90 spin_unlock_irq(&pool->lock); in pools_show()
164 spin_lock_init(&retval->lock); in dma_pool_create()
324 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc()
331 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc()
337 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc()
373 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc()
410 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free()
413 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free()
[all …]
Dpage_isolation.c25 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate()
33 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
52 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
56 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
78 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate()
123 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate()
311 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated()
313 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
Dzsmalloc.c201 spinlock_t lock; member
291 rwlock_t lock; member
609 spin_lock(&class->lock); in zs_stats_size_show()
615 spin_unlock(&class->lock); in zs_stats_size_show()
937 assert_spin_locked(&class->lock); in __free_zspage()
1427 spin_lock(&class->lock); in zs_malloc()
1434 spin_unlock(&class->lock); in zs_malloc()
1439 spin_unlock(&class->lock); in zs_malloc()
1447 spin_lock(&class->lock); in zs_malloc()
1459 spin_unlock(&class->lock); in zs_malloc()
[all …]
Dswapfile.c357 spin_lock(&ci->lock); in lock_cluster()
365 spin_unlock(&ci->lock); in unlock_cluster()
381 spin_lock(&si->lock); in lock_cluster_or_swap_info()
392 spin_unlock(&si->lock); in unlock_cluster_or_swap_info()
427 spin_lock_nested(&ci_tail->lock, SINGLE_DEPTH_NESTING); in cluster_list_add_tail()
429 spin_unlock(&ci_tail->lock); in cluster_list_add_tail()
489 spin_unlock(&si->lock); in swap_do_scheduled_discard()
494 spin_lock(&si->lock); in swap_do_scheduled_discard()
509 spin_lock(&si->lock); in swap_discard_work()
511 spin_unlock(&si->lock); in swap_discard_work()
[all …]
Dzswap.c194 spinlock_t lock; member
958 spin_lock(&tree->lock); in zswap_writeback_entry()
962 spin_unlock(&tree->lock); in zswap_writeback_entry()
967 spin_unlock(&tree->lock); in zswap_writeback_entry()
1018 spin_lock(&tree->lock); in zswap_writeback_entry()
1031 spin_unlock(&tree->lock); in zswap_writeback_entry()
1043 spin_lock(&tree->lock); in zswap_writeback_entry()
1045 spin_unlock(&tree->lock); in zswap_writeback_entry()
1216 spin_lock(&tree->lock); in zswap_frontswap_store()
1226 spin_unlock(&tree->lock); in zswap_frontswap_store()
[all …]
Dswap_cgroup.c12 spinlock_t lock; member
107 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_cmpxchg()
113 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_cmpxchg()
138 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_record()
151 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_record()
185 spin_lock_init(&ctrl->lock); in swap_cgroup_swapon()
Dpage_pinner.c49 spinlock_t lock; member
98 spin_lock_init(&pp_buffer.lock); in init_page_pinner()
146 spin_lock_irqsave(&pp_buf->lock, flags); in add_record()
150 spin_unlock_irqrestore(&pp_buf->lock, flags); in add_record()
347 spin_lock_irqsave(&pp_buffer.lock, flags); in read_buffer()
349 spin_unlock_irqrestore(&pp_buffer.lock, flags); in read_buffer()
390 spin_lock_irqsave(&pp_buffer.lock, flags); in buffer_size_set()
395 spin_unlock_irqrestore(&pp_buffer.lock, flags); in buffer_size_set()
Dmmap_lock.c44 local_lock_t lock; member
49 .lock = INIT_LOCAL_LOCK(lock),
159 local_lock(&memcg_paths.lock); \
166 local_unlock(&memcg_paths.lock); \
Dcma.c99 spin_lock_irqsave(&cma->lock, flags); in cma_clear_bitmap()
101 spin_unlock_irqrestore(&cma->lock, flags); in cma_clear_bitmap()
130 spin_lock_init(&cma->lock); in cma_activate_area()
404 spin_lock_irq(&cma->lock); in cma_debug_show_areas()
419 spin_unlock_irq(&cma->lock); in cma_debug_show_areas()
480 spin_lock_irq(&cma->lock); in cma_alloc()
486 spin_unlock_irq(&cma->lock); in cma_alloc()
507 spin_unlock_irq(&cma->lock); in cma_alloc()
517 spin_unlock_irq(&cma->lock); in cma_alloc()
Dpage_reporting.c133 spin_lock_irq(&zone->lock); in page_reporting_cycle()
195 spin_unlock_irq(&zone->lock); in page_reporting_cycle()
207 spin_lock_irq(&zone->lock); in page_reporting_cycle()
227 spin_unlock_irq(&zone->lock); in page_reporting_cycle()
272 spin_lock_irq(&zone->lock); in page_reporting_process_zone()
274 spin_unlock_irq(&zone->lock); in page_reporting_process_zone()
Dhighmem.c710 spinlock_t lock; /* Protect this bucket's list */ member
735 spin_lock_irqsave(&pas->lock, flags); in page_address()
747 spin_unlock_irqrestore(&pas->lock, flags); in page_address()
771 spin_lock_irqsave(&pas->lock, flags); in set_page_address()
773 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address()
775 spin_lock_irqsave(&pas->lock, flags); in set_page_address()
779 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address()
783 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address()
795 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
Dshuffle.c89 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone()
142 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
144 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone()
147 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
Dcma_debug.c39 spin_lock_irq(&cma->lock); in cma_used_get()
42 spin_unlock_irq(&cma->lock); in cma_used_get()
56 spin_lock_irq(&cma->lock); in cma_maxchunk_get()
64 spin_unlock_irq(&cma->lock); in cma_maxchunk_get()
/mm/kfence/
Dcore.c257 lockdep_assert_held(&meta->lock); in metadata_update_state()
307 lockdep_assert_held(&meta->lock); in for_each_canary()
352 if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) { in kfence_guarded_alloc()
408 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc()
438 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free()
445 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free()
478 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free()
566 raw_spin_lock_init(&meta->lock); in kfence_init_pool()
654 raw_spin_lock_irqsave(&meta->lock, flags); in show_object()
656 raw_spin_unlock_irqrestore(&meta->lock, flags); in show_object()
[all …]
Dkfence_test.c37 spinlock_t lock; member
41 .lock = __SPIN_LOCK_UNLOCKED(observed.lock),
50 spin_lock_irqsave(&observed.lock, flags); in probe_console()
67 spin_unlock_irqrestore(&observed.lock, flags); in probe_console()
162 spin_lock_irqsave(&observed.lock, flags); in report_matches()
169 spin_unlock_irqrestore(&observed.lock, flags); in report_matches()
806 spin_lock_irqsave(&observed.lock, flags); in test_init()
810 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
Dreport.c130 lockdep_assert_held(&meta->lock); in kfence_print_object()
200 lockdep_assert_held(&meta->lock); in kfence_report_error()
306 raw_spin_lock_irqsave(&meta->lock, flags); in __kfence_obj_info()
317 raw_spin_unlock_irqrestore(&meta->lock, flags); in __kfence_obj_info()
/mm/kasan/
Dquarantine.c104 raw_spinlock_t lock; member
109 .lock = __RAW_SPIN_LOCK_UNLOCKED(shrink_qlist.lock),
339 raw_spin_lock_irqsave(&sq->lock, flags); in __per_cpu_remove_cache()
341 raw_spin_unlock_irqrestore(&sq->lock, flags); in __per_cpu_remove_cache()
382 raw_spin_lock_irqsave(&sq->lock, flags); in kasan_quarantine_remove_cache()
384 raw_spin_unlock_irqrestore(&sq->lock, flags); in kasan_quarantine_remove_cache()
Dtags.c40 .lock = __RW_LOCK_UNLOCKED(stack_ring.lock)
110 read_lock_irqsave(&stack_ring.lock, flags); in save_stack_info()
133 read_unlock_irqrestore(&stack_ring.lock, flags); in save_stack_info()

12