/mm/ |
D | swap.c | 54 local_lock_t lock; member 58 .lock = INIT_LOCAL_LOCK(lock), 66 local_lock_t lock; member 76 .lock = INIT_LOCAL_LOCK(lock), 248 local_lock_irqsave(&lru_rotate.lock, flags); in rotate_reclaimable_page() 252 local_unlock_irqrestore(&lru_rotate.lock, flags); in rotate_reclaimable_page() 338 local_lock(&lru_pvecs.lock); in activate_page() 343 local_unlock(&lru_pvecs.lock); in activate_page() 371 local_lock(&lru_pvecs.lock); in __lru_cache_activate_page() 393 local_unlock(&lru_pvecs.lock); in __lru_cache_activate_page() [all …]
|
D | z3fold.c | 94 rwlock_t lock; member 159 spinlock_t lock; member 224 rwlock_init(&slots->lock); in alloc_slots() 270 read_lock(&slots->lock); in get_z3fold_header() 274 read_unlock(&slots->lock); in get_z3fold_header() 307 write_lock(&slots->lock); in free_handle() 311 write_unlock(&slots->lock); in free_handle() 325 write_unlock(&slots->lock); in free_handle() 457 write_lock(&slots->lock); in __encode_handle() 459 write_unlock(&slots->lock); in __encode_handle() [all …]
|
D | mmu_notifier.c | 42 spinlock_t lock; member 89 lockdep_assert_held(&subscriptions->lock); in mn_itree_is_invalidating() 101 spin_lock(&subscriptions->lock); in mn_itree_inv_start_range() 112 spin_unlock(&subscriptions->lock); in mn_itree_inv_start_range() 134 spin_lock(&subscriptions->lock); in mn_itree_inv_end() 137 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 161 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 232 spin_lock(&subscriptions->lock); in mmu_interval_read_begin() 236 spin_unlock(&subscriptions->lock); in mmu_interval_read_begin() 321 spin_lock(&subscriptions->lock); in mn_hlist_release() [all …]
|
D | list_lru.c | 54 lockdep_is_held(&nlru->lock)); in list_lru_from_memcg_idx() 122 spin_lock(&nlru->lock); in list_lru_add() 131 spin_unlock(&nlru->lock); in list_lru_add() 134 spin_unlock(&nlru->lock); in list_lru_add() 145 spin_lock(&nlru->lock); in list_lru_del() 151 spin_unlock(&nlru->lock); in list_lru_del() 154 spin_unlock(&nlru->lock); in list_lru_del() 222 ret = isolate(item, l, &nlru->lock, cb_arg); in __list_lru_walk_one() 225 assert_spin_locked(&nlru->lock); in __list_lru_walk_one() 248 assert_spin_locked(&nlru->lock); in __list_lru_walk_one() [all …]
|
D | mempool.c | 184 spin_lock_init(&pool->lock); in mempool_init_node() 308 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 312 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 314 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 319 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 327 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 330 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 341 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 345 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 349 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() [all …]
|
D | zbud.c | 102 spinlock_t lock; member 228 spin_lock_init(&pool->lock); in zbud_create_pool() 281 spin_lock(&pool->lock); in zbud_alloc() 298 spin_unlock(&pool->lock); in zbud_alloc() 302 spin_lock(&pool->lock); in zbud_alloc() 328 spin_unlock(&pool->lock); in zbud_alloc() 348 spin_lock(&pool->lock); in zbud_free() 359 spin_unlock(&pool->lock); in zbud_free() 377 spin_unlock(&pool->lock); in zbud_free() 421 spin_lock(&pool->lock); in zbud_reclaim_page() [all …]
|
D | kmemleak.c | 140 raw_spinlock_t lock; member 596 raw_spin_lock_init(&object->lock); in create_object() 679 raw_spin_lock_irqsave(&object->lock, flags); in __delete_object() 681 raw_spin_unlock_irqrestore(&object->lock, flags); in __delete_object() 751 raw_spin_lock_irqsave(&object->lock, flags); in paint_it() 753 raw_spin_unlock_irqrestore(&object->lock, flags); in paint_it() 815 raw_spin_lock_irqsave(&object->lock, flags); in add_scan_area() 837 raw_spin_unlock_irqrestore(&object->lock, flags); in add_scan_area() 859 raw_spin_lock_irqsave(&object->lock, flags); in object_set_excess_ref() 861 raw_spin_unlock_irqrestore(&object->lock, flags); in object_set_excess_ref() [all …]
|
D | dmapool.c | 45 spinlock_t lock; member 85 spin_lock_irq(&pool->lock); in pools_show() 90 spin_unlock_irq(&pool->lock); in pools_show() 164 spin_lock_init(&retval->lock); in dma_pool_create() 324 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 331 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 337 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 373 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 410 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free() 413 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free() [all …]
|
D | page_isolation.c | 25 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 33 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 52 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 56 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 78 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 123 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate() 311 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated() 313 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
|
D | zsmalloc.c | 201 spinlock_t lock; member 291 rwlock_t lock; member 609 spin_lock(&class->lock); in zs_stats_size_show() 615 spin_unlock(&class->lock); in zs_stats_size_show() 937 assert_spin_locked(&class->lock); in __free_zspage() 1427 spin_lock(&class->lock); in zs_malloc() 1434 spin_unlock(&class->lock); in zs_malloc() 1439 spin_unlock(&class->lock); in zs_malloc() 1447 spin_lock(&class->lock); in zs_malloc() 1459 spin_unlock(&class->lock); in zs_malloc() [all …]
|
D | swapfile.c | 357 spin_lock(&ci->lock); in lock_cluster() 365 spin_unlock(&ci->lock); in unlock_cluster() 381 spin_lock(&si->lock); in lock_cluster_or_swap_info() 392 spin_unlock(&si->lock); in unlock_cluster_or_swap_info() 427 spin_lock_nested(&ci_tail->lock, SINGLE_DEPTH_NESTING); in cluster_list_add_tail() 429 spin_unlock(&ci_tail->lock); in cluster_list_add_tail() 489 spin_unlock(&si->lock); in swap_do_scheduled_discard() 494 spin_lock(&si->lock); in swap_do_scheduled_discard() 509 spin_lock(&si->lock); in swap_discard_work() 511 spin_unlock(&si->lock); in swap_discard_work() [all …]
|
D | zswap.c | 194 spinlock_t lock; member 958 spin_lock(&tree->lock); in zswap_writeback_entry() 962 spin_unlock(&tree->lock); in zswap_writeback_entry() 967 spin_unlock(&tree->lock); in zswap_writeback_entry() 1018 spin_lock(&tree->lock); in zswap_writeback_entry() 1031 spin_unlock(&tree->lock); in zswap_writeback_entry() 1043 spin_lock(&tree->lock); in zswap_writeback_entry() 1045 spin_unlock(&tree->lock); in zswap_writeback_entry() 1216 spin_lock(&tree->lock); in zswap_frontswap_store() 1226 spin_unlock(&tree->lock); in zswap_frontswap_store() [all …]
|
D | swap_cgroup.c | 12 spinlock_t lock; member 107 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 113 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 138 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_record() 151 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_record() 185 spin_lock_init(&ctrl->lock); in swap_cgroup_swapon()
|
D | page_pinner.c | 49 spinlock_t lock; member 98 spin_lock_init(&pp_buffer.lock); in init_page_pinner() 146 spin_lock_irqsave(&pp_buf->lock, flags); in add_record() 150 spin_unlock_irqrestore(&pp_buf->lock, flags); in add_record() 347 spin_lock_irqsave(&pp_buffer.lock, flags); in read_buffer() 349 spin_unlock_irqrestore(&pp_buffer.lock, flags); in read_buffer() 390 spin_lock_irqsave(&pp_buffer.lock, flags); in buffer_size_set() 395 spin_unlock_irqrestore(&pp_buffer.lock, flags); in buffer_size_set()
|
D | mmap_lock.c | 44 local_lock_t lock; member 49 .lock = INIT_LOCAL_LOCK(lock), 159 local_lock(&memcg_paths.lock); \ 166 local_unlock(&memcg_paths.lock); \
|
D | cma.c | 99 spin_lock_irqsave(&cma->lock, flags); in cma_clear_bitmap() 101 spin_unlock_irqrestore(&cma->lock, flags); in cma_clear_bitmap() 130 spin_lock_init(&cma->lock); in cma_activate_area() 404 spin_lock_irq(&cma->lock); in cma_debug_show_areas() 419 spin_unlock_irq(&cma->lock); in cma_debug_show_areas() 480 spin_lock_irq(&cma->lock); in cma_alloc() 486 spin_unlock_irq(&cma->lock); in cma_alloc() 507 spin_unlock_irq(&cma->lock); in cma_alloc() 517 spin_unlock_irq(&cma->lock); in cma_alloc()
|
D | page_reporting.c | 133 spin_lock_irq(&zone->lock); in page_reporting_cycle() 195 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 207 spin_lock_irq(&zone->lock); in page_reporting_cycle() 227 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 272 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 274 spin_unlock_irq(&zone->lock); in page_reporting_process_zone()
|
D | highmem.c | 710 spinlock_t lock; /* Protect this bucket's list */ member 735 spin_lock_irqsave(&pas->lock, flags); in page_address() 747 spin_unlock_irqrestore(&pas->lock, flags); in page_address() 771 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 773 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 775 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 779 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 783 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 795 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
|
D | shuffle.c | 89 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 142 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone() 144 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 147 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
|
D | cma_debug.c | 39 spin_lock_irq(&cma->lock); in cma_used_get() 42 spin_unlock_irq(&cma->lock); in cma_used_get() 56 spin_lock_irq(&cma->lock); in cma_maxchunk_get() 64 spin_unlock_irq(&cma->lock); in cma_maxchunk_get()
|
/mm/kfence/ |
D | core.c | 257 lockdep_assert_held(&meta->lock); in metadata_update_state() 307 lockdep_assert_held(&meta->lock); in for_each_canary() 352 if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) { in kfence_guarded_alloc() 408 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc() 438 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free() 445 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 478 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 566 raw_spin_lock_init(&meta->lock); in kfence_init_pool() 654 raw_spin_lock_irqsave(&meta->lock, flags); in show_object() 656 raw_spin_unlock_irqrestore(&meta->lock, flags); in show_object() [all …]
|
D | kfence_test.c | 37 spinlock_t lock; member 41 .lock = __SPIN_LOCK_UNLOCKED(observed.lock), 50 spin_lock_irqsave(&observed.lock, flags); in probe_console() 67 spin_unlock_irqrestore(&observed.lock, flags); in probe_console() 162 spin_lock_irqsave(&observed.lock, flags); in report_matches() 169 spin_unlock_irqrestore(&observed.lock, flags); in report_matches() 806 spin_lock_irqsave(&observed.lock, flags); in test_init() 810 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
|
D | report.c | 130 lockdep_assert_held(&meta->lock); in kfence_print_object() 200 lockdep_assert_held(&meta->lock); in kfence_report_error() 306 raw_spin_lock_irqsave(&meta->lock, flags); in __kfence_obj_info() 317 raw_spin_unlock_irqrestore(&meta->lock, flags); in __kfence_obj_info()
|
/mm/kasan/ |
D | quarantine.c | 104 raw_spinlock_t lock; member 109 .lock = __RAW_SPIN_LOCK_UNLOCKED(shrink_qlist.lock), 339 raw_spin_lock_irqsave(&sq->lock, flags); in __per_cpu_remove_cache() 341 raw_spin_unlock_irqrestore(&sq->lock, flags); in __per_cpu_remove_cache() 382 raw_spin_lock_irqsave(&sq->lock, flags); in kasan_quarantine_remove_cache() 384 raw_spin_unlock_irqrestore(&sq->lock, flags); in kasan_quarantine_remove_cache()
|
D | tags.c | 40 .lock = __RW_LOCK_UNLOCKED(stack_ring.lock) 110 read_lock_irqsave(&stack_ring.lock, flags); in save_stack_info() 133 read_unlock_irqrestore(&stack_ring.lock, flags); in save_stack_info()
|