/mm/ |
D | swap.c | 55 local_lock_t lock; member 59 .lock = INIT_LOCAL_LOCK(lock), 67 local_lock_t lock; member 77 .lock = INIT_LOCAL_LOCK(lock), 265 local_lock_irqsave(&lru_rotate.lock, flags); in folio_rotate_reclaimable() 268 local_unlock_irqrestore(&lru_rotate.lock, flags); in folio_rotate_reclaimable() 362 local_lock(&cpu_fbatches.lock); in folio_activate() 365 local_unlock(&cpu_fbatches.lock); in folio_activate() 392 local_lock(&cpu_fbatches.lock); in __lru_cache_activate_folio() 414 local_unlock(&cpu_fbatches.lock); in __lru_cache_activate_folio() [all …]
|
D | mmu_notifier.c | 42 spinlock_t lock; member 89 lockdep_assert_held(&subscriptions->lock); in mn_itree_is_invalidating() 101 spin_lock(&subscriptions->lock); in mn_itree_inv_start_range() 112 spin_unlock(&subscriptions->lock); in mn_itree_inv_start_range() 134 spin_lock(&subscriptions->lock); in mn_itree_inv_end() 137 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 161 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 232 spin_lock(&subscriptions->lock); in mmu_interval_read_begin() 236 spin_unlock(&subscriptions->lock); in mmu_interval_read_begin() 321 spin_lock(&subscriptions->lock); in mn_hlist_release() [all …]
|
D | mempool.c | 189 spin_lock_init(&pool->lock); in mempool_init_node() 313 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 317 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 319 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 324 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 332 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 335 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 346 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 350 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 354 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() [all …]
|
D | list_lru.c | 126 spin_lock(&nlru->lock); in list_lru_add() 135 spin_unlock(&nlru->lock); in list_lru_add() 138 spin_unlock(&nlru->lock); in list_lru_add() 149 spin_lock(&nlru->lock); in list_lru_del() 155 spin_unlock(&nlru->lock); in list_lru_del() 158 spin_unlock(&nlru->lock); in list_lru_del() 231 ret = isolate(item, l, &nlru->lock, cb_arg); in __list_lru_walk_one() 234 assert_spin_locked(&nlru->lock); in __list_lru_walk_one() 257 assert_spin_locked(&nlru->lock); in __list_lru_walk_one() 275 spin_lock(&nlru->lock); in list_lru_walk_one() [all …]
|
D | z3fold.c | 87 rwlock_t lock; member 145 spinlock_t lock; member 204 rwlock_init(&slots->lock); in alloc_slots() 250 read_lock(&slots->lock); in get_z3fold_header() 254 read_unlock(&slots->lock); in get_z3fold_header() 289 write_lock(&slots->lock); in free_handle() 293 write_unlock(&slots->lock); in free_handle() 307 write_unlock(&slots->lock); in free_handle() 390 write_lock(&slots->lock); in __encode_handle() 392 write_unlock(&slots->lock); in __encode_handle() [all …]
|
D | kmemleak.c | 144 raw_spinlock_t lock; member 650 raw_spin_lock_init(&object->lock); in __create_object() 753 raw_spin_lock_irqsave(&object->lock, flags); in __delete_object() 755 raw_spin_unlock_irqrestore(&object->lock, flags); in __delete_object() 825 raw_spin_lock_irqsave(&object->lock, flags); in paint_it() 827 raw_spin_unlock_irqrestore(&object->lock, flags); in paint_it() 889 raw_spin_lock_irqsave(&object->lock, flags); in add_scan_area() 911 raw_spin_unlock_irqrestore(&object->lock, flags); in add_scan_area() 933 raw_spin_lock_irqsave(&object->lock, flags); in object_set_excess_ref() 935 raw_spin_unlock_irqrestore(&object->lock, flags); in object_set_excess_ref() [all …]
|
D | zsmalloc.c | 241 spinlock_t lock; member 258 rwlock_t lock; member 262 local_lock_t lock; member 432 .lock = INIT_LOCAL_LOCK(lock), 582 spin_lock(&pool->lock); in zs_stats_size_show() 593 spin_unlock(&pool->lock); in zs_stats_size_show() 858 assert_spin_locked(&pool->lock); in __free_zspage() 1214 spin_lock(&pool->lock); in zs_map_object() 1226 spin_unlock(&pool->lock); in zs_map_object() 1231 local_lock(&zs_map_area.lock); in zs_map_object() [all …]
|
D | swapfile.c | 314 spin_lock(&ci->lock); in lock_cluster() 322 spin_unlock(&ci->lock); in unlock_cluster() 338 spin_lock(&si->lock); in lock_cluster_or_swap_info() 349 spin_unlock(&si->lock); in unlock_cluster_or_swap_info() 396 spin_unlock(&si->lock); in swap_do_scheduled_discard() 401 spin_lock(&si->lock); in swap_do_scheduled_discard() 403 spin_lock(&ci->lock); in swap_do_scheduled_discard() 407 spin_unlock(&ci->lock); in swap_do_scheduled_discard() 417 spin_lock(&si->lock); in swap_discard_work() 419 spin_unlock(&si->lock); in swap_discard_work() [all …]
|
D | zbud.c | 92 spinlock_t lock; member 208 spin_lock_init(&pool->lock); in zbud_create_pool() 259 spin_lock(&pool->lock); in zbud_alloc() 276 spin_unlock(&pool->lock); in zbud_alloc() 280 spin_lock(&pool->lock); in zbud_alloc() 301 spin_unlock(&pool->lock); in zbud_alloc() 316 spin_lock(&pool->lock); in zbud_free() 338 spin_unlock(&pool->lock); in zbud_free()
|
D | zswap.c | 228 spinlock_t lock; member 663 spin_lock(&tree->lock); in zswap_reclaim_entry() 670 spin_unlock(&tree->lock); in zswap_reclaim_entry() 674 spin_lock(&tree->lock); in zswap_reclaim_entry() 694 spin_unlock(&tree->lock); in zswap_reclaim_entry() 1105 spin_lock(&tree->lock); in zswap_writeback_entry() 1107 spin_unlock(&tree->lock); in zswap_writeback_entry() 1114 spin_unlock(&tree->lock); in zswap_writeback_entry() 1234 spin_lock(&tree->lock); in zswap_store() 1240 spin_unlock(&tree->lock); in zswap_store() [all …]
|
D | swap_cgroup.c | 12 spinlock_t lock; member 107 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 113 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 138 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_record() 151 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_record() 186 spin_lock_init(&ctrl->lock); in swap_cgroup_swapon()
|
D | page_pinner.c | 49 spinlock_t lock; member 98 spin_lock_init(&pp_buffer.lock); in init_page_pinner() 146 spin_lock_irqsave(&pp_buf->lock, flags); in add_record() 150 spin_unlock_irqrestore(&pp_buf->lock, flags); in add_record() 347 spin_lock_irqsave(&pp_buffer.lock, flags); in read_buffer() 349 spin_unlock_irqrestore(&pp_buffer.lock, flags); in read_buffer() 390 spin_lock_irqsave(&pp_buffer.lock, flags); in buffer_size_set() 395 spin_unlock_irqrestore(&pp_buffer.lock, flags); in buffer_size_set()
|
D | page_isolation.c | 156 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 164 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 191 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 195 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 216 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 264 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate() 681 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated() 683 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
|
D | cma.c | 96 spin_lock_irqsave(&cma->lock, flags); in cma_clear_bitmap() 98 spin_unlock_irqrestore(&cma->lock, flags); in cma_clear_bitmap() 127 spin_lock_init(&cma->lock); in cma_activate_area() 408 spin_lock_irq(&cma->lock); in cma_debug_show_areas() 423 spin_unlock_irq(&cma->lock); in cma_debug_show_areas() 475 spin_lock_irq(&cma->lock); in __cma_alloc() 481 spin_unlock_irq(&cma->lock); in __cma_alloc() 500 spin_unlock_irq(&cma->lock); in __cma_alloc() 510 spin_unlock_irq(&cma->lock); in __cma_alloc()
|
D | dmapool.c | 50 spinlock_t lock; member 263 spin_lock_init(&retval->lock); in dma_pool_create() 413 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 420 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 426 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 430 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 455 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free() 460 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free()
|
D | page_reporting.c | 165 spin_lock_irq(&zone->lock); in page_reporting_cycle() 227 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 239 spin_lock_irq(&zone->lock); in page_reporting_cycle() 259 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 304 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 306 spin_unlock_irq(&zone->lock); in page_reporting_process_zone()
|
D | mlock.c | 34 local_lock_t lock; member 39 .lock = INIT_LOCAL_LOCK(lock), 219 local_lock(&mlock_fbatch.lock); in mlock_drain_local() 223 local_unlock(&mlock_fbatch.lock); in mlock_drain_local() 249 local_lock(&mlock_fbatch.lock); in mlock_folio() 263 local_unlock(&mlock_fbatch.lock); in mlock_folio() 275 local_lock(&mlock_fbatch.lock); in mlock_new_folio() 286 local_unlock(&mlock_fbatch.lock); in mlock_new_folio() 297 local_lock(&mlock_fbatch.lock); in munlock_folio() 307 local_unlock(&mlock_fbatch.lock); in munlock_folio()
|
D | shuffle.c | 88 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 141 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone() 143 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 146 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
|
D | highmem.c | 728 spinlock_t lock; /* Protect this bucket's list */ member 753 spin_lock_irqsave(&pas->lock, flags); in page_address() 765 spin_unlock_irqrestore(&pas->lock, flags); in page_address() 789 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 791 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 793 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 800 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 812 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
|
D | page_alloc.c | 170 pcpu_spin_lock(struct per_cpu_pages, lock, ptr) 173 pcpu_spin_trylock(struct per_cpu_pages, lock, ptr) 176 pcpu_spin_unlock(lock, ptr) 970 spin_lock_irqsave(&zone->lock, flags); in split_free_page() 998 spin_unlock_irqrestore(&zone->lock, flags); in split_free_page() 1306 spin_lock_irqsave(&zone->lock, flags); in free_pcppages_bulk() 1344 spin_unlock_irqrestore(&zone->lock, flags); in free_pcppages_bulk() 1354 spin_lock_irqsave(&zone->lock, flags); in free_one_page() 1360 spin_unlock_irqrestore(&zone->lock, flags); in free_one_page() 1396 spin_lock_irqsave(&zone->lock, flags); in __free_pages_ok() [all …]
|
/mm/kfence/ |
D | core.c | 283 lockdep_assert_held(&meta->lock); in metadata_update_state() 320 raw_spin_lock_irqsave(&meta->lock, flags); in check_canary_byte() 322 raw_spin_unlock_irqrestore(&meta->lock, flags); in check_canary_byte() 414 if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) { in kfence_guarded_alloc() 459 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc() 500 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free() 507 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 529 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 626 raw_spin_lock_init(&meta->lock); in kfence_init_pool() 740 raw_spin_lock_irqsave(&meta->lock, flags); in show_object() [all …]
|
D | kfence_test.c | 42 spinlock_t lock; member 46 .lock = __SPIN_LOCK_UNLOCKED(observed.lock), 55 spin_lock_irqsave(&observed.lock, flags); in probe_console() 72 spin_unlock_irqrestore(&observed.lock, flags); in probe_console() 167 spin_lock_irqsave(&observed.lock, flags); in report_matches() 174 spin_unlock_irqrestore(&observed.lock, flags); in report_matches() 810 spin_lock_irqsave(&observed.lock, flags); in test_init() 814 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
|
/mm/kasan/ |
D | quarantine.c | 103 raw_spinlock_t lock; member 108 .lock = __RAW_SPIN_LOCK_UNLOCKED(shrink_qlist.lock), 327 raw_spin_lock_irqsave(&sq->lock, flags); in __per_cpu_remove_cache() 329 raw_spin_unlock_irqrestore(&sq->lock, flags); in __per_cpu_remove_cache() 366 raw_spin_lock_irqsave(&sq->lock, flags); in kasan_quarantine_remove_cache() 368 raw_spin_unlock_irqrestore(&sq->lock, flags); in kasan_quarantine_remove_cache()
|
D | tags.c | 40 .lock = __RW_LOCK_UNLOCKED(stack_ring.lock) 110 read_lock_irqsave(&stack_ring.lock, flags); in save_stack_info() 133 read_unlock_irqrestore(&stack_ring.lock, flags); in save_stack_info()
|
/mm/kmsan/ |
D | kmsan_test.c | 32 spinlock_t lock; member 37 .lock = __SPIN_LOCK_UNLOCKED(observed.lock), 47 spin_lock_irqsave(&observed.lock, flags); in probe_console() 61 spin_unlock_irqrestore(&observed.lock, flags); in probe_console() 107 spin_lock_irqsave(&observed.lock, flags); in report_matches() 114 spin_unlock_irqrestore(&observed.lock, flags); in report_matches() 616 spin_lock_irqsave(&observed.lock, flags); in test_init() 620 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
|