/mm/ |
D | z3fold.c | 94 rwlock_t lock; member 157 spinlock_t lock; member 222 rwlock_init(&slots->lock); in alloc_slots() 258 bool lock) in __get_z3fold_header() argument 269 read_lock(&slots->lock); in __get_z3fold_header() 272 if (lock) in __get_z3fold_header() 274 read_unlock(&slots->lock); in __get_z3fold_header() 278 } while (lock); in __get_z3fold_header() 319 write_lock(&slots->lock); in free_handle() 323 write_unlock(&slots->lock); in free_handle() [all …]
|
D | swap.c | 51 local_lock_t lock; member 55 .lock = INIT_LOCAL_LOCK(lock), 63 local_lock_t lock; member 74 .lock = INIT_LOCAL_LOCK(lock), 287 local_lock_irqsave(&lru_rotate.lock, flags); in rotate_reclaimable_page() 291 local_unlock_irqrestore(&lru_rotate.lock, flags); in rotate_reclaimable_page() 371 local_lock(&lru_pvecs.lock); in activate_page() 376 local_unlock(&lru_pvecs.lock); in activate_page() 401 local_lock(&lru_pvecs.lock); in __lru_cache_activate_page() 423 local_unlock(&lru_pvecs.lock); in __lru_cache_activate_page() [all …]
|
D | mmu_notifier.c | 48 spinlock_t lock; member 95 lockdep_assert_held(&subscriptions->lock); in mn_itree_is_invalidating() 107 spin_lock(&subscriptions->lock); in mn_itree_inv_start_range() 118 spin_unlock(&subscriptions->lock); in mn_itree_inv_start_range() 140 spin_lock(&subscriptions->lock); in mn_itree_inv_end() 143 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 167 spin_unlock(&subscriptions->lock); in mn_itree_inv_end() 238 spin_lock(&subscriptions->lock); in mmu_interval_read_begin() 242 spin_unlock(&subscriptions->lock); in mmu_interval_read_begin() 327 spin_lock(&subscriptions->lock); in mn_hlist_release() [all …]
|
D | mempool.c | 184 spin_lock_init(&pool->lock); in mempool_init_node() 308 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 312 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 314 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 319 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 327 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 330 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 341 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 345 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 349 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() [all …]
|
D | list_lru.c | 54 lockdep_is_held(&nlru->lock)); in list_lru_from_memcg_idx() 122 spin_lock(&nlru->lock); in list_lru_add() 131 spin_unlock(&nlru->lock); in list_lru_add() 134 spin_unlock(&nlru->lock); in list_lru_add() 145 spin_lock(&nlru->lock); in list_lru_del() 151 spin_unlock(&nlru->lock); in list_lru_del() 154 spin_unlock(&nlru->lock); in list_lru_del() 222 ret = isolate(item, l, &nlru->lock, cb_arg); in __list_lru_walk_one() 225 assert_spin_locked(&nlru->lock); in __list_lru_walk_one() 248 assert_spin_locked(&nlru->lock); in __list_lru_walk_one() [all …]
|
D | zbud.c | 95 spinlock_t lock; member 314 spin_lock_init(&pool->lock); in zbud_create_pool() 367 spin_lock(&pool->lock); in zbud_alloc() 384 spin_unlock(&pool->lock); in zbud_alloc() 388 spin_lock(&pool->lock); in zbud_alloc() 414 spin_unlock(&pool->lock); in zbud_alloc() 434 spin_lock(&pool->lock); in zbud_free() 445 spin_unlock(&pool->lock); in zbud_free() 463 spin_unlock(&pool->lock); in zbud_free() 507 spin_lock(&pool->lock); in zbud_reclaim_page() [all …]
|
D | kmemleak.c | 139 raw_spinlock_t lock; member 589 raw_spin_lock_init(&object->lock); in create_object() 671 raw_spin_lock_irqsave(&object->lock, flags); in __delete_object() 673 raw_spin_unlock_irqrestore(&object->lock, flags); in __delete_object() 743 raw_spin_lock_irqsave(&object->lock, flags); in paint_it() 745 raw_spin_unlock_irqrestore(&object->lock, flags); in paint_it() 807 raw_spin_lock_irqsave(&object->lock, flags); in add_scan_area() 829 raw_spin_unlock_irqrestore(&object->lock, flags); in add_scan_area() 851 raw_spin_lock_irqsave(&object->lock, flags); in object_set_excess_ref() 853 raw_spin_unlock_irqrestore(&object->lock, flags); in object_set_excess_ref() [all …]
|
D | page_pinner.c | 42 spinlock_t lock; member 48 .lock = __SPIN_LOCK_UNLOCKED(lt_pinner.lock), 55 .lock = __SPIN_LOCK_UNLOCKED(acf_pinner.lock), 153 spin_lock_irqsave(<_pinner.lock, flags); in check_longterm_pin() 157 spin_unlock_irqrestore(<_pinner.lock, flags); in check_longterm_pin() 354 spin_lock_irqsave(&acf_pinner.lock, flags); in __page_pinner_migration_failed() 358 spin_unlock_irqrestore(&acf_pinner.lock, flags); in __page_pinner_migration_failed() 403 spin_lock_irqsave(<_pinner.lock, flags); in read_longterm_page_pinner() 405 spin_unlock_irqrestore(<_pinner.lock, flags); in read_longterm_page_pinner() 439 spin_lock_irqsave(&acf_pinner.lock, flags); in read_alloc_contig_failed() [all …]
|
D | dmapool.c | 44 spinlock_t lock; member 85 spin_lock_irq(&pool->lock); in show_pools() 90 spin_unlock_irq(&pool->lock); in show_pools() 164 spin_lock_init(&retval->lock); in dma_pool_create() 324 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 331 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 337 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 373 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 410 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free() 413 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free() [all …]
|
D | swapfile.c | 360 spin_lock(&ci->lock); in lock_cluster() 368 spin_unlock(&ci->lock); in unlock_cluster() 384 spin_lock(&si->lock); in lock_cluster_or_swap_info() 395 spin_unlock(&si->lock); in unlock_cluster_or_swap_info() 430 spin_lock_nested(&ci_tail->lock, SINGLE_DEPTH_NESTING); in cluster_list_add_tail() 432 spin_unlock(&ci_tail->lock); in cluster_list_add_tail() 492 spin_unlock(&si->lock); in swap_do_scheduled_discard() 497 spin_lock(&si->lock); in swap_do_scheduled_discard() 512 spin_lock(&si->lock); in swap_discard_work() 514 spin_unlock(&si->lock); in swap_discard_work() [all …]
|
D | page_isolation.c | 24 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 32 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 51 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 55 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 77 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 123 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate() 309 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated() 311 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
|
D | highmem.c | 394 spinlock_t lock; /* Protect this bucket's list */ member 419 spin_lock_irqsave(&pas->lock, flags); in page_address() 431 spin_unlock_irqrestore(&pas->lock, flags); in page_address() 456 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 458 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 460 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 464 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 468 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 480 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
|
D | zswap.c | 184 spinlock_t lock; member 894 spin_lock(&tree->lock); in zswap_writeback_entry() 898 spin_unlock(&tree->lock); in zswap_writeback_entry() 902 spin_unlock(&tree->lock); in zswap_writeback_entry() 942 spin_lock(&tree->lock); in zswap_writeback_entry() 955 spin_unlock(&tree->lock); in zswap_writeback_entry() 967 spin_lock(&tree->lock); in zswap_writeback_entry() 969 spin_unlock(&tree->lock); in zswap_writeback_entry() 1115 spin_lock(&tree->lock); in zswap_frontswap_store() 1125 spin_unlock(&tree->lock); in zswap_frontswap_store() [all …]
|
D | zsmalloc.c | 201 spinlock_t lock; member 291 rwlock_t lock; member 609 spin_lock(&class->lock); in zs_stats_size_show() 615 spin_unlock(&class->lock); in zs_stats_size_show() 940 assert_spin_locked(&class->lock); in __free_zspage() 1431 spin_lock(&class->lock); in zs_malloc() 1438 spin_unlock(&class->lock); in zs_malloc() 1443 spin_unlock(&class->lock); in zs_malloc() 1451 spin_lock(&class->lock); in zs_malloc() 1463 spin_unlock(&class->lock); in zs_malloc() [all …]
|
D | swap_cgroup.c | 12 spinlock_t lock; member 107 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 113 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 138 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_record() 151 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_record() 185 spin_lock_init(&ctrl->lock); in swap_cgroup_swapon()
|
D | cma.c | 101 mutex_lock(&cma->lock); in cma_clear_bitmap() 103 mutex_unlock(&cma->lock); in cma_clear_bitmap() 132 mutex_init(&cma->lock); in cma_activate_area() 406 mutex_lock(&cma->lock); in cma_debug_show_areas() 421 mutex_unlock(&cma->lock); in cma_debug_show_areas() 477 mutex_lock(&cma->lock); in cma_alloc() 483 mutex_unlock(&cma->lock); in cma_alloc() 502 mutex_unlock(&cma->lock); in cma_alloc() 512 mutex_unlock(&cma->lock); in cma_alloc()
|
D | page_reporting.c | 127 spin_lock_irq(&zone->lock); in page_reporting_cycle() 185 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 197 spin_lock_irq(&zone->lock); in page_reporting_cycle() 217 spin_unlock_irq(&zone->lock); in page_reporting_cycle() 262 spin_lock_irq(&zone->lock); in page_reporting_process_zone() 264 spin_unlock_irq(&zone->lock); in page_reporting_process_zone()
|
D | shuffle.c | 89 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 142 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone() 144 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone() 147 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
|
D | cma_debug.c | 39 mutex_lock(&cma->lock); in cma_used_get() 42 mutex_unlock(&cma->lock); in cma_used_get() 56 mutex_lock(&cma->lock); in cma_maxchunk_get() 64 mutex_unlock(&cma->lock); in cma_maxchunk_get()
|
D | vmalloc.c | 1491 spinlock_t lock; member 1496 spinlock_t lock; member 1571 spin_lock_init(&vb->lock); in new_vmap_block() 1590 spin_lock(&vbq->lock); in new_vmap_block() 1592 spin_unlock(&vbq->lock); in new_vmap_block() 1622 spin_lock(&vb->lock); in purge_fragmented_blocks() 1628 spin_lock(&vbq->lock); in purge_fragmented_blocks() 1630 spin_unlock(&vbq->lock); in purge_fragmented_blocks() 1631 spin_unlock(&vb->lock); in purge_fragmented_blocks() 1634 spin_unlock(&vb->lock); in purge_fragmented_blocks() [all …]
|
D | shmem.c | 393 spin_lock_irqsave(&info->lock, flags); in shmem_charge() 397 spin_unlock_irqrestore(&info->lock, flags); in shmem_charge() 409 spin_lock_irqsave(&info->lock, flags); in shmem_uncharge() 413 spin_unlock_irqrestore(&info->lock, flags); in shmem_uncharge() 1062 spin_lock_irq(&info->lock); in shmem_undo_range() 1065 spin_unlock_irq(&info->lock); in shmem_undo_range() 1083 spin_lock_irq(&info->lock); in shmem_getattr() 1085 spin_unlock_irq(&info->lock); in shmem_getattr() 1454 spin_lock_irq(&info->lock); in shmem_writepage() 1457 spin_unlock_irq(&info->lock); in shmem_writepage() [all …]
|
D | compaction.c | 504 static bool compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, in compact_lock_irqsave() argument 506 __acquires(lock) in compact_lock_irqsave() 510 if (spin_trylock_irqsave(lock, *flags)) in compact_lock_irqsave() 516 spin_lock_irqsave(lock, *flags); in compact_lock_irqsave() 535 static bool compact_unlock_should_abort(spinlock_t *lock, in compact_unlock_should_abort() argument 539 spin_unlock_irqrestore(lock, flags); in compact_unlock_should_abort() 589 && compact_unlock_should_abort(&cc->zone->lock, flags, in isolate_freepages_block() 624 locked = compact_lock_irqsave(&cc->zone->lock, in isolate_freepages_block() 661 spin_unlock_irqrestore(&cc->zone->lock, flags); in isolate_freepages_block() 1381 spin_lock_irqsave(&cc->zone->lock, flags); in fast_isolate_freepages() [all …]
|
D | page_alloc.c | 1496 spin_lock(&zone->lock); in free_pcppages_bulk() 1514 spin_unlock(&zone->lock); in free_pcppages_bulk() 1522 spin_lock(&zone->lock); in free_one_page() 1528 spin_unlock(&zone->lock); in free_one_page() 2761 spin_lock_irqsave(&zone->lock, flags); in reserve_highatomic_pageblock() 2777 spin_unlock_irqrestore(&zone->lock, flags); in reserve_highatomic_pageblock() 2816 spin_lock_irqsave(&zone->lock, flags); in unreserve_highatomic_pageblock() 2857 spin_unlock_irqrestore(&zone->lock, flags); in unreserve_highatomic_pageblock() 2861 spin_unlock_irqrestore(&zone->lock, flags); in unreserve_highatomic_pageblock() 3005 spin_lock(&zone->lock); in rmqueue_bulk() [all …]
|
/mm/kfence/ |
D | core.c | 191 lockdep_assert_held(&meta->lock); in metadata_update_state() 233 lockdep_assert_held(&meta->lock); in for_each_canary() 274 if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) { in kfence_guarded_alloc() 328 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc() 356 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free() 363 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 396 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free() 475 raw_spin_lock_init(&meta->lock); in kfence_init_pool() 553 raw_spin_lock_irqsave(&meta->lock, flags); in show_object() 555 raw_spin_unlock_irqrestore(&meta->lock, flags); in show_object() [all …]
|
D | kfence_test.c | 30 spinlock_t lock; member 34 .lock = __SPIN_LOCK_UNLOCKED(observed.lock), 43 spin_lock_irqsave(&observed.lock, flags); in probe_console() 60 spin_unlock_irqrestore(&observed.lock, flags); in probe_console() 151 spin_lock_irqsave(&observed.lock, flags); in report_matches() 158 spin_unlock_irqrestore(&observed.lock, flags); in report_matches() 791 spin_lock_irqsave(&observed.lock, flags); in test_init() 795 spin_unlock_irqrestore(&observed.lock, flags); in test_init()
|