Home
last modified time | relevance | path

Searched refs:lock (Results 1 – 25 of 42) sorted by relevance

12

/mm/
Dz3fold.c94 rwlock_t lock; member
157 spinlock_t lock; member
222 rwlock_init(&slots->lock); in alloc_slots()
258 bool lock) in __get_z3fold_header() argument
269 read_lock(&slots->lock); in __get_z3fold_header()
272 if (lock) in __get_z3fold_header()
274 read_unlock(&slots->lock); in __get_z3fold_header()
278 } while (lock); in __get_z3fold_header()
319 write_lock(&slots->lock); in free_handle()
323 write_unlock(&slots->lock); in free_handle()
[all …]
Dswap.c51 local_lock_t lock; member
55 .lock = INIT_LOCAL_LOCK(lock),
63 local_lock_t lock; member
74 .lock = INIT_LOCAL_LOCK(lock),
287 local_lock_irqsave(&lru_rotate.lock, flags); in rotate_reclaimable_page()
291 local_unlock_irqrestore(&lru_rotate.lock, flags); in rotate_reclaimable_page()
371 local_lock(&lru_pvecs.lock); in activate_page()
376 local_unlock(&lru_pvecs.lock); in activate_page()
401 local_lock(&lru_pvecs.lock); in __lru_cache_activate_page()
423 local_unlock(&lru_pvecs.lock); in __lru_cache_activate_page()
[all …]
Dmmu_notifier.c48 spinlock_t lock; member
95 lockdep_assert_held(&subscriptions->lock); in mn_itree_is_invalidating()
107 spin_lock(&subscriptions->lock); in mn_itree_inv_start_range()
118 spin_unlock(&subscriptions->lock); in mn_itree_inv_start_range()
140 spin_lock(&subscriptions->lock); in mn_itree_inv_end()
143 spin_unlock(&subscriptions->lock); in mn_itree_inv_end()
167 spin_unlock(&subscriptions->lock); in mn_itree_inv_end()
238 spin_lock(&subscriptions->lock); in mmu_interval_read_begin()
242 spin_unlock(&subscriptions->lock); in mmu_interval_read_begin()
327 spin_lock(&subscriptions->lock); in mn_hlist_release()
[all …]
Dmempool.c184 spin_lock_init(&pool->lock); in mempool_init_node()
308 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
312 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
314 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
319 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
327 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
330 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
341 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
345 spin_lock_irqsave(&pool->lock, flags); in mempool_resize()
349 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize()
[all …]
Dlist_lru.c54 lockdep_is_held(&nlru->lock)); in list_lru_from_memcg_idx()
122 spin_lock(&nlru->lock); in list_lru_add()
131 spin_unlock(&nlru->lock); in list_lru_add()
134 spin_unlock(&nlru->lock); in list_lru_add()
145 spin_lock(&nlru->lock); in list_lru_del()
151 spin_unlock(&nlru->lock); in list_lru_del()
154 spin_unlock(&nlru->lock); in list_lru_del()
222 ret = isolate(item, l, &nlru->lock, cb_arg); in __list_lru_walk_one()
225 assert_spin_locked(&nlru->lock); in __list_lru_walk_one()
248 assert_spin_locked(&nlru->lock); in __list_lru_walk_one()
[all …]
Dzbud.c95 spinlock_t lock; member
314 spin_lock_init(&pool->lock); in zbud_create_pool()
367 spin_lock(&pool->lock); in zbud_alloc()
384 spin_unlock(&pool->lock); in zbud_alloc()
388 spin_lock(&pool->lock); in zbud_alloc()
414 spin_unlock(&pool->lock); in zbud_alloc()
434 spin_lock(&pool->lock); in zbud_free()
445 spin_unlock(&pool->lock); in zbud_free()
463 spin_unlock(&pool->lock); in zbud_free()
507 spin_lock(&pool->lock); in zbud_reclaim_page()
[all …]
Dkmemleak.c139 raw_spinlock_t lock; member
589 raw_spin_lock_init(&object->lock); in create_object()
671 raw_spin_lock_irqsave(&object->lock, flags); in __delete_object()
673 raw_spin_unlock_irqrestore(&object->lock, flags); in __delete_object()
743 raw_spin_lock_irqsave(&object->lock, flags); in paint_it()
745 raw_spin_unlock_irqrestore(&object->lock, flags); in paint_it()
807 raw_spin_lock_irqsave(&object->lock, flags); in add_scan_area()
829 raw_spin_unlock_irqrestore(&object->lock, flags); in add_scan_area()
851 raw_spin_lock_irqsave(&object->lock, flags); in object_set_excess_ref()
853 raw_spin_unlock_irqrestore(&object->lock, flags); in object_set_excess_ref()
[all …]
Dpage_pinner.c42 spinlock_t lock; member
48 .lock = __SPIN_LOCK_UNLOCKED(lt_pinner.lock),
55 .lock = __SPIN_LOCK_UNLOCKED(acf_pinner.lock),
153 spin_lock_irqsave(&lt_pinner.lock, flags); in check_longterm_pin()
157 spin_unlock_irqrestore(&lt_pinner.lock, flags); in check_longterm_pin()
354 spin_lock_irqsave(&acf_pinner.lock, flags); in __page_pinner_migration_failed()
358 spin_unlock_irqrestore(&acf_pinner.lock, flags); in __page_pinner_migration_failed()
403 spin_lock_irqsave(&lt_pinner.lock, flags); in read_longterm_page_pinner()
405 spin_unlock_irqrestore(&lt_pinner.lock, flags); in read_longterm_page_pinner()
439 spin_lock_irqsave(&acf_pinner.lock, flags); in read_alloc_contig_failed()
[all …]
Ddmapool.c44 spinlock_t lock; member
85 spin_lock_irq(&pool->lock); in show_pools()
90 spin_unlock_irq(&pool->lock); in show_pools()
164 spin_lock_init(&retval->lock); in dma_pool_create()
324 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc()
331 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc()
337 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc()
373 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc()
410 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free()
413 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free()
[all …]
Dswapfile.c360 spin_lock(&ci->lock); in lock_cluster()
368 spin_unlock(&ci->lock); in unlock_cluster()
384 spin_lock(&si->lock); in lock_cluster_or_swap_info()
395 spin_unlock(&si->lock); in unlock_cluster_or_swap_info()
430 spin_lock_nested(&ci_tail->lock, SINGLE_DEPTH_NESTING); in cluster_list_add_tail()
432 spin_unlock(&ci_tail->lock); in cluster_list_add_tail()
492 spin_unlock(&si->lock); in swap_do_scheduled_discard()
497 spin_lock(&si->lock); in swap_do_scheduled_discard()
512 spin_lock(&si->lock); in swap_discard_work()
514 spin_unlock(&si->lock); in swap_discard_work()
[all …]
Dpage_isolation.c24 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate()
32 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
51 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
55 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate()
77 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate()
123 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate()
309 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated()
311 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
Dhighmem.c394 spinlock_t lock; /* Protect this bucket's list */ member
419 spin_lock_irqsave(&pas->lock, flags); in page_address()
431 spin_unlock_irqrestore(&pas->lock, flags); in page_address()
456 spin_lock_irqsave(&pas->lock, flags); in set_page_address()
458 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address()
460 spin_lock_irqsave(&pas->lock, flags); in set_page_address()
464 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address()
468 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address()
480 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
Dzswap.c184 spinlock_t lock; member
894 spin_lock(&tree->lock); in zswap_writeback_entry()
898 spin_unlock(&tree->lock); in zswap_writeback_entry()
902 spin_unlock(&tree->lock); in zswap_writeback_entry()
942 spin_lock(&tree->lock); in zswap_writeback_entry()
955 spin_unlock(&tree->lock); in zswap_writeback_entry()
967 spin_lock(&tree->lock); in zswap_writeback_entry()
969 spin_unlock(&tree->lock); in zswap_writeback_entry()
1115 spin_lock(&tree->lock); in zswap_frontswap_store()
1125 spin_unlock(&tree->lock); in zswap_frontswap_store()
[all …]
Dzsmalloc.c201 spinlock_t lock; member
291 rwlock_t lock; member
609 spin_lock(&class->lock); in zs_stats_size_show()
615 spin_unlock(&class->lock); in zs_stats_size_show()
940 assert_spin_locked(&class->lock); in __free_zspage()
1431 spin_lock(&class->lock); in zs_malloc()
1438 spin_unlock(&class->lock); in zs_malloc()
1443 spin_unlock(&class->lock); in zs_malloc()
1451 spin_lock(&class->lock); in zs_malloc()
1463 spin_unlock(&class->lock); in zs_malloc()
[all …]
Dswap_cgroup.c12 spinlock_t lock; member
107 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_cmpxchg()
113 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_cmpxchg()
138 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_record()
151 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_record()
185 spin_lock_init(&ctrl->lock); in swap_cgroup_swapon()
Dcma.c101 mutex_lock(&cma->lock); in cma_clear_bitmap()
103 mutex_unlock(&cma->lock); in cma_clear_bitmap()
132 mutex_init(&cma->lock); in cma_activate_area()
406 mutex_lock(&cma->lock); in cma_debug_show_areas()
421 mutex_unlock(&cma->lock); in cma_debug_show_areas()
477 mutex_lock(&cma->lock); in cma_alloc()
483 mutex_unlock(&cma->lock); in cma_alloc()
502 mutex_unlock(&cma->lock); in cma_alloc()
512 mutex_unlock(&cma->lock); in cma_alloc()
Dpage_reporting.c127 spin_lock_irq(&zone->lock); in page_reporting_cycle()
185 spin_unlock_irq(&zone->lock); in page_reporting_cycle()
197 spin_lock_irq(&zone->lock); in page_reporting_cycle()
217 spin_unlock_irq(&zone->lock); in page_reporting_cycle()
262 spin_lock_irq(&zone->lock); in page_reporting_process_zone()
264 spin_unlock_irq(&zone->lock); in page_reporting_process_zone()
Dshuffle.c89 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone()
142 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
144 spin_lock_irqsave(&z->lock, flags); in __shuffle_zone()
147 spin_unlock_irqrestore(&z->lock, flags); in __shuffle_zone()
Dcma_debug.c39 mutex_lock(&cma->lock); in cma_used_get()
42 mutex_unlock(&cma->lock); in cma_used_get()
56 mutex_lock(&cma->lock); in cma_maxchunk_get()
64 mutex_unlock(&cma->lock); in cma_maxchunk_get()
Dvmalloc.c1491 spinlock_t lock; member
1496 spinlock_t lock; member
1571 spin_lock_init(&vb->lock); in new_vmap_block()
1590 spin_lock(&vbq->lock); in new_vmap_block()
1592 spin_unlock(&vbq->lock); in new_vmap_block()
1622 spin_lock(&vb->lock); in purge_fragmented_blocks()
1628 spin_lock(&vbq->lock); in purge_fragmented_blocks()
1630 spin_unlock(&vbq->lock); in purge_fragmented_blocks()
1631 spin_unlock(&vb->lock); in purge_fragmented_blocks()
1634 spin_unlock(&vb->lock); in purge_fragmented_blocks()
[all …]
Dshmem.c393 spin_lock_irqsave(&info->lock, flags); in shmem_charge()
397 spin_unlock_irqrestore(&info->lock, flags); in shmem_charge()
409 spin_lock_irqsave(&info->lock, flags); in shmem_uncharge()
413 spin_unlock_irqrestore(&info->lock, flags); in shmem_uncharge()
1062 spin_lock_irq(&info->lock); in shmem_undo_range()
1065 spin_unlock_irq(&info->lock); in shmem_undo_range()
1083 spin_lock_irq(&info->lock); in shmem_getattr()
1085 spin_unlock_irq(&info->lock); in shmem_getattr()
1454 spin_lock_irq(&info->lock); in shmem_writepage()
1457 spin_unlock_irq(&info->lock); in shmem_writepage()
[all …]
Dcompaction.c504 static bool compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, in compact_lock_irqsave() argument
506 __acquires(lock) in compact_lock_irqsave()
510 if (spin_trylock_irqsave(lock, *flags)) in compact_lock_irqsave()
516 spin_lock_irqsave(lock, *flags); in compact_lock_irqsave()
535 static bool compact_unlock_should_abort(spinlock_t *lock, in compact_unlock_should_abort() argument
539 spin_unlock_irqrestore(lock, flags); in compact_unlock_should_abort()
589 && compact_unlock_should_abort(&cc->zone->lock, flags, in isolate_freepages_block()
624 locked = compact_lock_irqsave(&cc->zone->lock, in isolate_freepages_block()
661 spin_unlock_irqrestore(&cc->zone->lock, flags); in isolate_freepages_block()
1381 spin_lock_irqsave(&cc->zone->lock, flags); in fast_isolate_freepages()
[all …]
Dpage_alloc.c1496 spin_lock(&zone->lock); in free_pcppages_bulk()
1514 spin_unlock(&zone->lock); in free_pcppages_bulk()
1522 spin_lock(&zone->lock); in free_one_page()
1528 spin_unlock(&zone->lock); in free_one_page()
2761 spin_lock_irqsave(&zone->lock, flags); in reserve_highatomic_pageblock()
2777 spin_unlock_irqrestore(&zone->lock, flags); in reserve_highatomic_pageblock()
2816 spin_lock_irqsave(&zone->lock, flags); in unreserve_highatomic_pageblock()
2857 spin_unlock_irqrestore(&zone->lock, flags); in unreserve_highatomic_pageblock()
2861 spin_unlock_irqrestore(&zone->lock, flags); in unreserve_highatomic_pageblock()
3005 spin_lock(&zone->lock); in rmqueue_bulk()
[all …]
/mm/kfence/
Dcore.c191 lockdep_assert_held(&meta->lock); in metadata_update_state()
233 lockdep_assert_held(&meta->lock); in for_each_canary()
274 if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) { in kfence_guarded_alloc()
328 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_alloc()
356 raw_spin_lock_irqsave(&meta->lock, flags); in kfence_guarded_free()
363 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free()
396 raw_spin_unlock_irqrestore(&meta->lock, flags); in kfence_guarded_free()
475 raw_spin_lock_init(&meta->lock); in kfence_init_pool()
553 raw_spin_lock_irqsave(&meta->lock, flags); in show_object()
555 raw_spin_unlock_irqrestore(&meta->lock, flags); in show_object()
[all …]
Dkfence_test.c30 spinlock_t lock; member
34 .lock = __SPIN_LOCK_UNLOCKED(observed.lock),
43 spin_lock_irqsave(&observed.lock, flags); in probe_console()
60 spin_unlock_irqrestore(&observed.lock, flags); in probe_console()
151 spin_lock_irqsave(&observed.lock, flags); in report_matches()
158 spin_unlock_irqrestore(&observed.lock, flags); in report_matches()
791 spin_lock_irqsave(&observed.lock, flags); in test_init()
795 spin_unlock_irqrestore(&observed.lock, flags); in test_init()

12