/mm/ |
D | list_lru.c | 18 spin_lock(&nlru->lock); in list_lru_add() 24 spin_unlock(&nlru->lock); in list_lru_add() 27 spin_unlock(&nlru->lock); in list_lru_add() 37 spin_lock(&nlru->lock); in list_lru_del() 43 spin_unlock(&nlru->lock); in list_lru_del() 46 spin_unlock(&nlru->lock); in list_lru_del() 57 spin_lock(&nlru->lock); in list_lru_count_node() 60 spin_unlock(&nlru->lock); in list_lru_count_node() 75 spin_lock(&nlru->lock); in list_lru_walk_node() 88 ret = isolate(item, &nlru->lock, cb_arg); in list_lru_walk_node() [all …]
|
D | mempool.c | 86 spin_lock_init(&pool->lock); in mempool_create_node() 134 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 138 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 140 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 145 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 152 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 155 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 166 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() 170 spin_lock_irqsave(&pool->lock, flags); in mempool_resize() 174 spin_unlock_irqrestore(&pool->lock, flags); in mempool_resize() [all …]
|
D | kmemleak.c | 140 spinlock_t lock; member 535 spin_lock_init(&object->lock); in create_object() 585 spin_lock(&object->lock); in create_object() 587 spin_unlock(&object->lock); in create_object() 620 spin_lock_irqsave(&object->lock, flags); in __delete_object() 622 spin_unlock_irqrestore(&object->lock, flags); in __delete_object() 696 spin_lock_irqsave(&object->lock, flags); in paint_it() 698 spin_unlock_irqrestore(&object->lock, flags); in paint_it() 758 spin_lock_irqsave(&object->lock, flags); in add_scan_area() 774 spin_unlock_irqrestore(&object->lock, flags); in add_scan_area() [all …]
|
D | zbud.c | 94 spinlock_t lock; member 298 spin_lock_init(&pool->lock); in zbud_create_pool() 351 spin_lock(&pool->lock); in zbud_alloc() 369 spin_unlock(&pool->lock); in zbud_alloc() 373 spin_lock(&pool->lock); in zbud_alloc() 399 spin_unlock(&pool->lock); in zbud_alloc() 419 spin_lock(&pool->lock); in zbud_free() 430 spin_unlock(&pool->lock); in zbud_free() 448 spin_unlock(&pool->lock); in zbud_free() 495 spin_lock(&pool->lock); in zbud_reclaim_page() [all …]
|
D | zswap.c | 199 spinlock_t lock; member 556 spin_lock(&tree->lock); in zswap_writeback_entry() 560 spin_unlock(&tree->lock); in zswap_writeback_entry() 563 spin_unlock(&tree->lock); in zswap_writeback_entry() 603 spin_lock(&tree->lock); in zswap_writeback_entry() 616 spin_unlock(&tree->lock); in zswap_writeback_entry() 628 spin_lock(&tree->lock); in zswap_writeback_entry() 630 spin_unlock(&tree->lock); in zswap_writeback_entry() 710 spin_lock(&tree->lock); in zswap_frontswap_store() 720 spin_unlock(&tree->lock); in zswap_frontswap_store() [all …]
|
D | dmapool.c | 47 spinlock_t lock; member 88 spin_lock_irq(&pool->lock); in show_pools() 93 spin_unlock_irq(&pool->lock); in show_pools() 167 spin_lock_init(&retval->lock); in dma_pool_create() 328 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 335 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 341 spin_lock_irqsave(&pool->lock, flags); in dma_pool_alloc() 377 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_alloc() 410 spin_lock_irqsave(&pool->lock, flags); in dma_pool_free() 413 spin_unlock_irqrestore(&pool->lock, flags); in dma_pool_free() [all …]
|
D | highmem.c | 398 spinlock_t lock; /* Protect this bucket's list */ member 423 spin_lock_irqsave(&pas->lock, flags); in page_address() 435 spin_unlock_irqrestore(&pas->lock, flags); in page_address() 460 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 462 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 464 spin_lock_irqsave(&pas->lock, flags); in set_page_address() 468 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 472 spin_unlock_irqrestore(&pas->lock, flags); in set_page_address() 484 spin_lock_init(&page_address_htable[i].lock); in page_address_init()
|
D | mmu_notifier.c | 76 spin_lock(&mm->mmu_notifier_mm->lock); in __mmu_notifier_release() 89 spin_unlock(&mm->mmu_notifier_mm->lock); in __mmu_notifier_release() 231 spin_lock_init(&mmu_notifier_mm->lock); in do_mmu_notifier_register() 246 spin_lock(&mm->mmu_notifier_mm->lock); in do_mmu_notifier_register() 248 spin_unlock(&mm->mmu_notifier_mm->lock); in do_mmu_notifier_register() 327 spin_lock(&mm->mmu_notifier_mm->lock); in mmu_notifier_unregister() 333 spin_unlock(&mm->mmu_notifier_mm->lock); in mmu_notifier_unregister() 354 spin_lock(&mm->mmu_notifier_mm->lock); in mmu_notifier_unregister_no_release() 360 spin_unlock(&mm->mmu_notifier_mm->lock); in mmu_notifier_unregister_no_release()
|
D | swapfile.c | 305 spin_unlock(&si->lock); in swap_do_scheduled_discard() 310 spin_lock(&si->lock); in swap_do_scheduled_discard() 336 spin_lock(&si->lock); in swap_discard_work() 338 spin_unlock(&si->lock); in swap_discard_work() 521 spin_unlock(&si->lock); in scan_swap_map() 537 spin_lock(&si->lock); in scan_swap_map() 550 spin_lock(&si->lock); in scan_swap_map() 569 spin_unlock(&si->lock); in scan_swap_map() 571 spin_lock(&si->lock); in scan_swap_map() 601 spin_unlock(&si->lock); in scan_swap_map() [all …]
|
D | cma.c | 43 struct mutex lock; member 98 mutex_lock(&cma->lock); in cma_clear_bitmap() 100 mutex_unlock(&cma->lock); in cma_clear_bitmap() 136 mutex_init(&cma->lock); in cma_activate_area() 385 mutex_lock(&cma->lock); in cma_alloc() 390 mutex_unlock(&cma->lock); in cma_alloc() 399 mutex_unlock(&cma->lock); in cma_alloc()
|
D | page_isolation.c | 22 spin_lock_irqsave(&zone->lock, flags); in set_migratetype_isolate() 69 spin_unlock_irqrestore(&zone->lock, flags); in set_migratetype_isolate() 85 spin_lock_irqsave(&zone->lock, flags); in unset_migratetype_isolate() 125 spin_unlock_irqrestore(&zone->lock, flags); in unset_migratetype_isolate() 284 spin_lock_irqsave(&zone->lock, flags); in test_pages_isolated() 287 spin_unlock_irqrestore(&zone->lock, flags); in test_pages_isolated()
|
D | page_cgroup.c | 331 spinlock_t lock; member 420 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 426 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_cmpxchg() 447 spin_lock_irqsave(&ctrl->lock, flags); in swap_cgroup_record() 450 spin_unlock_irqrestore(&ctrl->lock, flags); in swap_cgroup_record() 487 spin_lock_init(&ctrl->lock); in swap_cgroup_swapon()
|
D | vmalloc.c | 756 spinlock_t lock; member 761 spinlock_t lock; member 825 spin_lock_init(&vb->lock); in new_vmap_block() 840 spin_lock(&vbq->lock); in new_vmap_block() 842 spin_unlock(&vbq->lock); in new_vmap_block() 876 spin_lock(&vb->lock); in purge_fragmented_blocks() 881 spin_lock(&vbq->lock); in purge_fragmented_blocks() 883 spin_unlock(&vbq->lock); in purge_fragmented_blocks() 884 spin_unlock(&vb->lock); in purge_fragmented_blocks() 887 spin_unlock(&vb->lock); in purge_fragmented_blocks() [all …]
|
D | zsmalloc.c | 224 spinlock_t lock; member 506 spin_lock(&class->lock); in zs_stats_size_show() 511 spin_unlock(&class->lock); in zs_stats_size_show() 1403 spin_lock(&class->lock); in zs_malloc() 1407 spin_unlock(&class->lock); in zs_malloc() 1418 spin_lock(&class->lock); in zs_malloc() 1427 spin_unlock(&class->lock); in zs_malloc() 1484 spin_lock(&class->lock); in zs_free() 1494 spin_unlock(&class->lock); in zs_free() 1715 spin_lock(&class->lock); in __zs_compact() [all …]
|
D | shmem.c | 537 spin_lock(&info->lock); in shmem_undo_range() 540 spin_unlock(&info->lock); in shmem_undo_range() 683 spin_lock(&info->lock); in shmem_unuse_inode() 685 spin_unlock(&info->lock); in shmem_unuse_inode() 831 spin_lock(&info->lock); in shmem_writepage() 834 spin_unlock(&info->lock); in shmem_writepage() 1148 spin_lock(&info->lock); in shmem_getpage_gfp() 1151 spin_unlock(&info->lock); in shmem_getpage_gfp() 1201 spin_lock(&info->lock); in shmem_getpage_gfp() 1205 spin_unlock(&info->lock); in shmem_getpage_gfp() [all …]
|
D | compaction.c | 234 static bool compact_trylock_irqsave(spinlock_t *lock, unsigned long *flags, in compact_trylock_irqsave() argument 238 if (!spin_trylock_irqsave(lock, *flags)) { in compact_trylock_irqsave() 243 spin_lock_irqsave(lock, *flags); in compact_trylock_irqsave() 264 static bool compact_unlock_should_abort(spinlock_t *lock, in compact_unlock_should_abort() argument 268 spin_unlock_irqrestore(lock, flags); in compact_unlock_should_abort() 364 && compact_unlock_should_abort(&cc->zone->lock, flags, in isolate_freepages_block() 411 locked = compact_trylock_irqsave(&cc->zone->lock, in isolate_freepages_block() 450 spin_unlock_irqrestore(&cc->zone->lock, flags); in isolate_freepages_block()
|
D | hugetlb.c | 68 spin_unlock(&spool->lock); in unlock_or_release_subpool() 84 spin_lock_init(&spool->lock); in hugepage_new_subpool() 94 spin_lock(&spool->lock); in hugepage_put_subpool() 108 spin_lock(&spool->lock); in hugepage_subpool_get_pages() 114 spin_unlock(&spool->lock); in hugepage_subpool_get_pages() 125 spin_lock(&spool->lock); in hugepage_subpool_put_pages() 160 spin_lock(&resv->lock); in region_add() 190 spin_unlock(&resv->lock); in region_add() 201 spin_lock(&resv->lock); in region_chg() 212 spin_unlock(&resv->lock); in region_chg() [all …]
|
D | memory_hotplug.c | 55 struct mutex lock; /* Synchronizes accesses to refcount, */ member 67 .lock = __MUTEX_INITIALIZER(mem_hotplug.lock), 85 mutex_lock(&mem_hotplug.lock); in get_online_mems() 87 mutex_unlock(&mem_hotplug.lock); in get_online_mems() 95 mutex_lock(&mem_hotplug.lock); in put_online_mems() 102 mutex_unlock(&mem_hotplug.lock); in put_online_mems() 113 mutex_lock(&mem_hotplug.lock); in mem_hotplug_begin() 117 mutex_unlock(&mem_hotplug.lock); in mem_hotplug_begin() 125 mutex_unlock(&mem_hotplug.lock); in mem_hotplug_done()
|
D | memory-failure.c | 1310 spinlock_t lock; member 1344 spin_lock_irqsave(&mf_cpu->lock, proc_flags); in memory_failure_queue() 1350 spin_unlock_irqrestore(&mf_cpu->lock, proc_flags); in memory_failure_queue() 1364 spin_lock_irqsave(&mf_cpu->lock, proc_flags); in memory_failure_work_func() 1366 spin_unlock_irqrestore(&mf_cpu->lock, proc_flags); in memory_failure_work_func() 1383 spin_lock_init(&mf_cpu->lock); in memory_failure_init()
|
D | page_alloc.c | 709 spin_lock(&zone->lock); in free_pcppages_bulk() 751 spin_unlock(&zone->lock); in free_pcppages_bulk() 760 spin_lock(&zone->lock); in free_one_page() 770 spin_unlock(&zone->lock); in free_one_page() 1254 spin_lock(&zone->lock); in rmqueue_bulk() 1279 spin_unlock(&zone->lock); in rmqueue_bulk() 1400 spin_lock_irqsave(&zone->lock, flags); in mark_free_pages() 1420 spin_unlock_irqrestore(&zone->lock, flags); in mark_free_pages() 1634 spin_lock_irqsave(&zone->lock, flags); in buffered_rmqueue() 1636 spin_unlock(&zone->lock); in buffered_rmqueue() [all …]
|
D | memcontrol.c | 186 spinlock_t lock; member 430 spinlock_t lock; /* for from, to */ member 440 .lock = __SPIN_LOCK_UNLOCKED(mc.lock), 753 spin_lock_irqsave(&mctz->lock, flags); in mem_cgroup_remove_exceeded() 755 spin_unlock_irqrestore(&mctz->lock, flags); in mem_cgroup_remove_exceeded() 780 spin_lock_irqsave(&mctz->lock, flags); in mem_cgroup_update_tree() 789 spin_unlock_irqrestore(&mctz->lock, flags); in mem_cgroup_update_tree() 840 spin_lock_irq(&mctz->lock); in mem_cgroup_largest_soft_limit_node() 842 spin_unlock_irq(&mctz->lock); in mem_cgroup_largest_soft_limit_node() 1571 spin_lock(&mc.lock); in mem_cgroup_under_move() [all …]
|
D | mlock.c | 562 int lock = !!(newflags & VM_LOCKED); in mlock_fixup() local 594 if (!lock) in mlock_fixup() 604 if (lock) in mlock_fixup()
|
D | mempolicy.c | 2232 spin_lock(&sp->lock); in mpol_shared_policy_lookup() 2238 spin_unlock(&sp->lock); in mpol_shared_policy_lookup() 2381 spin_lock(&sp->lock); in shared_policy_replace() 2414 spin_unlock(&sp->lock); in shared_policy_replace() 2426 spin_unlock(&sp->lock); in shared_policy_replace() 2452 spin_lock_init(&sp->lock); in mpol_shared_policy_init() 2518 spin_lock(&p->lock); in mpol_free_shared_policy() 2525 spin_unlock(&p->lock); in mpol_free_shared_policy()
|
D | swap.c | 861 static DEFINE_MUTEX(lock); in lru_add_drain_all() 865 mutex_lock(&lock); in lru_add_drain_all() 886 mutex_unlock(&lock); in lru_add_drain_all()
|
D | slab.c | 206 spinlock_t lock; member 873 spin_lock_init(&alc->lock); in __alloc_alien_cache() 949 if (ac->avail && spin_trylock_irq(&alc->lock)) { in reap_alien() 953 spin_unlock_irq(&alc->lock); in reap_alien() 974 spin_lock_irqsave(&alc->lock, flags); in drain_alien_cache() 976 spin_unlock_irqrestore(&alc->lock, flags); in drain_alien_cache() 995 spin_lock(&alien->lock); in __cache_free_alien() 1001 spin_unlock(&alien->lock); in __cache_free_alien()
|