/mm/ |
D | mremap.c | 262 bool *locked, struct vm_userfaultfd_ctx *uf, in move_vma() argument 368 *locked = true; in move_vma() 417 unsigned long locked, lock_limit; in vma_to_resize() local 418 locked = mm->locked_vm << PAGE_SHIFT; in vma_to_resize() 420 locked += new_len - old_len; in vma_to_resize() 421 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) in vma_to_resize() 440 unsigned long new_addr, unsigned long new_len, bool *locked, in mremap_to() argument 488 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, uf, in mremap_to() 527 bool locked = false; in SYSCALL_DEFINE5() local 557 &locked, &uf, &uf_unmap_early, &uf_unmap); in SYSCALL_DEFINE5() [all …]
|
D | gup.c | 866 int *locked, bool notify_drop, in __get_user_pages_locked() argument 872 if (locked) { in __get_user_pages_locked() 876 BUG_ON(*locked != 1); in __get_user_pages_locked() 886 vmas, locked); in __get_user_pages_locked() 887 if (!locked) in __get_user_pages_locked() 892 if (!*locked) { in __get_user_pages_locked() 907 if (*locked) { in __get_user_pages_locked() 922 *locked = 1; in __get_user_pages_locked() 940 if (notify_drop && lock_dropped && *locked) { in __get_user_pages_locked() 946 *locked = 0; in __get_user_pages_locked() [all …]
|
D | compaction.c | 357 unsigned long flags, bool *locked, struct compact_control *cc) in compact_unlock_should_abort() argument 359 if (*locked) { in compact_unlock_should_abort() 361 *locked = false; in compact_unlock_should_abort() 418 bool locked = false; in isolate_freepages_block() local 436 &locked, cc)) in isolate_freepages_block() 473 if (!locked) { in isolate_freepages_block() 482 locked = compact_trylock_irqsave(&cc->zone->lock, in isolate_freepages_block() 484 if (!locked) in isolate_freepages_block() 520 if (locked) in isolate_freepages_block() 675 bool locked = false; in isolate_migratepages_block() local [all …]
|
D | mlock.c | 666 unsigned long locked; in do_mlock() local 680 locked = len >> PAGE_SHIFT; in do_mlock() 685 locked += current->mm->locked_vm; in do_mlock() 686 if ((locked > lock_limit) && (!capable(CAP_IPC_LOCK))) { in do_mlock() 693 locked -= count_mm_mlocked_page_nr(current->mm, in do_mlock() 698 if ((locked <= lock_limit) || capable(CAP_IPC_LOCK)) in do_mlock() 840 unsigned long lock_limit, locked; in user_shm_lock() local 843 locked = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in user_shm_lock() 850 locked + user->locked_shm > lock_limit && !capable(CAP_IPC_LOCK)) in user_shm_lock() 853 user->locked_shm += locked; in user_shm_lock()
|
D | frame_vector.c | 41 int locked; in get_vaddr_frames() local 50 locked = 1; in get_vaddr_frames() 74 gup_flags, (struct page **)(vec->ptrs), &locked); in get_vaddr_frames() 102 if (locked) in get_vaddr_frames()
|
D | process_vm_access.c | 104 int locked = 1; in process_vm_rw_single_vec() local 114 process_pages, NULL, &locked); in process_vm_rw_single_vec() 115 if (locked) in process_vm_rw_single_vec()
|
D | rmap.c | 1747 bool locked) in rmap_walk_anon() argument 1753 if (locked) { in rmap_walk_anon() 1781 if (!locked) in rmap_walk_anon() 1799 bool locked) in rmap_walk_file() argument 1818 if (!locked) in rmap_walk_file() 1836 if (!locked) in rmap_walk_file()
|
D | z3fold.c | 258 static void __release_z3fold_page(struct z3fold_header *zhdr, bool locked) in __release_z3fold_page() argument 270 if (locked) in __release_z3fold_page() 407 static void do_compact_page(struct z3fold_header *zhdr, bool locked) in do_compact_page() argument 415 if (locked) in do_compact_page()
|
D | mmap.c | 1308 unsigned long locked, lock_limit; in mlock_future_check() local 1312 locked = len >> PAGE_SHIFT; in mlock_future_check() 1313 locked += mm->locked_vm; in mlock_future_check() 1316 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) in mlock_future_check() 2229 unsigned long locked; in acct_stack_growth() local 2231 locked = mm->locked_vm + grow; in acct_stack_growth() 2234 if (locked > limit && !capable(CAP_IPC_LOCK)) in acct_stack_growth()
|
D | memcontrol.c | 1593 bool locked; in mem_cgroup_oom_synchronize() local 1611 locked = mem_cgroup_oom_trylock(memcg); in mem_cgroup_oom_synchronize() 1613 if (locked) in mem_cgroup_oom_synchronize() 1616 if (locked && !memcg->oom_kill_disable) { in mem_cgroup_oom_synchronize() 1627 if (locked) { in mem_cgroup_oom_synchronize()
|
D | migrate.c | 289 void remove_migration_ptes(struct page *old, struct page *new, bool locked) in remove_migration_ptes() argument 296 if (locked) in remove_migration_ptes()
|
D | memory.c | 2880 int locked; in do_swap_page() local 2954 locked = lock_page_or_retry(page, vma->vm_mm, vmf->flags); in do_swap_page() 2957 if (!locked) { in do_swap_page()
|
D | nommu.c | 174 int *locked) in get_user_pages_locked() argument
|