/mm/ |
D | mremap.c | 321 bool *locked, struct vm_userfaultfd_ctx *uf, in move_vma() argument 427 *locked = true; in move_vma() 476 unsigned long locked, lock_limit; in vma_to_resize() local 477 locked = mm->locked_vm << PAGE_SHIFT; in vma_to_resize() 479 locked += new_len - old_len; in vma_to_resize() 480 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) in vma_to_resize() 499 unsigned long new_addr, unsigned long new_len, bool *locked, in mremap_to() argument 564 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, uf, in mremap_to() 603 bool locked = false; in SYSCALL_DEFINE5() local 637 &locked, &uf, &uf_unmap_early, &uf_unmap); in SYSCALL_DEFINE5() [all …]
|
D | gup.c | 1004 int *locked, in __get_user_pages_locked() argument 1010 if (locked) { in __get_user_pages_locked() 1014 BUG_ON(*locked != 1); in __get_user_pages_locked() 1024 vmas, locked); in __get_user_pages_locked() 1025 if (!locked) in __get_user_pages_locked() 1030 if (!*locked) { in __get_user_pages_locked() 1041 if (*locked) { in __get_user_pages_locked() 1063 *locked = 1; in __get_user_pages_locked() 1082 if (lock_dropped && *locked) { in __get_user_pages_locked() 1088 *locked = 0; in __get_user_pages_locked() [all …]
|
D | mlock.c | 673 unsigned long locked; in do_mlock() local 687 locked = len >> PAGE_SHIFT; in do_mlock() 692 locked += current->mm->locked_vm; in do_mlock() 693 if ((locked > lock_limit) && (!capable(CAP_IPC_LOCK))) { in do_mlock() 700 locked -= count_mm_mlocked_page_nr(current->mm, in do_mlock() 705 if ((locked <= lock_limit) || capable(CAP_IPC_LOCK)) in do_mlock() 847 unsigned long lock_limit, locked; in user_shm_lock() local 850 locked = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in user_shm_lock() 857 locked + user->locked_shm > lock_limit && !capable(CAP_IPC_LOCK)) in user_shm_lock() 860 user->locked_shm += locked; in user_shm_lock()
|
D | frame_vector.c | 41 int locked; in get_vaddr_frames() local 52 locked = 1; in get_vaddr_frames() 76 gup_flags, (struct page **)(vec->ptrs), &locked); in get_vaddr_frames() 104 if (locked) in get_vaddr_frames()
|
D | compaction.c | 513 unsigned long flags, bool *locked, struct compact_control *cc) in compact_unlock_should_abort() argument 515 if (*locked) { in compact_unlock_should_abort() 517 *locked = false; in compact_unlock_should_abort() 545 bool locked = false; in isolate_freepages_block() local 567 &locked, cc)) in isolate_freepages_block() 600 if (!locked) { in isolate_freepages_block() 601 locked = compact_lock_irqsave(&cc->zone->lock, in isolate_freepages_block() 637 if (locked) in isolate_freepages_block() 789 bool locked = false; in isolate_migratepages_block() local 851 flags, &locked, cc)) { in isolate_migratepages_block() [all …]
|
D | process_vm_access.c | 100 int locked = 1; in process_vm_rw_single_vec() local 110 process_pages, NULL, &locked); in process_vm_rw_single_vec() 111 if (locked) in process_vm_rw_single_vec()
|
D | rmap.c | 1813 bool locked) in rmap_walk_anon() argument 1819 if (locked) { in rmap_walk_anon() 1847 if (!locked) in rmap_walk_anon() 1865 bool locked) in rmap_walk_file() argument 1884 if (!locked) in rmap_walk_file() 1902 if (!locked) in rmap_walk_file()
|
D | z3fold.c | 441 static void __release_z3fold_page(struct z3fold_header *zhdr, bool locked) in __release_z3fold_page() argument 453 if (locked) in __release_z3fold_page() 611 static void do_compact_page(struct z3fold_header *zhdr, bool locked) in do_compact_page() argument 617 if (locked) in do_compact_page()
|
D | mmap.c | 1351 unsigned long locked, lock_limit; in mlock_future_check() local 1355 locked = len >> PAGE_SHIFT; in mlock_future_check() 1356 locked += mm->locked_vm; in mlock_future_check() 1359 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) in mlock_future_check() 2322 unsigned long locked; in acct_stack_growth() local 2324 locked = mm->locked_vm + grow; in acct_stack_growth() 2327 if (locked > limit && !capable(CAP_IPC_LOCK)) in acct_stack_growth()
|
D | memcontrol.c | 1881 bool locked; in mem_cgroup_oom() local 1919 locked = mem_cgroup_oom_trylock(memcg); in mem_cgroup_oom() 1921 if (locked) in mem_cgroup_oom() 1930 if (locked) in mem_cgroup_oom() 1957 bool locked; in mem_cgroup_oom_synchronize() local 1975 locked = mem_cgroup_oom_trylock(memcg); in mem_cgroup_oom_synchronize() 1977 if (locked) in mem_cgroup_oom_synchronize() 1980 if (locked && !memcg->oom_kill_disable) { in mem_cgroup_oom_synchronize() 1991 if (locked) { in mem_cgroup_oom_synchronize()
|
D | mempolicy.c | 860 int locked = 1; in lookup_node() local 861 err = get_user_pages_locked(addr & PAGE_MASK, 1, 0, &p, &locked); in lookup_node() 866 if (locked) in lookup_node()
|
D | migrate.c | 289 void remove_migration_ptes(struct page *old, struct page *new, bool locked) in remove_migration_ptes() argument 296 if (locked) in remove_migration_ptes()
|
D | memory.c | 2777 int locked; in do_swap_page() local 2854 locked = lock_page_or_retry(page, vma->vm_mm, vmf->flags); in do_swap_page() 2857 if (!locked) { in do_swap_page()
|