/mm/ |
D | gup.c | 898 unsigned long address, unsigned int *flags, int *locked) in faultin_page() argument 910 if (locked) in faultin_page() 932 if (locked && !(fault_flags & FAULT_FLAG_RETRY_NOWAIT)) in faultin_page() 933 *locked = 0; in faultin_page() 1064 struct vm_area_struct **vmas, int *locked) in __get_user_pages() argument 1114 gup_flags, locked); in __get_user_pages() 1115 if (locked && *locked == 0) { in __get_user_pages() 1141 ret = faultin_page(vma, start, &foll_flags, locked); in __get_user_pages() 1296 int *locked, in __get_user_pages_locked() argument 1302 if (locked) { in __get_user_pages_locked() [all …]
|
D | mremap.c | 542 bool *locked, unsigned long flags, in move_vma() argument 679 *locked = true; in move_vma() 742 unsigned long locked, lock_limit; in vma_to_resize() local 743 locked = mm->locked_vm << PAGE_SHIFT; in vma_to_resize() 745 locked += new_len - old_len; in vma_to_resize() 746 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) in vma_to_resize() 765 unsigned long new_addr, unsigned long new_len, bool *locked, in mremap_to() argument 845 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, flags, uf, in mremap_to() 886 bool locked = false; in SYSCALL_DEFINE5() local 938 &locked, flags, &uf, &uf_unmap_early, in SYSCALL_DEFINE5() [all …]
|
D | mlock.c | 689 unsigned long locked; in do_mlock() local 703 locked = len >> PAGE_SHIFT; in do_mlock() 708 locked += current->mm->locked_vm; in do_mlock() 709 if ((locked > lock_limit) && (!capable(CAP_IPC_LOCK))) { in do_mlock() 716 locked -= count_mm_mlocked_page_nr(current->mm, in do_mlock() 721 if ((locked <= lock_limit) || capable(CAP_IPC_LOCK)) in do_mlock() 863 unsigned long lock_limit, locked; in user_shm_lock() local 866 locked = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in user_shm_lock() 873 locked + user->locked_shm > lock_limit && !capable(CAP_IPC_LOCK)) in user_shm_lock() 876 user->locked_shm += locked; in user_shm_lock()
|
D | frame_vector.c | 44 int locked; in get_vaddr_frames() local 55 locked = 1; in get_vaddr_frames() 79 gup_flags, (struct page **)(vec->ptrs), &locked); in get_vaddr_frames() 87 if (locked) in get_vaddr_frames()
|
D | compaction.c | 536 unsigned long flags, bool *locked, struct compact_control *cc) in compact_unlock_should_abort() argument 538 if (*locked) { in compact_unlock_should_abort() 540 *locked = false; in compact_unlock_should_abort() 568 bool locked = false; in isolate_freepages_block() local 590 &locked, cc)) in isolate_freepages_block() 623 if (!locked) { in isolate_freepages_block() 624 locked = compact_lock_irqsave(&cc->zone->lock, in isolate_freepages_block() 660 if (locked) in isolate_freepages_block() 837 bool locked = false; in isolate_migratepages_block() local 903 flags, &locked, cc)) { in isolate_migratepages_block() [all …]
|
D | process_vm_access.c | 97 int locked = 1; in process_vm_rw_single_vec() local 108 NULL, &locked); in process_vm_rw_single_vec() 109 if (locked) in process_vm_rw_single_vec()
|
D | z3fold.c | 262 int locked = 0; in __get_z3fold_header() local 273 locked = z3fold_page_trylock(zhdr); in __get_z3fold_header() 275 if (locked) in __get_z3fold_header() 522 static void __release_z3fold_page(struct z3fold_header *zhdr, bool locked) in __release_z3fold_page() argument 535 if (locked) in __release_z3fold_page() 819 static void do_compact_page(struct z3fold_header *zhdr, bool locked) in do_compact_page() argument 825 if (locked) in do_compact_page()
|
D | rmap.c | 1959 bool locked) in rmap_walk_anon() argument 1965 if (locked) { in rmap_walk_anon() 1994 if (!locked) in rmap_walk_anon() 2012 bool locked) in rmap_walk_file() argument 2032 if (!locked) { in rmap_walk_file() 2068 if (!locked) in rmap_walk_file()
|
D | mmap.c | 1421 unsigned long locked, lock_limit; in mlock_future_check() local 1425 locked = len >> PAGE_SHIFT; in mlock_future_check() 1426 locked += mm->locked_vm; in mlock_future_check() 1429 if (locked > lock_limit && !capable(CAP_IPC_LOCK)) in mlock_future_check() 2490 unsigned long locked; in acct_stack_growth() local 2492 locked = mm->locked_vm + grow; in acct_stack_growth() 2495 if (locked > limit && !capable(CAP_IPC_LOCK)) in acct_stack_growth()
|
D | memcontrol.c | 1959 bool locked; in mem_cgroup_oom() local 1997 locked = mem_cgroup_oom_trylock(memcg); in mem_cgroup_oom() 1999 if (locked) in mem_cgroup_oom() 2008 if (locked) in mem_cgroup_oom() 2035 bool locked; in mem_cgroup_oom_synchronize() local 2053 locked = mem_cgroup_oom_trylock(memcg); in mem_cgroup_oom_synchronize() 2055 if (locked) in mem_cgroup_oom_synchronize() 2058 if (locked && !memcg->oom_kill_disable) { in mem_cgroup_oom_synchronize() 2069 if (locked) { in mem_cgroup_oom_synchronize()
|
D | memory.c | 2907 bool locked = false; in cow_user_page() local 2934 locked = true; in cow_user_page() 2957 if (locked) in cow_user_page() 2962 locked = true; in cow_user_page() 2988 if (locked) in cow_user_page() 3620 int locked; in do_swap_page() local 3761 locked = lock_page_or_retry(page, vma->vm_mm, vmf->flags); in do_swap_page() 3764 if (!locked) { in do_swap_page()
|
D | mempolicy.c | 933 int locked = 1; in lookup_node() local 934 err = get_user_pages_locked(addr & PAGE_MASK, 1, 0, &p, &locked); in lookup_node() 939 if (locked) in lookup_node()
|
D | hugetlb.c | 4896 long i, unsigned int flags, int *locked) in follow_hugetlb_page() argument 4967 if (locked) in follow_hugetlb_page() 4987 if (locked && in follow_hugetlb_page() 4989 *locked = 0; in follow_hugetlb_page()
|
D | migrate.c | 272 void remove_migration_ptes(struct page *old, struct page *new, bool locked) in remove_migration_ptes() argument 279 if (locked) in remove_migration_ptes()
|
D | Kconfig | 193 # a per-page lock leads to problems when multiple tables need to be locked
|