/mm/ |
D | maccess.c | 123 long ret; in copy_from_kernel_nofault() local 131 ret = __copy_from_user_inatomic(dst, (__force const void __user *)src, in copy_from_kernel_nofault() 136 if (ret) in copy_from_kernel_nofault() 153 long ret; in copy_to_kernel_nofault() local 158 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); in copy_to_kernel_nofault() 162 if (ret) in copy_to_kernel_nofault() 190 long ret; in strncpy_from_kernel_nofault() local 201 ret = __get_user(*dst++, (const char __user __force *)src++); in strncpy_from_kernel_nofault() 202 } while (dst[-1] && ret == 0 && src - unsafe_addr < count); in strncpy_from_kernel_nofault() 208 return ret ? -EFAULT : src - unsafe_addr; in strncpy_from_kernel_nofault() [all …]
|
D | memory-failure.c | 72 int ret; in __page_handle_poison() local 75 ret = dissolve_free_huge_page(page); in __page_handle_poison() 76 if (!ret) in __page_handle_poison() 77 ret = take_page_off_buddy(page); in __page_handle_poison() 80 return ret > 0; in __page_handle_poison() 255 int ret = 0; in kill_proc() local 262 ret = force_sig_mceerr(BUS_MCEERR_AR, in kill_proc() 266 ret = send_sig_mceerr(BUS_MCEERR_AO, (void __user *)tk->addr, in kill_proc() 275 ret = send_sig_mceerr(BUS_MCEERR_AO, (void __user *)tk->addr, in kill_proc() 278 if (ret < 0) in kill_proc() [all …]
|
D | page_io.c | 92 int ret; in generic_swapfile_activate() local 112 ret = bmap(inode, &first_block); in generic_swapfile_activate() 113 if (ret || !first_block) in generic_swapfile_activate() 129 ret = bmap(inode, &block); in generic_swapfile_activate() 130 if (ret || !block) in generic_swapfile_activate() 151 ret = add_swap_extent(sis, page_no, 1, first_block); in generic_swapfile_activate() 152 if (ret < 0) in generic_swapfile_activate() 154 nr_extents += ret; in generic_swapfile_activate() 160 ret = nr_extents; in generic_swapfile_activate() 168 return ret; in generic_swapfile_activate() [all …]
|
D | backing-dev.c | 139 ssize_t ret; in read_ahead_kb_store() local 141 ret = kstrtoul(buf, 10, &read_ahead_kb); in read_ahead_kb_store() 142 if (ret < 0) in read_ahead_kb_store() 143 return ret; in read_ahead_kb_store() 167 ssize_t ret; in min_ratio_store() local 169 ret = kstrtouint(buf, 10, &ratio); in min_ratio_store() 170 if (ret < 0) in min_ratio_store() 171 return ret; in min_ratio_store() 173 ret = bdi_set_min_ratio(bdi, ratio); in min_ratio_store() 174 if (!ret) in min_ratio_store() [all …]
|
D | memory_hotplug.c | 67 int ret = sysfs_match_string(online_policy_to_str, val); in set_online_policy() local 69 if (ret < 0) in set_online_policy() 70 return ret; in set_online_policy() 71 *((int *)kp->arg) = ret; in set_online_policy() 1035 int ret; in mhp_init_memmap_on_memory() local 1037 ret = kasan_add_zero_shadow(__va(PFN_PHYS(pfn)), PFN_PHYS(nr_pages)); in mhp_init_memmap_on_memory() 1038 if (ret) in mhp_init_memmap_on_memory() 1039 return ret; in mhp_init_memmap_on_memory() 1051 return ret; in mhp_init_memmap_on_memory() 1080 int ret; in online_pages() local [all …]
|
D | util.c | 487 int ret = 0; in __account_locked_vm() local 496 ret = -ENOMEM; in __account_locked_vm() 498 if (!ret) in __account_locked_vm() 508 ret ? " - exceeded" : ""); in __account_locked_vm() 510 return ret; in __account_locked_vm() 528 int ret; in account_locked_vm() local 534 ret = __account_locked_vm(mm, pages, inc, current, in account_locked_vm() 538 return ret; in account_locked_vm() 546 unsigned long ret; in vm_mmap_pgoff() local 551 ret = security_mmap_file(file, prot, flag); in vm_mmap_pgoff() [all …]
|
D | gup.c | 506 int ret; in follow_page_pte() local 577 ret = follow_pfn_pte(vma, address, ptep, flags); in follow_page_pte() 578 page = ERR_PTR(ret); in follow_page_pte() 594 ret = arch_make_page_accessible(page); in follow_page_pte() 595 if (ret) { in follow_page_pte() 597 page = ERR_PTR(ret); in follow_page_pte() 728 int ret; in follow_pmd_mask() local 732 ret = 0; in follow_pmd_mask() 735 ret = -EBUSY; in follow_pmd_mask() 739 ret = pte_alloc(mm, pmd) ? -ENOMEM : 0; in follow_pmd_mask() [all …]
|
D | memory.c | 1073 int progress, ret = 0; in copy_pte_range() local 1084 ret = -ENOMEM; in copy_pte_range() 1110 ret = copy_nonpresent_pte(dst_mm, src_mm, in copy_pte_range() 1114 if (ret == -EIO) { in copy_pte_range() 1117 } else if (ret == -EBUSY) { in copy_pte_range() 1119 } else if (!ret) { in copy_pte_range() 1128 WARN_ON_ONCE(ret != -ENOENT); in copy_pte_range() 1131 ret = copy_present_pte(dst_vma, src_vma, dst_pte, src_pte, in copy_pte_range() 1137 if (unlikely(ret == -EAGAIN)) in copy_pte_range() 1159 if (ret == -EIO) { in copy_pte_range() [all …]
|
D | mmu_notifier.c | 273 bool ret; in mn_itree_release() local 279 ret = interval_sub->ops->invalidate(interval_sub, &range, in mn_itree_release() 281 WARN_ON(!ret); in mn_itree_release() 454 bool ret; in mn_itree_invalidate() local 456 ret = interval_sub->ops->invalidate(interval_sub, range, in mn_itree_invalidate() 458 if (!ret) { in mn_itree_invalidate() 480 int ret = 0; in mn_hlist_invalidate_range_start() local 511 ret = _ret; in mn_hlist_invalidate_range_start() 516 if (ret) { in mn_hlist_invalidate_range_start() 533 return ret; in mn_hlist_invalidate_range_start() [all …]
|
D | page-writeback.c | 509 int ret; in dirty_background_ratio_handler() local 511 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler() 512 if (ret == 0 && write) in dirty_background_ratio_handler() 514 return ret; in dirty_background_ratio_handler() 520 int ret; in dirty_background_bytes_handler() local 522 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler() 523 if (ret == 0 && write) in dirty_background_bytes_handler() 525 return ret; in dirty_background_bytes_handler() 532 int ret; in dirty_ratio_handler() local 534 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler() [all …]
|
D | huge_memory.c | 196 ssize_t ret = count; in enabled_store() local 208 ret = -EINVAL; in enabled_store() 210 if (ret > 0) { in enabled_store() 213 ret = err; in enabled_store() 215 return ret; in enabled_store() 234 int ret; in single_hugepage_flag_store() local 236 ret = kstrtoul(buf, 10, &value); in single_hugepage_flag_store() 237 if (ret < 0) in single_hugepage_flag_store() 238 return ret; in single_hugepage_flag_store() 465 int ret = 0; in setup_transparent_hugepage() local [all …]
|
D | userfaultfd.c | 62 int ret; in mfill_atomic_install_pte() local 90 ret = -EFAULT; in mfill_atomic_install_pte() 95 ret = -EEXIST; in mfill_atomic_install_pte() 117 ret = 0; in mfill_atomic_install_pte() 120 return ret; in mfill_atomic_install_pte() 132 int ret; in mcopy_atomic_pte() local 136 ret = -ENOMEM; in mcopy_atomic_pte() 142 ret = copy_from_user(page_kaddr, in mcopy_atomic_pte() 148 if (unlikely(ret)) { in mcopy_atomic_pte() 149 ret = -ENOENT; in mcopy_atomic_pte() [all …]
|
D | nommu.c | 169 void *ret; in __vmalloc_user_flags() local 171 ret = __vmalloc(size, flags); in __vmalloc_user_flags() 172 if (ret) { in __vmalloc_user_flags() 176 vma = find_vma(current->mm, (unsigned long)ret); in __vmalloc_user_flags() 182 return ret; in __vmalloc_user_flags() 411 int ret; in mmap_init() local 413 ret = percpu_counter_init(&vm_committed_as, 0, GFP_KERNEL); in mmap_init() 414 VM_BUG_ON(ret); in mmap_init() 762 int ret; in validate_mmap_request() local 898 ret = security_mmap_addr(addr); in validate_mmap_request() [all …]
|
D | zswap.c | 590 int ret; in zswap_pool_create() local 625 ret = cpuhp_state_add_instance(CPUHP_MM_ZSWP_POOL_PREPARE, in zswap_pool_create() 627 if (ret) in zswap_pool_create() 758 int ret; in __zswap_param_set() local 807 ret = param_set_charp(s, kp); in __zswap_param_set() 809 ret = -EINVAL; in __zswap_param_set() 813 if (!ret) { in __zswap_param_set() 836 ret = param_set_charp(s, kp); in __zswap_param_set() 845 return ret; in __zswap_param_set() 940 int ret; in zswap_writeback_entry() local [all …]
|
D | frontswap.c | 251 int ret = -1; in __frontswap_store() local 276 ret = ops->store(type, offset, page); in __frontswap_store() 277 if (!ret) /* successful store */ in __frontswap_store() 280 if (ret == 0) { in __frontswap_store() 288 ret = -1; in __frontswap_store() 289 return ret; in __frontswap_store() 300 int ret = -1; in __frontswap_load() local 316 ret = ops->load(type, offset, page); in __frontswap_load() 317 if (!ret) /* successful load */ in __frontswap_load() 320 if (ret == 0) { in __frontswap_load() [all …]
|
D | page_pinner.c | 193 int ret; in print_page_pinner() local 204 ret = snprintf(kbuf, count, "At least, pinned for %llu us\n", in print_page_pinner() 210 ret = snprintf(kbuf, count, in print_page_pinner() 217 if (ret >= count) in print_page_pinner() 221 ret += snprintf(kbuf + ret, count - ret, in print_page_pinner() 229 if (ret >= count) in print_page_pinner() 233 ret += stack_trace_snprint(kbuf + ret, count - ret, entries, in print_page_pinner() 235 if (ret >= count) in print_page_pinner() 238 ret += snprintf(kbuf + ret, count - ret, "\n"); in print_page_pinner() 239 if (ret >= count) in print_page_pinner() [all …]
|
D | oom_kill.c | 97 bool ret = false; in oom_cpuset_eligible() local 112 ret = mempolicy_in_oom_domain(tsk, mask); in oom_cpuset_eligible() 118 ret = cpuset_mems_allowed_intersects(current, tsk); in oom_cpuset_eligible() 120 if (ret) in oom_cpuset_eligible() 125 return ret; in oom_cpuset_eligible() 522 bool ret = true; in __oom_reap_task_mm() local 556 ret = false; in __oom_reap_task_mm() 565 return ret; in __oom_reap_task_mm() 576 bool ret = true; in oom_reap_task_mm() local 597 ret = __oom_reap_task_mm(mm); in oom_reap_task_mm() [all …]
|
D | cma.c | 251 int ret = 0; in cma_declare_contiguous_nid() local 283 ret = -EINVAL; in cma_declare_contiguous_nid() 304 ret = -EINVAL; in cma_declare_contiguous_nid() 319 ret = -EINVAL; in cma_declare_contiguous_nid() 329 ret = -EBUSY; in cma_declare_contiguous_nid() 368 ret = -ENOMEM; in cma_declare_contiguous_nid() 381 ret = cma_init_reserved_mem(base, size, order_per_bit, name, res_cma); in cma_declare_contiguous_nid() 382 if (ret) in cma_declare_contiguous_nid() 393 return ret; in cma_declare_contiguous_nid() 444 int ret = -ENOMEM; in cma_alloc() local [all …]
|
D | mremap.c | 845 unsigned long ret = -EINVAL; in mremap_to() local 877 ret = do_munmap(mm, new_addr, new_len, uf_unmap_early); in mremap_to() 878 if (ret) in mremap_to() 883 ret = do_munmap(mm, addr+new_len, old_len - new_len, uf_unmap); in mremap_to() 884 if (ret && old_len != new_len) in mremap_to() 891 ret = PTR_ERR(vma); in mremap_to() 898 ret = -ENOMEM; in mremap_to() 908 ret = get_unmapped_area(vma->vm_file, new_addr, new_len, vma->vm_pgoff + in mremap_to() 911 if (IS_ERR_VALUE(ret)) in mremap_to() 916 new_addr = ret; in mremap_to() [all …]
|
D | secretmem.c | 58 vm_fault_t ret; in secretmem_fault() local 71 ret = VM_FAULT_OOM; in secretmem_fault() 78 ret = vmf_error(err); in secretmem_fault() 95 ret = vmf_error(err); in secretmem_fault() 104 ret = VM_FAULT_LOCKED; in secretmem_fault() 108 return ret; in secretmem_fault() 178 int ret; in secretmem_setattr() local 183 ret = -EINVAL; in secretmem_setattr() 185 ret = simple_setattr(mnt_userns, dentry, iattr); in secretmem_setattr() 189 return ret; in secretmem_setattr() [all …]
|
D | page_owner.c | 369 int ret, pageblock_mt, page_mt; in print_page_owner() local 379 ret = snprintf(kbuf, count, in print_page_owner() 385 if (ret >= count) in print_page_owner() 391 ret += snprintf(kbuf + ret, count - ret, in print_page_owner() 399 if (ret >= count) in print_page_owner() 403 ret += stack_trace_snprint(kbuf + ret, count - ret, entries, nr_entries, 0); in print_page_owner() 404 if (ret >= count) in print_page_owner() 408 ret += snprintf(kbuf + ret, count - ret, in print_page_owner() 411 if (ret >= count) in print_page_owner() 415 ret += snprintf(kbuf + ret, count - ret, "\n"); in print_page_owner() [all …]
|
D | sparse-vmemmap.c | 125 int ret; in vmemmap_pmd_range() local 127 ret = split_vmemmap_huge_pmd(pmd, addr & PMD_MASK, walk); in vmemmap_pmd_range() 128 if (ret) in vmemmap_pmd_range() 129 return ret; in vmemmap_pmd_range() 147 int ret; in vmemmap_pud_range() local 150 ret = vmemmap_pmd_range(pud, addr, next, walk); in vmemmap_pud_range() 151 if (ret) in vmemmap_pud_range() 152 return ret; in vmemmap_pud_range() 167 int ret; in vmemmap_p4d_range() local 170 ret = vmemmap_pud_range(p4d, addr, next, walk); in vmemmap_p4d_range() [all …]
|
D | vmpressure.c | 162 bool ret = false; in vmpressure_event() local 173 ret = true; in vmpressure_event() 177 return ret; in vmpressure_event() 383 int ret = 0; in vmpressure_register_event() local 391 ret = match_string(vmpressure_str_levels, VMPRESSURE_NUM_LEVELS, token); in vmpressure_register_event() 392 if (ret < 0) in vmpressure_register_event() 394 level = ret; in vmpressure_register_event() 399 ret = match_string(vmpressure_str_modes, VMPRESSURE_NUM_MODES, token); in vmpressure_register_event() 400 if (ret < 0) in vmpressure_register_event() 402 mode = ret; in vmpressure_register_event() [all …]
|
D | compaction.c | 836 int ret = 0; in isolate_migratepages_block() local 905 ret = -EINTR; in isolate_migratepages_block() 933 ret = isolate_or_dissolve_huge_page(page, &cc->migratepages); in isolate_migratepages_block() 939 if (ret < 0) { in isolate_migratepages_block() 941 if (ret == -EBUSY) in isolate_migratepages_block() 942 ret = 0; in isolate_migratepages_block() 1162 if (!skip_on_failure && ret != -ENOMEM) in isolate_migratepages_block() 1190 if (ret == -ENOMEM) in isolate_migratepages_block() 1235 return ret; in isolate_migratepages_block() 1253 int ret = 0; in isolate_migratepages_range() local [all …]
|
/mm/damon/ |
D | dbgfs.c | 29 ssize_t ret; in user_input_str() local 39 ret = simple_write_to_buffer(kbuf, count + 1, ppos, buf, count); in user_input_str() 40 if (ret != count) { in user_input_str() 44 kbuf[ret] = '\0'; in user_input_str() 54 int ret; in dbgfs_attrs_read() local 57 ret = scnprintf(kbuf, ARRAY_SIZE(kbuf), "%lu %lu %lu %lu %lu\n", in dbgfs_attrs_read() 63 return simple_read_from_buffer(buf, count, ppos, kbuf, ret); in dbgfs_attrs_read() 72 ssize_t ret; in dbgfs_attrs_write() local 80 ret = -EINVAL; in dbgfs_attrs_write() 86 ret = -EBUSY; in dbgfs_attrs_write() [all …]
|