Home
last modified time | relevance | path

Searched refs:ret (Results 1 – 25 of 86) sorted by relevance

1234

/mm/
Dmaccess.c123 long ret; in copy_from_kernel_nofault() local
131 ret = __copy_from_user_inatomic(dst, (__force const void __user *)src, in copy_from_kernel_nofault()
136 if (ret) in copy_from_kernel_nofault()
153 long ret; in copy_to_kernel_nofault() local
158 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); in copy_to_kernel_nofault()
162 if (ret) in copy_to_kernel_nofault()
190 long ret; in strncpy_from_kernel_nofault() local
201 ret = __get_user(*dst++, (const char __user __force *)src++); in strncpy_from_kernel_nofault()
202 } while (dst[-1] && ret == 0 && src - unsafe_addr < count); in strncpy_from_kernel_nofault()
208 return ret ? -EFAULT : src - unsafe_addr; in strncpy_from_kernel_nofault()
[all …]
Dmemory-failure.c72 int ret; in __page_handle_poison() local
75 ret = dissolve_free_huge_page(page); in __page_handle_poison()
76 if (!ret) in __page_handle_poison()
77 ret = take_page_off_buddy(page); in __page_handle_poison()
80 return ret > 0; in __page_handle_poison()
255 int ret = 0; in kill_proc() local
262 ret = force_sig_mceerr(BUS_MCEERR_AR, in kill_proc()
266 ret = send_sig_mceerr(BUS_MCEERR_AO, (void __user *)tk->addr, in kill_proc()
275 ret = send_sig_mceerr(BUS_MCEERR_AO, (void __user *)tk->addr, in kill_proc()
278 if (ret < 0) in kill_proc()
[all …]
Dpage_io.c92 int ret; in generic_swapfile_activate() local
112 ret = bmap(inode, &first_block); in generic_swapfile_activate()
113 if (ret || !first_block) in generic_swapfile_activate()
129 ret = bmap(inode, &block); in generic_swapfile_activate()
130 if (ret || !block) in generic_swapfile_activate()
151 ret = add_swap_extent(sis, page_no, 1, first_block); in generic_swapfile_activate()
152 if (ret < 0) in generic_swapfile_activate()
154 nr_extents += ret; in generic_swapfile_activate()
160 ret = nr_extents; in generic_swapfile_activate()
168 return ret; in generic_swapfile_activate()
[all …]
Dbacking-dev.c139 ssize_t ret; in read_ahead_kb_store() local
141 ret = kstrtoul(buf, 10, &read_ahead_kb); in read_ahead_kb_store()
142 if (ret < 0) in read_ahead_kb_store()
143 return ret; in read_ahead_kb_store()
167 ssize_t ret; in min_ratio_store() local
169 ret = kstrtouint(buf, 10, &ratio); in min_ratio_store()
170 if (ret < 0) in min_ratio_store()
171 return ret; in min_ratio_store()
173 ret = bdi_set_min_ratio(bdi, ratio); in min_ratio_store()
174 if (!ret) in min_ratio_store()
[all …]
Dmemory_hotplug.c67 int ret = sysfs_match_string(online_policy_to_str, val); in set_online_policy() local
69 if (ret < 0) in set_online_policy()
70 return ret; in set_online_policy()
71 *((int *)kp->arg) = ret; in set_online_policy()
1035 int ret; in mhp_init_memmap_on_memory() local
1037 ret = kasan_add_zero_shadow(__va(PFN_PHYS(pfn)), PFN_PHYS(nr_pages)); in mhp_init_memmap_on_memory()
1038 if (ret) in mhp_init_memmap_on_memory()
1039 return ret; in mhp_init_memmap_on_memory()
1051 return ret; in mhp_init_memmap_on_memory()
1080 int ret; in online_pages() local
[all …]
Dutil.c487 int ret = 0; in __account_locked_vm() local
496 ret = -ENOMEM; in __account_locked_vm()
498 if (!ret) in __account_locked_vm()
508 ret ? " - exceeded" : ""); in __account_locked_vm()
510 return ret; in __account_locked_vm()
528 int ret; in account_locked_vm() local
534 ret = __account_locked_vm(mm, pages, inc, current, in account_locked_vm()
538 return ret; in account_locked_vm()
546 unsigned long ret; in vm_mmap_pgoff() local
551 ret = security_mmap_file(file, prot, flag); in vm_mmap_pgoff()
[all …]
Dgup.c506 int ret; in follow_page_pte() local
577 ret = follow_pfn_pte(vma, address, ptep, flags); in follow_page_pte()
578 page = ERR_PTR(ret); in follow_page_pte()
594 ret = arch_make_page_accessible(page); in follow_page_pte()
595 if (ret) { in follow_page_pte()
597 page = ERR_PTR(ret); in follow_page_pte()
728 int ret; in follow_pmd_mask() local
732 ret = 0; in follow_pmd_mask()
735 ret = -EBUSY; in follow_pmd_mask()
739 ret = pte_alloc(mm, pmd) ? -ENOMEM : 0; in follow_pmd_mask()
[all …]
Dmemory.c1073 int progress, ret = 0; in copy_pte_range() local
1084 ret = -ENOMEM; in copy_pte_range()
1110 ret = copy_nonpresent_pte(dst_mm, src_mm, in copy_pte_range()
1114 if (ret == -EIO) { in copy_pte_range()
1117 } else if (ret == -EBUSY) { in copy_pte_range()
1119 } else if (!ret) { in copy_pte_range()
1128 WARN_ON_ONCE(ret != -ENOENT); in copy_pte_range()
1131 ret = copy_present_pte(dst_vma, src_vma, dst_pte, src_pte, in copy_pte_range()
1137 if (unlikely(ret == -EAGAIN)) in copy_pte_range()
1159 if (ret == -EIO) { in copy_pte_range()
[all …]
Dmmu_notifier.c273 bool ret; in mn_itree_release() local
279 ret = interval_sub->ops->invalidate(interval_sub, &range, in mn_itree_release()
281 WARN_ON(!ret); in mn_itree_release()
454 bool ret; in mn_itree_invalidate() local
456 ret = interval_sub->ops->invalidate(interval_sub, range, in mn_itree_invalidate()
458 if (!ret) { in mn_itree_invalidate()
480 int ret = 0; in mn_hlist_invalidate_range_start() local
511 ret = _ret; in mn_hlist_invalidate_range_start()
516 if (ret) { in mn_hlist_invalidate_range_start()
533 return ret; in mn_hlist_invalidate_range_start()
[all …]
Dpage-writeback.c509 int ret; in dirty_background_ratio_handler() local
511 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler()
512 if (ret == 0 && write) in dirty_background_ratio_handler()
514 return ret; in dirty_background_ratio_handler()
520 int ret; in dirty_background_bytes_handler() local
522 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler()
523 if (ret == 0 && write) in dirty_background_bytes_handler()
525 return ret; in dirty_background_bytes_handler()
532 int ret; in dirty_ratio_handler() local
534 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler()
[all …]
Dhuge_memory.c196 ssize_t ret = count; in enabled_store() local
208 ret = -EINVAL; in enabled_store()
210 if (ret > 0) { in enabled_store()
213 ret = err; in enabled_store()
215 return ret; in enabled_store()
234 int ret; in single_hugepage_flag_store() local
236 ret = kstrtoul(buf, 10, &value); in single_hugepage_flag_store()
237 if (ret < 0) in single_hugepage_flag_store()
238 return ret; in single_hugepage_flag_store()
465 int ret = 0; in setup_transparent_hugepage() local
[all …]
Duserfaultfd.c62 int ret; in mfill_atomic_install_pte() local
90 ret = -EFAULT; in mfill_atomic_install_pte()
95 ret = -EEXIST; in mfill_atomic_install_pte()
117 ret = 0; in mfill_atomic_install_pte()
120 return ret; in mfill_atomic_install_pte()
132 int ret; in mcopy_atomic_pte() local
136 ret = -ENOMEM; in mcopy_atomic_pte()
142 ret = copy_from_user(page_kaddr, in mcopy_atomic_pte()
148 if (unlikely(ret)) { in mcopy_atomic_pte()
149 ret = -ENOENT; in mcopy_atomic_pte()
[all …]
Dnommu.c169 void *ret; in __vmalloc_user_flags() local
171 ret = __vmalloc(size, flags); in __vmalloc_user_flags()
172 if (ret) { in __vmalloc_user_flags()
176 vma = find_vma(current->mm, (unsigned long)ret); in __vmalloc_user_flags()
182 return ret; in __vmalloc_user_flags()
411 int ret; in mmap_init() local
413 ret = percpu_counter_init(&vm_committed_as, 0, GFP_KERNEL); in mmap_init()
414 VM_BUG_ON(ret); in mmap_init()
762 int ret; in validate_mmap_request() local
898 ret = security_mmap_addr(addr); in validate_mmap_request()
[all …]
Dzswap.c590 int ret; in zswap_pool_create() local
625 ret = cpuhp_state_add_instance(CPUHP_MM_ZSWP_POOL_PREPARE, in zswap_pool_create()
627 if (ret) in zswap_pool_create()
758 int ret; in __zswap_param_set() local
807 ret = param_set_charp(s, kp); in __zswap_param_set()
809 ret = -EINVAL; in __zswap_param_set()
813 if (!ret) { in __zswap_param_set()
836 ret = param_set_charp(s, kp); in __zswap_param_set()
845 return ret; in __zswap_param_set()
940 int ret; in zswap_writeback_entry() local
[all …]
Dfrontswap.c251 int ret = -1; in __frontswap_store() local
276 ret = ops->store(type, offset, page); in __frontswap_store()
277 if (!ret) /* successful store */ in __frontswap_store()
280 if (ret == 0) { in __frontswap_store()
288 ret = -1; in __frontswap_store()
289 return ret; in __frontswap_store()
300 int ret = -1; in __frontswap_load() local
316 ret = ops->load(type, offset, page); in __frontswap_load()
317 if (!ret) /* successful load */ in __frontswap_load()
320 if (ret == 0) { in __frontswap_load()
[all …]
Dpage_pinner.c193 int ret; in print_page_pinner() local
204 ret = snprintf(kbuf, count, "At least, pinned for %llu us\n", in print_page_pinner()
210 ret = snprintf(kbuf, count, in print_page_pinner()
217 if (ret >= count) in print_page_pinner()
221 ret += snprintf(kbuf + ret, count - ret, in print_page_pinner()
229 if (ret >= count) in print_page_pinner()
233 ret += stack_trace_snprint(kbuf + ret, count - ret, entries, in print_page_pinner()
235 if (ret >= count) in print_page_pinner()
238 ret += snprintf(kbuf + ret, count - ret, "\n"); in print_page_pinner()
239 if (ret >= count) in print_page_pinner()
[all …]
Doom_kill.c97 bool ret = false; in oom_cpuset_eligible() local
112 ret = mempolicy_in_oom_domain(tsk, mask); in oom_cpuset_eligible()
118 ret = cpuset_mems_allowed_intersects(current, tsk); in oom_cpuset_eligible()
120 if (ret) in oom_cpuset_eligible()
125 return ret; in oom_cpuset_eligible()
522 bool ret = true; in __oom_reap_task_mm() local
556 ret = false; in __oom_reap_task_mm()
565 return ret; in __oom_reap_task_mm()
576 bool ret = true; in oom_reap_task_mm() local
597 ret = __oom_reap_task_mm(mm); in oom_reap_task_mm()
[all …]
Dcma.c251 int ret = 0; in cma_declare_contiguous_nid() local
283 ret = -EINVAL; in cma_declare_contiguous_nid()
304 ret = -EINVAL; in cma_declare_contiguous_nid()
319 ret = -EINVAL; in cma_declare_contiguous_nid()
329 ret = -EBUSY; in cma_declare_contiguous_nid()
368 ret = -ENOMEM; in cma_declare_contiguous_nid()
381 ret = cma_init_reserved_mem(base, size, order_per_bit, name, res_cma); in cma_declare_contiguous_nid()
382 if (ret) in cma_declare_contiguous_nid()
393 return ret; in cma_declare_contiguous_nid()
444 int ret = -ENOMEM; in cma_alloc() local
[all …]
Dmremap.c845 unsigned long ret = -EINVAL; in mremap_to() local
877 ret = do_munmap(mm, new_addr, new_len, uf_unmap_early); in mremap_to()
878 if (ret) in mremap_to()
883 ret = do_munmap(mm, addr+new_len, old_len - new_len, uf_unmap); in mremap_to()
884 if (ret && old_len != new_len) in mremap_to()
891 ret = PTR_ERR(vma); in mremap_to()
898 ret = -ENOMEM; in mremap_to()
908 ret = get_unmapped_area(vma->vm_file, new_addr, new_len, vma->vm_pgoff + in mremap_to()
911 if (IS_ERR_VALUE(ret)) in mremap_to()
916 new_addr = ret; in mremap_to()
[all …]
Dsecretmem.c58 vm_fault_t ret; in secretmem_fault() local
71 ret = VM_FAULT_OOM; in secretmem_fault()
78 ret = vmf_error(err); in secretmem_fault()
95 ret = vmf_error(err); in secretmem_fault()
104 ret = VM_FAULT_LOCKED; in secretmem_fault()
108 return ret; in secretmem_fault()
178 int ret; in secretmem_setattr() local
183 ret = -EINVAL; in secretmem_setattr()
185 ret = simple_setattr(mnt_userns, dentry, iattr); in secretmem_setattr()
189 return ret; in secretmem_setattr()
[all …]
Dpage_owner.c369 int ret, pageblock_mt, page_mt; in print_page_owner() local
379 ret = snprintf(kbuf, count, in print_page_owner()
385 if (ret >= count) in print_page_owner()
391 ret += snprintf(kbuf + ret, count - ret, in print_page_owner()
399 if (ret >= count) in print_page_owner()
403 ret += stack_trace_snprint(kbuf + ret, count - ret, entries, nr_entries, 0); in print_page_owner()
404 if (ret >= count) in print_page_owner()
408 ret += snprintf(kbuf + ret, count - ret, in print_page_owner()
411 if (ret >= count) in print_page_owner()
415 ret += snprintf(kbuf + ret, count - ret, "\n"); in print_page_owner()
[all …]
Dsparse-vmemmap.c125 int ret; in vmemmap_pmd_range() local
127 ret = split_vmemmap_huge_pmd(pmd, addr & PMD_MASK, walk); in vmemmap_pmd_range()
128 if (ret) in vmemmap_pmd_range()
129 return ret; in vmemmap_pmd_range()
147 int ret; in vmemmap_pud_range() local
150 ret = vmemmap_pmd_range(pud, addr, next, walk); in vmemmap_pud_range()
151 if (ret) in vmemmap_pud_range()
152 return ret; in vmemmap_pud_range()
167 int ret; in vmemmap_p4d_range() local
170 ret = vmemmap_pud_range(p4d, addr, next, walk); in vmemmap_p4d_range()
[all …]
Dvmpressure.c162 bool ret = false; in vmpressure_event() local
173 ret = true; in vmpressure_event()
177 return ret; in vmpressure_event()
383 int ret = 0; in vmpressure_register_event() local
391 ret = match_string(vmpressure_str_levels, VMPRESSURE_NUM_LEVELS, token); in vmpressure_register_event()
392 if (ret < 0) in vmpressure_register_event()
394 level = ret; in vmpressure_register_event()
399 ret = match_string(vmpressure_str_modes, VMPRESSURE_NUM_MODES, token); in vmpressure_register_event()
400 if (ret < 0) in vmpressure_register_event()
402 mode = ret; in vmpressure_register_event()
[all …]
Dcompaction.c836 int ret = 0; in isolate_migratepages_block() local
905 ret = -EINTR; in isolate_migratepages_block()
933 ret = isolate_or_dissolve_huge_page(page, &cc->migratepages); in isolate_migratepages_block()
939 if (ret < 0) { in isolate_migratepages_block()
941 if (ret == -EBUSY) in isolate_migratepages_block()
942 ret = 0; in isolate_migratepages_block()
1162 if (!skip_on_failure && ret != -ENOMEM) in isolate_migratepages_block()
1190 if (ret == -ENOMEM) in isolate_migratepages_block()
1235 return ret; in isolate_migratepages_block()
1253 int ret = 0; in isolate_migratepages_range() local
[all …]
/mm/damon/
Ddbgfs.c29 ssize_t ret; in user_input_str() local
39 ret = simple_write_to_buffer(kbuf, count + 1, ppos, buf, count); in user_input_str()
40 if (ret != count) { in user_input_str()
44 kbuf[ret] = '\0'; in user_input_str()
54 int ret; in dbgfs_attrs_read() local
57 ret = scnprintf(kbuf, ARRAY_SIZE(kbuf), "%lu %lu %lu %lu %lu\n", in dbgfs_attrs_read()
63 return simple_read_from_buffer(buf, count, ppos, kbuf, ret); in dbgfs_attrs_read()
72 ssize_t ret; in dbgfs_attrs_write() local
80 ret = -EINVAL; in dbgfs_attrs_write()
86 ret = -EBUSY; in dbgfs_attrs_write()
[all …]

1234