Home
last modified time | relevance | path

Searched refs:ret (Results 1 – 25 of 69) sorted by relevance

123

/mm/
Dmemory_hotplug.c256 int ret; in __add_section() local
262 ret = sparse_add_one_section(NODE_DATA(nid), phys_start_pfn); in __add_section()
263 if (ret < 0) in __add_section()
264 return ret; in __add_section()
508 int ret = -EINVAL; in __remove_section() local
511 return ret; in __remove_section()
513 ret = unregister_memory_section(ms); in __remove_section()
514 if (ret) in __remove_section()
515 return ret; in __remove_section()
541 int sections_to_remove, ret = 0; in __remove_pages() local
[all …]
Dpage_io.c161 int ret; in generic_swapfile_activate() local
217 ret = add_swap_extent(sis, page_no, 1, first_block); in generic_swapfile_activate()
218 if (ret < 0) in generic_swapfile_activate()
220 nr_extents += ret; in generic_swapfile_activate()
226 ret = nr_extents; in generic_swapfile_activate()
234 return ret; in generic_swapfile_activate()
237 ret = -EINVAL; in generic_swapfile_activate()
247 int ret = 0; in swap_writepage() local
259 ret = __swap_writepage(page, wbc, end_swap_bio_write); in swap_writepage()
261 return ret; in swap_writepage()
[all …]
Dbacking-dev.c147 ssize_t ret; in read_ahead_kb_store() local
149 ret = kstrtoul(buf, 10, &read_ahead_kb); in read_ahead_kb_store()
150 if (ret < 0) in read_ahead_kb_store()
151 return ret; in read_ahead_kb_store()
177 ssize_t ret; in min_ratio_store() local
179 ret = kstrtouint(buf, 10, &ratio); in min_ratio_store()
180 if (ret < 0) in min_ratio_store()
181 return ret; in min_ratio_store()
183 ret = bdi_set_min_ratio(bdi, ratio); in min_ratio_store()
184 if (!ret) in min_ratio_store()
[all …]
Dgup.c134 int ret; in follow_page_pte() local
136 ret = follow_pfn_pte(vma, address, ptep, flags); in follow_page_pte()
137 page = ERR_PTR(ret); in follow_page_pte()
143 int ret; in follow_page_pte() local
147 ret = split_huge_page(page); in follow_page_pte()
150 if (ret) in follow_page_pte()
151 return ERR_PTR(ret); in follow_page_pte()
277 int ret; in follow_pmd_mask() local
281 ret = 0; in follow_pmd_mask()
284 ret = -EBUSY; in follow_pmd_mask()
[all …]
Dhmm.c314 int ret; in hmm_vma_walk_hole() local
316 ret = hmm_vma_do_fault(walk, addr, &pfns[i]); in hmm_vma_walk_hole()
317 if (ret != -EAGAIN) in hmm_vma_walk_hole()
318 return ret; in hmm_vma_walk_hole()
339 int ret; in hmm_vma_walk_clear() local
341 ret = hmm_vma_do_fault(walk, addr, &pfns[i]); in hmm_vma_walk_clear()
342 if (ret != -EAGAIN) in hmm_vma_walk_clear()
343 return ret; in hmm_vma_walk_clear()
675 int ret; in hmm_vma_fault() local
723 ret = walk_page_range(start, end, &mm_walk); in hmm_vma_fault()
[all …]
Dmemory-failure.c185 int ret; in kill_proc() local
199 ret = force_sig_info(SIGBUS, &si, current); in kill_proc()
208 ret = send_sig_info(SIGBUS, &si, t); /* synchronous? */ in kill_proc()
210 if (ret < 0) in kill_proc()
212 t->comm, t->pid, ret); in kill_proc()
213 return ret; in kill_proc()
562 int ret = MF_FAILED; in truncate_error_page() local
575 ret = MF_RECOVERED; in truncate_error_page()
583 ret = MF_RECOVERED; in truncate_error_page()
589 return ret; in truncate_error_page()
[all …]
Dzswap.c503 int ret; in zswap_pool_create() local
537 ret = cpuhp_state_add_instance(CPUHP_MM_ZSWP_POOL_PREPARE, in zswap_pool_create()
539 if (ret) in zswap_pool_create()
665 int ret; in __zswap_param_set() local
714 ret = param_set_charp(s, kp); in __zswap_param_set()
716 ret = -EINVAL; in __zswap_param_set()
720 if (!ret) { in __zswap_param_set()
743 ret = param_set_charp(s, kp); in __zswap_param_set()
752 return ret; in __zswap_param_set()
845 int ret; in zswap_writeback_entry() local
[all …]
Dframe_vector.c39 int ret = 0; in get_vaddr_frames() local
53 ret = -EFAULT; in get_vaddr_frames()
66 ret = -EOPNOTSUPP; in get_vaddr_frames()
73 ret = get_user_pages_locked(start, nr_frames, in get_vaddr_frames()
83 while (ret < nr_frames && start + PAGE_SIZE <= vma->vm_end) { in get_vaddr_frames()
84 err = follow_pfn(vma, start, &nums[ret]); in get_vaddr_frames()
86 if (ret == 0) in get_vaddr_frames()
87 ret = err; in get_vaddr_frames()
91 ret++; in get_vaddr_frames()
97 if (ret >= nr_frames || start < vma->vm_end) in get_vaddr_frames()
[all …]
Dmemory.c1230 int ret; in copy_page_range() local
1250 ret = track_pfn_copy(vma); in copy_page_range()
1251 if (ret) in copy_page_range()
1252 return ret; in copy_page_range()
1268 ret = 0; in copy_page_range()
1277 ret = -ENOMEM; in copy_page_range()
1284 return ret; in copy_page_range()
1887 int ret; in vm_insert_pfn_prot() local
1908 ret = insert_pfn(vma, addr, __pfn_to_pfn_t(pfn, PFN_DEV), pgprot, in vm_insert_pfn_prot()
1911 return ret; in vm_insert_pfn_prot()
[all …]
Dpage-writeback.c518 int ret; in dirty_background_ratio_handler() local
520 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler()
521 if (ret == 0 && write) in dirty_background_ratio_handler()
523 return ret; in dirty_background_ratio_handler()
530 int ret; in dirty_background_bytes_handler() local
532 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler()
533 if (ret == 0 && write) in dirty_background_bytes_handler()
535 return ret; in dirty_background_bytes_handler()
543 int ret; in dirty_ratio_handler() local
545 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler()
[all …]
Dutil.c324 unsigned long ret; in vm_mmap_pgoff() local
329 ret = security_mmap_file(file, prot, flag); in vm_mmap_pgoff()
330 if (!ret) { in vm_mmap_pgoff()
333 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, in vm_mmap_pgoff()
338 mm_populate(ret, populate); in vm_mmap_pgoff()
340 return ret; in vm_mmap_pgoff()
375 void *ret; in kvmalloc_node() local
397 ret = kmalloc_node(size, kmalloc_flags, node); in kvmalloc_node()
403 if (ret || size <= PAGE_SIZE) in kvmalloc_node()
404 return ret; in kvmalloc_node()
[all …]
Dnommu.c185 long ret; in __get_user_pages_unlocked() local
187 ret = __get_user_pages(tsk, mm, start, nr_pages, gup_flags, pages, in __get_user_pages_unlocked()
190 return ret; in __get_user_pages_unlocked()
247 void *ret; in vmalloc_user() local
249 ret = __vmalloc(size, GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL); in vmalloc_user()
250 if (ret) { in vmalloc_user()
254 vma = find_vma(current->mm, (unsigned long)ret); in vmalloc_user()
260 return ret; in vmalloc_user()
533 int ret; in mmap_init() local
535 ret = percpu_counter_init(&vm_committed_as, 0, GFP_KERNEL); in mmap_init()
[all …]
Dhuge_memory.c164 ssize_t ret = count; in enabled_store() local
176 ret = -EINVAL; in enabled_store()
178 if (ret > 0) { in enabled_store()
181 ret = err; in enabled_store()
183 return ret; in enabled_store()
202 int ret; in single_hugepage_flag_store() local
204 ret = kstrtoul(buf, 10, &value); in single_hugepage_flag_store()
205 if (ret < 0) in single_hugepage_flag_store()
206 return ret; in single_hugepage_flag_store()
441 int ret = 0; in setup_transparent_hugepage() local
[all …]
Dmremap.c447 unsigned long ret = -EINVAL; in mremap_to() local
461 ret = do_munmap(mm, new_addr, new_len, uf_unmap_early); in mremap_to()
462 if (ret) in mremap_to()
466 ret = do_munmap(mm, addr+new_len, old_len - new_len, uf_unmap); in mremap_to()
467 if (ret && old_len != new_len) in mremap_to()
474 ret = PTR_ERR(vma); in mremap_to()
482 ret = get_unmapped_area(vma->vm_file, new_addr, new_len, vma->vm_pgoff + in mremap_to()
485 if (offset_in_page(ret)) in mremap_to()
488 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, uf, in mremap_to()
490 if (!(offset_in_page(ret))) in mremap_to()
[all …]
Dcma.c155 int ret = cma_activate_area(&cma_areas[i]); in cma_init_reserved_areas() local
157 if (ret) in cma_init_reserved_areas()
158 return ret; in cma_init_reserved_areas()
249 int ret = 0; in cma_declare_contiguous() local
281 ret = -EINVAL; in cma_declare_contiguous()
302 ret = -EINVAL; in cma_declare_contiguous()
317 ret = -EINVAL; in cma_declare_contiguous()
327 ret = -EBUSY; in cma_declare_contiguous()
351 ret = -ENOMEM; in cma_declare_contiguous()
364 ret = cma_init_reserved_mem(base, size, order_per_bit, name, res_cma); in cma_declare_contiguous()
[all …]
Dfrontswap.c248 int ret = -1; in __frontswap_store() local
273 ret = ops->store(type, offset, page); in __frontswap_store()
274 if (!ret) /* successful store */ in __frontswap_store()
277 if (ret == 0) { in __frontswap_store()
285 ret = -1; in __frontswap_store()
286 return ret; in __frontswap_store()
297 int ret = -1; in __frontswap_load() local
313 ret = ops->load(type, offset, page); in __frontswap_load()
314 if (!ret) /* successful load */ in __frontswap_load()
317 if (ret == 0) { in __frontswap_load()
[all …]
Dpage_owner.c344 int ret; in print_page_owner() local
359 ret = snprintf(kbuf, count, in print_page_owner()
364 if (ret >= count) in print_page_owner()
370 ret += snprintf(kbuf + ret, count - ret, in print_page_owner()
378 if (ret >= count) in print_page_owner()
382 ret += snprint_stack_trace(kbuf + ret, count - ret, &trace, 0); in print_page_owner()
383 if (ret >= count) in print_page_owner()
387 ret += snprintf(kbuf + ret, count - ret, in print_page_owner()
390 if (ret >= count) in print_page_owner()
394 ret += snprintf(kbuf + ret, count - ret, "\n"); in print_page_owner()
[all …]
Dmaccess.c28 long ret; in __probe_kernel_read() local
33 ret = __copy_from_user_inatomic(dst, in __probe_kernel_read()
38 return ret ? -EFAULT : 0; in __probe_kernel_read()
56 long ret; in __probe_kernel_write() local
61 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); in __probe_kernel_write()
65 return ret ? -EFAULT : 0; in __probe_kernel_write()
90 long ret; in strncpy_from_unsafe() local
99 ret = __get_user(*dst++, (const char __user __force *)src++); in strncpy_from_unsafe()
100 } while (dst[-1] && ret == 0 && src - unsafe_addr < count); in strncpy_from_unsafe()
106 return ret ? -EFAULT : src - unsafe_addr; in strncpy_from_unsafe()
Dhugetlb.c135 long ret = delta; in hugepage_subpool_get_pages() local
138 return ret; in hugepage_subpool_get_pages()
146 ret = -ENOMEM; in hugepage_subpool_get_pages()
158 ret = delta - spool->rsv_hpages; in hugepage_subpool_get_pages()
161 ret = 0; /* reserves already accounted for */ in hugepage_subpool_get_pages()
168 return ret; in hugepage_subpool_get_pages()
180 long ret = delta; in hugepage_subpool_put_pages() local
193 ret = 0; in hugepage_subpool_put_pages()
195 ret = spool->rsv_hpages + delta - spool->min_hpages; in hugepage_subpool_put_pages()
208 return ret; in hugepage_subpool_put_pages()
[all …]
Dfilemap.c295 int ret = 0; in filemap_check_errors() local
299 ret = -ENOSPC; in filemap_check_errors()
302 ret = -EIO; in filemap_check_errors()
303 return ret; in filemap_check_errors()
335 int ret; in __filemap_fdatawrite_range() local
348 ret = do_writepages(mapping, &wbc); in __filemap_fdatawrite_range()
350 return ret; in __filemap_fdatawrite_range()
827 int ret; in add_to_page_cache_lru() local
830 ret = __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_lru()
832 if (unlikely(ret)) in add_to_page_cache_lru()
[all …]
Dreadahead.c87 int ret = 0; in read_cache_pages() local
99 ret = filler(data, page); in read_cache_pages()
100 if (unlikely(ret)) { in read_cache_pages()
106 return ret; in read_cache_pages()
116 int ret; in read_pages() local
121 ret = mapping->a_ops->readpages(filp, mapping, pages, nr_pages); in read_pages()
134 ret = 0; in read_pages()
139 return ret; in read_pages()
159 int ret = 0; in __do_page_cache_readahead() local
190 ret++; in __do_page_cache_readahead()
[all …]
Dhugetlb_cgroup.c79 int ret; in hugetlb_cgroup_init() local
87 ret = page_counter_limit(counter, limit); in hugetlb_cgroup_init()
88 VM_BUG_ON(ret); in hugetlb_cgroup_init()
184 int ret = 0; in hugetlb_cgroup_charge_cgroup() local
206 ret = -ENOMEM; in hugetlb_cgroup_charge_cgroup()
210 return ret; in hugetlb_cgroup_charge_cgroup()
291 int ret, idx; in hugetlb_cgroup_write() local
299 ret = page_counter_memparse(buf, "-1", &nr_pages); in hugetlb_cgroup_write()
300 if (ret) in hugetlb_cgroup_write()
301 return ret; in hugetlb_cgroup_write()
[all …]
Dworkingset.c451 int ret; in shadow_lru_isolate() local
471 ret = LRU_RETRY; in shadow_lru_isolate()
509 ret = LRU_REMOVED_RETRY; in shadow_lru_isolate()
515 return ret; in shadow_lru_isolate()
521 unsigned long ret; in scan_shadow_nodes() local
525 ret = list_lru_shrink_walk(&shadow_nodes, sc, shadow_lru_isolate, NULL); in scan_shadow_nodes()
527 return ret; in scan_shadow_nodes()
547 int ret; in workingset_init() local
564 ret = __list_lru_init(&shadow_nodes, true, &shadow_nodes_key); in workingset_init()
565 if (ret) in workingset_init()
[all …]
Dmemcontrol.c945 int ret = 0; in mem_cgroup_scan_tasks() local
954 while (!ret && (task = css_task_iter_next(&it))) in mem_cgroup_scan_tasks()
955 ret = fn(task, arg); in mem_cgroup_scan_tasks()
957 if (ret) { in mem_cgroup_scan_tasks()
962 return ret; in mem_cgroup_scan_tasks()
1049 bool ret; in task_in_mem_cgroup() local
1066 ret = mem_cgroup_is_descendant(task_memcg, memcg); in task_in_mem_cgroup()
1068 return ret; in task_in_mem_cgroup()
1112 bool ret = false; in mem_cgroup_under_move() local
1123 ret = mem_cgroup_is_descendant(from, memcg) || in mem_cgroup_under_move()
[all …]
Dcompaction.c1384 int ret; in compact_finished() local
1386 ret = __compact_finished(zone, cc); in compact_finished()
1387 trace_mm_compaction_finished(zone, cc->order, ret); in compact_finished()
1388 if (ret == COMPACT_NO_SUITABLE_PAGE) in compact_finished()
1389 ret = COMPACT_CONTINUE; in compact_finished()
1391 return ret; in compact_finished()
1448 enum compact_result ret; in compaction_suitable() local
1451 ret = __compaction_suitable(zone, order, alloc_flags, classzone_idx, in compaction_suitable()
1469 if (ret == COMPACT_CONTINUE && (order > PAGE_ALLOC_COSTLY_ORDER)) { in compaction_suitable()
1472 ret = COMPACT_NOT_SUITABLE_ZONE; in compaction_suitable()
[all …]

123