/mm/ |
D | memory_hotplug.c | 256 int ret; in __add_section() local 262 ret = sparse_add_one_section(NODE_DATA(nid), phys_start_pfn); in __add_section() 263 if (ret < 0) in __add_section() 264 return ret; in __add_section() 508 int ret = -EINVAL; in __remove_section() local 511 return ret; in __remove_section() 513 ret = unregister_memory_section(ms); in __remove_section() 514 if (ret) in __remove_section() 515 return ret; in __remove_section() 541 int sections_to_remove, ret = 0; in __remove_pages() local [all …]
|
D | page_io.c | 161 int ret; in generic_swapfile_activate() local 217 ret = add_swap_extent(sis, page_no, 1, first_block); in generic_swapfile_activate() 218 if (ret < 0) in generic_swapfile_activate() 220 nr_extents += ret; in generic_swapfile_activate() 226 ret = nr_extents; in generic_swapfile_activate() 234 return ret; in generic_swapfile_activate() 237 ret = -EINVAL; in generic_swapfile_activate() 247 int ret = 0; in swap_writepage() local 259 ret = __swap_writepage(page, wbc, end_swap_bio_write); in swap_writepage() 261 return ret; in swap_writepage() [all …]
|
D | backing-dev.c | 147 ssize_t ret; in read_ahead_kb_store() local 149 ret = kstrtoul(buf, 10, &read_ahead_kb); in read_ahead_kb_store() 150 if (ret < 0) in read_ahead_kb_store() 151 return ret; in read_ahead_kb_store() 177 ssize_t ret; in min_ratio_store() local 179 ret = kstrtouint(buf, 10, &ratio); in min_ratio_store() 180 if (ret < 0) in min_ratio_store() 181 return ret; in min_ratio_store() 183 ret = bdi_set_min_ratio(bdi, ratio); in min_ratio_store() 184 if (!ret) in min_ratio_store() [all …]
|
D | gup.c | 134 int ret; in follow_page_pte() local 136 ret = follow_pfn_pte(vma, address, ptep, flags); in follow_page_pte() 137 page = ERR_PTR(ret); in follow_page_pte() 143 int ret; in follow_page_pte() local 147 ret = split_huge_page(page); in follow_page_pte() 150 if (ret) in follow_page_pte() 151 return ERR_PTR(ret); in follow_page_pte() 277 int ret; in follow_pmd_mask() local 281 ret = 0; in follow_pmd_mask() 284 ret = -EBUSY; in follow_pmd_mask() [all …]
|
D | hmm.c | 314 int ret; in hmm_vma_walk_hole() local 316 ret = hmm_vma_do_fault(walk, addr, &pfns[i]); in hmm_vma_walk_hole() 317 if (ret != -EAGAIN) in hmm_vma_walk_hole() 318 return ret; in hmm_vma_walk_hole() 339 int ret; in hmm_vma_walk_clear() local 341 ret = hmm_vma_do_fault(walk, addr, &pfns[i]); in hmm_vma_walk_clear() 342 if (ret != -EAGAIN) in hmm_vma_walk_clear() 343 return ret; in hmm_vma_walk_clear() 675 int ret; in hmm_vma_fault() local 723 ret = walk_page_range(start, end, &mm_walk); in hmm_vma_fault() [all …]
|
D | memory-failure.c | 185 int ret; in kill_proc() local 199 ret = force_sig_info(SIGBUS, &si, current); in kill_proc() 208 ret = send_sig_info(SIGBUS, &si, t); /* synchronous? */ in kill_proc() 210 if (ret < 0) in kill_proc() 212 t->comm, t->pid, ret); in kill_proc() 213 return ret; in kill_proc() 562 int ret = MF_FAILED; in truncate_error_page() local 575 ret = MF_RECOVERED; in truncate_error_page() 583 ret = MF_RECOVERED; in truncate_error_page() 589 return ret; in truncate_error_page() [all …]
|
D | zswap.c | 503 int ret; in zswap_pool_create() local 537 ret = cpuhp_state_add_instance(CPUHP_MM_ZSWP_POOL_PREPARE, in zswap_pool_create() 539 if (ret) in zswap_pool_create() 665 int ret; in __zswap_param_set() local 714 ret = param_set_charp(s, kp); in __zswap_param_set() 716 ret = -EINVAL; in __zswap_param_set() 720 if (!ret) { in __zswap_param_set() 743 ret = param_set_charp(s, kp); in __zswap_param_set() 752 return ret; in __zswap_param_set() 845 int ret; in zswap_writeback_entry() local [all …]
|
D | frame_vector.c | 39 int ret = 0; in get_vaddr_frames() local 53 ret = -EFAULT; in get_vaddr_frames() 66 ret = -EOPNOTSUPP; in get_vaddr_frames() 73 ret = get_user_pages_locked(start, nr_frames, in get_vaddr_frames() 83 while (ret < nr_frames && start + PAGE_SIZE <= vma->vm_end) { in get_vaddr_frames() 84 err = follow_pfn(vma, start, &nums[ret]); in get_vaddr_frames() 86 if (ret == 0) in get_vaddr_frames() 87 ret = err; in get_vaddr_frames() 91 ret++; in get_vaddr_frames() 97 if (ret >= nr_frames || start < vma->vm_end) in get_vaddr_frames() [all …]
|
D | memory.c | 1230 int ret; in copy_page_range() local 1250 ret = track_pfn_copy(vma); in copy_page_range() 1251 if (ret) in copy_page_range() 1252 return ret; in copy_page_range() 1268 ret = 0; in copy_page_range() 1277 ret = -ENOMEM; in copy_page_range() 1284 return ret; in copy_page_range() 1887 int ret; in vm_insert_pfn_prot() local 1908 ret = insert_pfn(vma, addr, __pfn_to_pfn_t(pfn, PFN_DEV), pgprot, in vm_insert_pfn_prot() 1911 return ret; in vm_insert_pfn_prot() [all …]
|
D | page-writeback.c | 518 int ret; in dirty_background_ratio_handler() local 520 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_ratio_handler() 521 if (ret == 0 && write) in dirty_background_ratio_handler() 523 return ret; in dirty_background_ratio_handler() 530 int ret; in dirty_background_bytes_handler() local 532 ret = proc_doulongvec_minmax(table, write, buffer, lenp, ppos); in dirty_background_bytes_handler() 533 if (ret == 0 && write) in dirty_background_bytes_handler() 535 return ret; in dirty_background_bytes_handler() 543 int ret; in dirty_ratio_handler() local 545 ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); in dirty_ratio_handler() [all …]
|
D | util.c | 324 unsigned long ret; in vm_mmap_pgoff() local 329 ret = security_mmap_file(file, prot, flag); in vm_mmap_pgoff() 330 if (!ret) { in vm_mmap_pgoff() 333 ret = do_mmap_pgoff(file, addr, len, prot, flag, pgoff, in vm_mmap_pgoff() 338 mm_populate(ret, populate); in vm_mmap_pgoff() 340 return ret; in vm_mmap_pgoff() 375 void *ret; in kvmalloc_node() local 397 ret = kmalloc_node(size, kmalloc_flags, node); in kvmalloc_node() 403 if (ret || size <= PAGE_SIZE) in kvmalloc_node() 404 return ret; in kvmalloc_node() [all …]
|
D | nommu.c | 185 long ret; in __get_user_pages_unlocked() local 187 ret = __get_user_pages(tsk, mm, start, nr_pages, gup_flags, pages, in __get_user_pages_unlocked() 190 return ret; in __get_user_pages_unlocked() 247 void *ret; in vmalloc_user() local 249 ret = __vmalloc(size, GFP_KERNEL | __GFP_ZERO, PAGE_KERNEL); in vmalloc_user() 250 if (ret) { in vmalloc_user() 254 vma = find_vma(current->mm, (unsigned long)ret); in vmalloc_user() 260 return ret; in vmalloc_user() 533 int ret; in mmap_init() local 535 ret = percpu_counter_init(&vm_committed_as, 0, GFP_KERNEL); in mmap_init() [all …]
|
D | huge_memory.c | 164 ssize_t ret = count; in enabled_store() local 176 ret = -EINVAL; in enabled_store() 178 if (ret > 0) { in enabled_store() 181 ret = err; in enabled_store() 183 return ret; in enabled_store() 202 int ret; in single_hugepage_flag_store() local 204 ret = kstrtoul(buf, 10, &value); in single_hugepage_flag_store() 205 if (ret < 0) in single_hugepage_flag_store() 206 return ret; in single_hugepage_flag_store() 441 int ret = 0; in setup_transparent_hugepage() local [all …]
|
D | mremap.c | 447 unsigned long ret = -EINVAL; in mremap_to() local 461 ret = do_munmap(mm, new_addr, new_len, uf_unmap_early); in mremap_to() 462 if (ret) in mremap_to() 466 ret = do_munmap(mm, addr+new_len, old_len - new_len, uf_unmap); in mremap_to() 467 if (ret && old_len != new_len) in mremap_to() 474 ret = PTR_ERR(vma); in mremap_to() 482 ret = get_unmapped_area(vma->vm_file, new_addr, new_len, vma->vm_pgoff + in mremap_to() 485 if (offset_in_page(ret)) in mremap_to() 488 ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, uf, in mremap_to() 490 if (!(offset_in_page(ret))) in mremap_to() [all …]
|
D | cma.c | 155 int ret = cma_activate_area(&cma_areas[i]); in cma_init_reserved_areas() local 157 if (ret) in cma_init_reserved_areas() 158 return ret; in cma_init_reserved_areas() 249 int ret = 0; in cma_declare_contiguous() local 281 ret = -EINVAL; in cma_declare_contiguous() 302 ret = -EINVAL; in cma_declare_contiguous() 317 ret = -EINVAL; in cma_declare_contiguous() 327 ret = -EBUSY; in cma_declare_contiguous() 351 ret = -ENOMEM; in cma_declare_contiguous() 364 ret = cma_init_reserved_mem(base, size, order_per_bit, name, res_cma); in cma_declare_contiguous() [all …]
|
D | frontswap.c | 248 int ret = -1; in __frontswap_store() local 273 ret = ops->store(type, offset, page); in __frontswap_store() 274 if (!ret) /* successful store */ in __frontswap_store() 277 if (ret == 0) { in __frontswap_store() 285 ret = -1; in __frontswap_store() 286 return ret; in __frontswap_store() 297 int ret = -1; in __frontswap_load() local 313 ret = ops->load(type, offset, page); in __frontswap_load() 314 if (!ret) /* successful load */ in __frontswap_load() 317 if (ret == 0) { in __frontswap_load() [all …]
|
D | page_owner.c | 344 int ret; in print_page_owner() local 359 ret = snprintf(kbuf, count, in print_page_owner() 364 if (ret >= count) in print_page_owner() 370 ret += snprintf(kbuf + ret, count - ret, in print_page_owner() 378 if (ret >= count) in print_page_owner() 382 ret += snprint_stack_trace(kbuf + ret, count - ret, &trace, 0); in print_page_owner() 383 if (ret >= count) in print_page_owner() 387 ret += snprintf(kbuf + ret, count - ret, in print_page_owner() 390 if (ret >= count) in print_page_owner() 394 ret += snprintf(kbuf + ret, count - ret, "\n"); in print_page_owner() [all …]
|
D | maccess.c | 28 long ret; in __probe_kernel_read() local 33 ret = __copy_from_user_inatomic(dst, in __probe_kernel_read() 38 return ret ? -EFAULT : 0; in __probe_kernel_read() 56 long ret; in __probe_kernel_write() local 61 ret = __copy_to_user_inatomic((__force void __user *)dst, src, size); in __probe_kernel_write() 65 return ret ? -EFAULT : 0; in __probe_kernel_write() 90 long ret; in strncpy_from_unsafe() local 99 ret = __get_user(*dst++, (const char __user __force *)src++); in strncpy_from_unsafe() 100 } while (dst[-1] && ret == 0 && src - unsafe_addr < count); in strncpy_from_unsafe() 106 return ret ? -EFAULT : src - unsafe_addr; in strncpy_from_unsafe()
|
D | hugetlb.c | 135 long ret = delta; in hugepage_subpool_get_pages() local 138 return ret; in hugepage_subpool_get_pages() 146 ret = -ENOMEM; in hugepage_subpool_get_pages() 158 ret = delta - spool->rsv_hpages; in hugepage_subpool_get_pages() 161 ret = 0; /* reserves already accounted for */ in hugepage_subpool_get_pages() 168 return ret; in hugepage_subpool_get_pages() 180 long ret = delta; in hugepage_subpool_put_pages() local 193 ret = 0; in hugepage_subpool_put_pages() 195 ret = spool->rsv_hpages + delta - spool->min_hpages; in hugepage_subpool_put_pages() 208 return ret; in hugepage_subpool_put_pages() [all …]
|
D | filemap.c | 295 int ret = 0; in filemap_check_errors() local 299 ret = -ENOSPC; in filemap_check_errors() 302 ret = -EIO; in filemap_check_errors() 303 return ret; in filemap_check_errors() 335 int ret; in __filemap_fdatawrite_range() local 348 ret = do_writepages(mapping, &wbc); in __filemap_fdatawrite_range() 350 return ret; in __filemap_fdatawrite_range() 827 int ret; in add_to_page_cache_lru() local 830 ret = __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_lru() 832 if (unlikely(ret)) in add_to_page_cache_lru() [all …]
|
D | readahead.c | 87 int ret = 0; in read_cache_pages() local 99 ret = filler(data, page); in read_cache_pages() 100 if (unlikely(ret)) { in read_cache_pages() 106 return ret; in read_cache_pages() 116 int ret; in read_pages() local 121 ret = mapping->a_ops->readpages(filp, mapping, pages, nr_pages); in read_pages() 134 ret = 0; in read_pages() 139 return ret; in read_pages() 159 int ret = 0; in __do_page_cache_readahead() local 190 ret++; in __do_page_cache_readahead() [all …]
|
D | hugetlb_cgroup.c | 79 int ret; in hugetlb_cgroup_init() local 87 ret = page_counter_limit(counter, limit); in hugetlb_cgroup_init() 88 VM_BUG_ON(ret); in hugetlb_cgroup_init() 184 int ret = 0; in hugetlb_cgroup_charge_cgroup() local 206 ret = -ENOMEM; in hugetlb_cgroup_charge_cgroup() 210 return ret; in hugetlb_cgroup_charge_cgroup() 291 int ret, idx; in hugetlb_cgroup_write() local 299 ret = page_counter_memparse(buf, "-1", &nr_pages); in hugetlb_cgroup_write() 300 if (ret) in hugetlb_cgroup_write() 301 return ret; in hugetlb_cgroup_write() [all …]
|
D | workingset.c | 451 int ret; in shadow_lru_isolate() local 471 ret = LRU_RETRY; in shadow_lru_isolate() 509 ret = LRU_REMOVED_RETRY; in shadow_lru_isolate() 515 return ret; in shadow_lru_isolate() 521 unsigned long ret; in scan_shadow_nodes() local 525 ret = list_lru_shrink_walk(&shadow_nodes, sc, shadow_lru_isolate, NULL); in scan_shadow_nodes() 527 return ret; in scan_shadow_nodes() 547 int ret; in workingset_init() local 564 ret = __list_lru_init(&shadow_nodes, true, &shadow_nodes_key); in workingset_init() 565 if (ret) in workingset_init() [all …]
|
D | memcontrol.c | 945 int ret = 0; in mem_cgroup_scan_tasks() local 954 while (!ret && (task = css_task_iter_next(&it))) in mem_cgroup_scan_tasks() 955 ret = fn(task, arg); in mem_cgroup_scan_tasks() 957 if (ret) { in mem_cgroup_scan_tasks() 962 return ret; in mem_cgroup_scan_tasks() 1049 bool ret; in task_in_mem_cgroup() local 1066 ret = mem_cgroup_is_descendant(task_memcg, memcg); in task_in_mem_cgroup() 1068 return ret; in task_in_mem_cgroup() 1112 bool ret = false; in mem_cgroup_under_move() local 1123 ret = mem_cgroup_is_descendant(from, memcg) || in mem_cgroup_under_move() [all …]
|
D | compaction.c | 1384 int ret; in compact_finished() local 1386 ret = __compact_finished(zone, cc); in compact_finished() 1387 trace_mm_compaction_finished(zone, cc->order, ret); in compact_finished() 1388 if (ret == COMPACT_NO_SUITABLE_PAGE) in compact_finished() 1389 ret = COMPACT_CONTINUE; in compact_finished() 1391 return ret; in compact_finished() 1448 enum compact_result ret; in compaction_suitable() local 1451 ret = __compaction_suitable(zone, order, alloc_flags, classzone_idx, in compaction_suitable() 1469 if (ret == COMPACT_CONTINUE && (order > PAGE_ALLOC_COSTLY_ORDER)) { in compaction_suitable() 1472 ret = COMPACT_NOT_SUITABLE_ZONE; in compaction_suitable() [all …]
|