/mm/ |
D | pagewalk.c | 27 int err = 0; in walk_pte_range_inner() local 30 err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); in walk_pte_range_inner() 31 if (err) in walk_pte_range_inner() 38 return err; in walk_pte_range_inner() 45 int err = 0; in walk_pte_range() local 50 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range() 54 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range() 58 return err; in walk_pte_range() 67 int err = 0; in walk_pmd_range() local 76 err = ops->pte_hole(addr, next, depth, walk); in walk_pmd_range() [all …]
|
D | userfaultfd.c | 289 ssize_t err; in __mcopy_atomic_hugetlb() local 319 err = -EINVAL; in __mcopy_atomic_hugetlb() 329 err = -ENOENT; in __mcopy_atomic_hugetlb() 334 err = -EINVAL; in __mcopy_atomic_hugetlb() 344 err = -ENOMEM; in __mcopy_atomic_hugetlb() 365 err = -ENOMEM; in __mcopy_atomic_hugetlb() 375 err = -EEXIST; in __mcopy_atomic_hugetlb() 381 err = hugetlb_mcopy_atomic_pte(dst_mm, dst_pte, dst_vma, in __mcopy_atomic_hugetlb() 390 if (unlikely(err == -ENOENT)) { in __mcopy_atomic_hugetlb() 394 err = copy_huge_page_from_user(page, in __mcopy_atomic_hugetlb() [all …]
|
D | page_reporting.c | 118 int err = 0; in page_reporting_cycle() local 125 return err; in page_reporting_cycle() 188 err = prdev->report(prdev, sgl, PAGE_REPORTING_CAPACITY); in page_reporting_cycle() 200 page_reporting_drain(prdev, sgl, PAGE_REPORTING_CAPACITY, !err); in page_reporting_cycle() 209 if (err) in page_reporting_cycle() 219 return err; in page_reporting_cycle() 228 int err = 0; in page_reporting_process_zone() local 239 return err; in page_reporting_process_zone() 248 err = page_reporting_cycle(prdev, zone, order, mt, in page_reporting_process_zone() 250 if (err) in page_reporting_process_zone() [all …]
|
D | ksm.c | 849 int err = 0; in unmerge_ksm_pages() local 851 for (addr = start; addr < end && !err; addr += PAGE_SIZE) { in unmerge_ksm_pages() 855 err = -ERESTARTSYS; in unmerge_ksm_pages() 857 err = break_ksm(vma, addr); in unmerge_ksm_pages() 859 return err; in unmerge_ksm_pages() 880 int err; in remove_stable_node() local 895 err = -EBUSY; in remove_stable_node() 907 err = 0; in remove_stable_node() 912 return err; in remove_stable_node() 944 int err = 0; in remove_all_stable_nodes() local [all …]
|
D | mempolicy.c | 750 int err; in queue_pages_range() local 760 err = walk_page_range(mm, start, end, &queue_pages_walk_ops, &qp); in queue_pages_range() 764 err = -EFAULT; in queue_pages_range() 766 return err; in queue_pages_range() 776 int err; in vma_replace_policy() local 791 err = vma->vm_ops->set_policy(vma, new); in vma_replace_policy() 792 if (err) in vma_replace_policy() 809 return err; in vma_replace_policy() 818 int err = 0; in mbind_range() local 848 err = split_vma(vma->vm_mm, vma, vmstart, 1); in mbind_range() [all …]
|
D | slab_common.c | 240 int err; in create_cache() local 245 err = -ENOMEM; in create_cache() 257 err = __kmem_cache_create(s, flags); in create_cache() 258 if (err) in create_cache() 264 if (err) in create_cache() 265 return ERR_PTR(err); in create_cache() 310 int err; in kmem_cache_create_usercopy() local 327 err = kmem_cache_sanity_check(name, size); in kmem_cache_create_usercopy() 328 if (err) { in kmem_cache_create_usercopy() 334 err = -EINVAL; in kmem_cache_create_usercopy() [all …]
|
D | migrate.c | 1587 int err; in do_move_pages_to_node() local 1593 err = migrate_pages(pagelist, alloc_migration_target, NULL, in do_move_pages_to_node() 1595 if (err) in do_move_pages_to_node() 1597 return err; in do_move_pages_to_node() 1615 int err; in add_page_for_migration() local 1618 err = -EFAULT; in add_page_for_migration() 1627 err = PTR_ERR(page); in add_page_for_migration() 1631 err = -ENOENT; in add_page_for_migration() 1635 err = 0; in add_page_for_migration() 1639 err = -EACCES; in add_page_for_migration() [all …]
|
D | filemap.c | 657 int err = 0; in filemap_write_and_wait_range() local 660 err = __filemap_fdatawrite_range(mapping, lstart, lend, in filemap_write_and_wait_range() 668 if (err != -EIO) { in filemap_write_and_wait_range() 671 if (!err) in filemap_write_and_wait_range() 672 err = err2; in filemap_write_and_wait_range() 678 err = filemap_check_errors(mapping); in filemap_write_and_wait_range() 680 return err; in filemap_write_and_wait_range() 684 void __filemap_set_wb_err(struct address_space *mapping, int err) in __filemap_set_wb_err() argument 686 errseq_t eseq = errseq_set(&mapping->wb_err, err); in __filemap_set_wb_err() 718 int err = 0; in file_check_and_advance_wb_err() local [all …]
|
D | percpu-vm.c | 96 goto err; in pcpu_alloc_pages() 101 err: in pcpu_alloc_pages() 217 int i, err; in pcpu_map_pages() local 220 err = __pcpu_map_pages(pcpu_chunk_addr(chunk, cpu, page_start), in pcpu_map_pages() 223 if (err < 0) in pcpu_map_pages() 224 goto err; in pcpu_map_pages() 231 err: in pcpu_map_pages() 239 return err; in pcpu_map_pages()
|
D | cma.c | 288 goto err; in cma_declare_contiguous_nid() 309 goto err; in cma_declare_contiguous_nid() 324 goto err; in cma_declare_contiguous_nid() 332 goto err; in cma_declare_contiguous_nid() 371 goto err; in cma_declare_contiguous_nid() 393 err: in cma_declare_contiguous_nid() 521 if (info.err) { in cma_alloc() 522 if (info.err & ACR_ERR_ISOLATE) in cma_alloc() 524 if (info.err & ACR_ERR_MIGRATE) in cma_alloc() 526 if (info.err & ACR_ERR_TEST) in cma_alloc()
|
D | cma_sysfs.c | 81 int i, err; in cma_sysfs_init() local 93 err = -ENOMEM; in cma_sysfs_init() 100 err = kobject_init_and_add(&cma_kobj->kobj, &cma_ktype, in cma_sysfs_init() 102 if (err) { in cma_sysfs_init() 116 return err; in cma_sysfs_init()
|
D | memory.c | 1107 int err; in copy_pmd_range() local 1109 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range() 1111 if (err == -ENOMEM) in copy_pmd_range() 1113 if (!err) in copy_pmd_range() 1143 int err; in copy_pud_range() local 1146 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range() 1148 if (err == -ENOMEM) in copy_pud_range() 1150 if (!err) in copy_pud_range() 1765 int err; in insert_page_in_batch_locked() local 1769 err = validate_page_before_insert(page); in insert_page_in_batch_locked() [all …]
|
D | khugepaged.c | 135 int err; in scan_sleep_millisecs_store() local 137 err = kstrtoul(buf, 10, &msecs); in scan_sleep_millisecs_store() 138 if (err || msecs > UINT_MAX) in scan_sleep_millisecs_store() 163 int err; in alloc_sleep_millisecs_store() local 165 err = kstrtoul(buf, 10, &msecs); in alloc_sleep_millisecs_store() 166 if (err || msecs > UINT_MAX) in alloc_sleep_millisecs_store() 189 int err; in pages_to_scan_store() local 192 err = kstrtoul(buf, 10, &pages); in pages_to_scan_store() 193 if (err || !pages || pages > UINT_MAX) in pages_to_scan_store() 257 int err; in khugepaged_max_ptes_none_store() local [all …]
|
D | swap_state.c | 220 int err; in add_to_swap() local 240 err = add_to_swap_cache(page, entry, in add_to_swap() 242 if (err) in add_to_swap() 463 int err; in __read_swap_cache_async() local 501 err = swapcache_prepare(entry); in __read_swap_cache_async() 502 if (!err) in __read_swap_cache_async() 506 if (err != -EEXIST) in __read_swap_cache_async() 938 int err; in swap_init_sysfs() local 946 err = sysfs_create_group(swap_kobj, &swap_attr_group); in swap_init_sysfs() 947 if (err) { in swap_init_sysfs() [all …]
|
D | page_idle.c | 216 int err; in page_idle_init() local 218 err = sysfs_create_group(mm_kobj, &page_idle_attr_group); in page_idle_init() 219 if (err) { in page_idle_init() 221 return err; in page_idle_init()
|
D | swapfile.c | 177 int err = 0; in discard_swap() local 184 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap() 186 if (err) in discard_swap() 187 return err; in discard_swap() 195 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap() 197 if (err) in discard_swap() 202 return err; /* That will often be -EOPNOTSUPP */ in discard_swap() 2605 int err, found = 0; in SYSCALL_DEFINE1() local 2619 err = PTR_ERR(victim); in SYSCALL_DEFINE1() 2634 err = -EINVAL; in SYSCALL_DEFINE1() [all …]
|
D | hwpoison-inject.c | 19 int err; in hwpoison_inject() local 45 err = hwpoison_filter(hpage); in hwpoison_inject() 46 if (err) in hwpoison_inject()
|
D | mincore.c | 188 int err; in do_mincore() local 199 err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); in do_mincore() 200 if (err < 0) in do_mincore() 201 return err; in do_mincore()
|
D | mremap.c | 553 int err; in move_vma() local 570 err = ksm_madvise(vma, old_addr, old_addr + old_len, in move_vma() 572 if (err) in move_vma() 573 return err; in move_vma() 592 err = -ENOMEM; in move_vma() 594 err = vma->vm_ops->mremap(new_vma); in move_vma() 597 if (unlikely(err)) { in move_vma() 610 new_addr = err; in move_vma() 645 if (unlikely(!err && (flags & MREMAP_DONTUNMAP))) { in move_vma()
|
D | backing-dev.c | 237 int err; in default_bdi_init() local 244 err = bdi_init(&noop_backing_dev_info); in default_bdi_init() 246 return err; in default_bdi_init() 283 int i, err; in wb_init() local 308 err = fprop_local_init_percpu(&wb->completions, gfp); in wb_init() 309 if (err) in wb_init() 313 err = percpu_counter_init(&wb->stat[i], 0, gfp); in wb_init() 314 if (err) in wb_init() 327 return err; in wb_init()
|
D | ioremap.c | 228 int err; in ioremap_page_range() local 238 err = ioremap_p4d_range(pgd, addr, next, phys_addr, prot, in ioremap_page_range() 240 if (err) in ioremap_page_range() 249 return err; in ioremap_page_range()
|
D | slub.c | 1689 int err; in init_cache_random_seq() local 1695 err = cache_random_seq_create(s, count, GFP_KERNEL); in init_cache_random_seq() 1696 if (err) { in init_cache_random_seq() 1699 return err; in init_cache_random_seq() 4510 int err; in __kmem_cache_create() local 4512 err = kmem_cache_open(s, flags); in __kmem_cache_create() 4513 if (err) in __kmem_cache_create() 4514 return err; in __kmem_cache_create() 4520 err = sysfs_slab_add(s); in __kmem_cache_create() 4521 if (err) { in __kmem_cache_create() [all …]
|
D | page_owner.c | 408 goto err; in print_page_owner() 422 goto err; in print_page_owner() 427 goto err; in print_page_owner() 434 goto err; in print_page_owner() 439 goto err; in print_page_owner() 447 err: in print_page_owner()
|
/mm/damon/ |
D | reclaim.c | 296 int err; in damon_reclaim_turn() local 299 err = damon_stop(&ctx, 1); in damon_reclaim_turn() 300 if (!err) in damon_reclaim_turn() 302 return err; in damon_reclaim_turn() 305 err = damon_set_attrs(ctx, sample_interval, aggr_interval, 0, in damon_reclaim_turn() 307 if (err) in damon_reclaim_turn() 308 return err; in damon_reclaim_turn() 325 err = -ENOMEM; in damon_reclaim_turn() 328 err = damon_set_schemes(ctx, &scheme, 1); in damon_reclaim_turn() 329 if (err) in damon_reclaim_turn() [all …]
|
D | core.c | 383 int err = -EBUSY; in __damon_start() local 387 err = 0; in __damon_start() 391 err = PTR_ERR(ctx->kdamond); in __damon_start() 397 return err; in __damon_start() 416 int err = 0; in damon_start() local 425 err = __damon_start(ctxs[i]); in damon_start() 426 if (err) in damon_start() 432 return err; in damon_start() 468 int i, err = 0; in damon_stop() local 472 err = __damon_stop(ctxs[i]); in damon_stop() [all …]
|