/mm/ |
D | pagewalk.c | 27 int err = 0; in walk_pte_range_inner() local 30 err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); in walk_pte_range_inner() 31 if (err) in walk_pte_range_inner() 38 return err; in walk_pte_range_inner() 45 int err = 0; in walk_pte_range() local 50 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range() 54 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range() 58 return err; in walk_pte_range() 65 int err = 0; in walk_hugepd_range() local 81 err = ops->pte_entry(pte, addr, addr + page_size, walk); in walk_hugepd_range() [all …]
|
D | userfaultfd.c | 296 ssize_t err; in __mcopy_atomic_hugetlb() local 326 err = -EINVAL; in __mcopy_atomic_hugetlb() 336 err = -ENOENT; in __mcopy_atomic_hugetlb() 341 err = -EINVAL; in __mcopy_atomic_hugetlb() 351 err = -ENOMEM; in __mcopy_atomic_hugetlb() 372 err = -ENOMEM; in __mcopy_atomic_hugetlb() 382 err = -EEXIST; in __mcopy_atomic_hugetlb() 388 err = hugetlb_mcopy_atomic_pte(dst_mm, dst_pte, dst_vma, in __mcopy_atomic_hugetlb() 396 if (unlikely(err == -ENOENT)) { in __mcopy_atomic_hugetlb() 400 err = copy_huge_page_from_user(page, in __mcopy_atomic_hugetlb() [all …]
|
D | page_reporting.c | 123 int err = 0; in page_reporting_cycle() local 130 return err; in page_reporting_cycle() 193 err = prdev->report(prdev, sgl, PAGE_REPORTING_CAPACITY); in page_reporting_cycle() 205 page_reporting_drain(prdev, sgl, PAGE_REPORTING_CAPACITY, !err); in page_reporting_cycle() 214 if (err) in page_reporting_cycle() 224 return err; in page_reporting_cycle() 233 int err = 0; in page_reporting_process_zone() local 244 return err; in page_reporting_process_zone() 253 err = page_reporting_cycle(prdev, zone, order, mt, in page_reporting_process_zone() 255 if (err) in page_reporting_process_zone() [all …]
|
D | mempolicy.c | 735 int err; in queue_pages_range() local 745 err = walk_page_range(mm, start, end, &queue_pages_walk_ops, &qp); in queue_pages_range() 749 err = -EFAULT; in queue_pages_range() 751 return err; in queue_pages_range() 761 int err; in vma_replace_policy() local 775 err = vma->vm_ops->set_policy(vma, new); in vma_replace_policy() 776 if (err) in vma_replace_policy() 787 return err; in vma_replace_policy() 796 int err = 0; in mbind_range() local 826 err = split_vma(vma->vm_mm, vma, vmstart, 1); in mbind_range() [all …]
|
D | ksm.c | 842 int err = 0; in unmerge_ksm_pages() local 844 for (addr = start; addr < end && !err; addr += PAGE_SIZE) { in unmerge_ksm_pages() 848 err = -ERESTARTSYS; in unmerge_ksm_pages() 850 err = break_ksm(vma, addr); in unmerge_ksm_pages() 852 return err; in unmerge_ksm_pages() 873 int err; in remove_stable_node() local 888 err = -EBUSY; in remove_stable_node() 900 err = 0; in remove_stable_node() 905 return err; in remove_stable_node() 937 int err = 0; in remove_all_stable_nodes() local [all …]
|
D | secretmem.c | 59 int err; in secretmem_fault() local 75 err = set_direct_map_invalid_noflush(page); in secretmem_fault() 76 if (err) { in secretmem_fault() 78 ret = vmf_error(err); in secretmem_fault() 83 err = add_to_page_cache_lru(page, mapping, offset, gfp); in secretmem_fault() 84 if (unlikely(err)) { in secretmem_fault() 92 if (err == -EEXIST) in secretmem_fault() 95 ret = vmf_error(err); in secretmem_fault() 232 int fd, err; in SYSCALL_DEFINE1() local 251 err = PTR_ERR(file); in SYSCALL_DEFINE1() [all …]
|
D | slab_common.c | 248 int err; in create_cache() local 253 err = -ENOMEM; in create_cache() 265 err = __kmem_cache_create(s, flags); in create_cache() 266 if (err) in create_cache() 272 if (err) in create_cache() 273 return ERR_PTR(err); in create_cache() 318 int err; in kmem_cache_create_usercopy() local 332 err = kmem_cache_sanity_check(name, size); in kmem_cache_create_usercopy() 333 if (err) { in kmem_cache_create_usercopy() 339 err = -EINVAL; in kmem_cache_create_usercopy() [all …]
|
D | migrate.c | 1679 int err; in do_move_pages_to_node() local 1685 err = migrate_pages(pagelist, alloc_migration_target, NULL, in do_move_pages_to_node() 1687 if (err) in do_move_pages_to_node() 1689 return err; in do_move_pages_to_node() 1707 int err; in add_page_for_migration() local 1710 err = -EFAULT; in add_page_for_migration() 1719 err = PTR_ERR(page); in add_page_for_migration() 1723 err = -ENOENT; in add_page_for_migration() 1727 err = 0; in add_page_for_migration() 1731 err = -EACCES; in add_page_for_migration() [all …]
|
D | filemap.c | 705 int err = 0; in filemap_write_and_wait_range() local 708 err = __filemap_fdatawrite_range(mapping, lstart, lend, in filemap_write_and_wait_range() 716 if (err != -EIO) { in filemap_write_and_wait_range() 719 if (!err) in filemap_write_and_wait_range() 720 err = err2; in filemap_write_and_wait_range() 726 err = filemap_check_errors(mapping); in filemap_write_and_wait_range() 728 return err; in filemap_write_and_wait_range() 732 void __filemap_set_wb_err(struct address_space *mapping, int err) in __filemap_set_wb_err() argument 734 errseq_t eseq = errseq_set(&mapping->wb_err, err); in __filemap_set_wb_err() 766 int err = 0; in file_check_and_advance_wb_err() local [all …]
|
D | hwpoison-inject.c | 19 int err; in hwpoison_inject() local 45 err = hwpoison_filter(hpage); in hwpoison_inject() 46 if (err) in hwpoison_inject() 51 err = memory_failure(pfn, 0); in hwpoison_inject() 52 return (err == -EOPNOTSUPP) ? 0 : err; in hwpoison_inject()
|
D | percpu-vm.c | 97 goto err; in pcpu_alloc_pages() 102 err: in pcpu_alloc_pages() 218 int i, err; in pcpu_map_pages() local 221 err = __pcpu_map_pages(pcpu_chunk_addr(chunk, cpu, page_start), in pcpu_map_pages() 224 if (err < 0) in pcpu_map_pages() 225 goto err; in pcpu_map_pages() 232 err: in pcpu_map_pages() 240 return err; in pcpu_map_pages()
|
D | cma_sysfs.c | 78 int i, err; in cma_sysfs_init() local 87 err = -ENOMEM; in cma_sysfs_init() 94 err = kobject_init_and_add(&cma_kobj->kobj, &cma_ktype, in cma_sysfs_init() 96 if (err) { in cma_sysfs_init() 110 return err; in cma_sysfs_init()
|
D | memory.c | 1203 int err; in copy_pmd_range() local 1205 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range() 1207 if (err == -ENOMEM) in copy_pmd_range() 1209 if (!err) in copy_pmd_range() 1239 int err; in copy_pud_range() local 1242 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range() 1244 if (err == -ENOMEM) in copy_pud_range() 1246 if (!err) in copy_pud_range() 1864 int err; in insert_page_in_batch_locked() local 1868 err = validate_page_before_insert(page); in insert_page_in_batch_locked() [all …]
|
D | mremap.c | 630 int err = 0; in move_vma() local 642 err = vma->vm_ops->may_split(vma, old_addr); in move_vma() 643 if (!err && vma->vm_end != old_addr + old_len) in move_vma() 644 err = vma->vm_ops->may_split(vma, old_addr + old_len); in move_vma() 645 if (err) in move_vma() 646 return err; in move_vma() 656 err = ksm_madvise(vma, old_addr, old_addr + old_len, in move_vma() 658 if (err) in move_vma() 659 return err; in move_vma() 678 err = -ENOMEM; in move_vma() [all …]
|
D | swap_state.c | 187 int err; in add_to_swap() local 207 err = add_to_swap_cache(page, entry, in add_to_swap() 209 if (err) in add_to_swap() 428 int err; in __read_swap_cache_async() local 466 err = swapcache_prepare(entry); in __read_swap_cache_async() 467 if (!err) in __read_swap_cache_async() 471 if (err != -EEXIST) in __read_swap_cache_async() 892 int err; in swap_init_sysfs() local 900 err = sysfs_create_group(swap_kobj, &swap_attr_group); in swap_init_sysfs() 901 if (err) { in swap_init_sysfs() [all …]
|
D | khugepaged.c | 137 int err; in scan_sleep_millisecs_store() local 139 err = kstrtouint(buf, 10, &msecs); in scan_sleep_millisecs_store() 140 if (err) in scan_sleep_millisecs_store() 165 int err; in alloc_sleep_millisecs_store() local 167 err = kstrtouint(buf, 10, &msecs); in alloc_sleep_millisecs_store() 168 if (err) in alloc_sleep_millisecs_store() 192 int err; in pages_to_scan_store() local 194 err = kstrtouint(buf, 10, &pages); in pages_to_scan_store() 195 if (err || !pages) in pages_to_scan_store() 259 int err; in khugepaged_max_ptes_none_store() local [all …]
|
D | page_idle.c | 212 int err; in page_idle_init() local 214 err = sysfs_create_group(mm_kobj, &page_idle_attr_group); in page_idle_init() 215 if (err) { in page_idle_init() 217 return err; in page_idle_init()
|
D | swapfile.c | 175 int err = 0; in discard_swap() local 182 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap() 184 if (err) in discard_swap() 185 return err; in discard_swap() 193 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap() 195 if (err) in discard_swap() 200 return err; /* That will often be -EOPNOTSUPP */ in discard_swap() 2531 int err, found = 0; in SYSCALL_DEFINE1() local 2545 err = PTR_ERR(victim); in SYSCALL_DEFINE1() 2560 err = -EINVAL; in SYSCALL_DEFINE1() [all …]
|
D | mincore.c | 189 int err; in do_mincore() local 200 err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); in do_mincore() 201 if (err < 0) in do_mincore() 202 return err; in do_mincore()
|
D | cma.c | 287 goto err; in cma_declare_contiguous_nid() 308 goto err; in cma_declare_contiguous_nid() 323 goto err; in cma_declare_contiguous_nid() 331 goto err; in cma_declare_contiguous_nid() 370 goto err; in cma_declare_contiguous_nid() 392 err: in cma_declare_contiguous_nid()
|
D | page_pinner.c | 218 goto err; in print_page_pinner() 230 goto err; in print_page_pinner() 236 goto err; in print_page_pinner() 240 goto err; in print_page_pinner() 248 err: in print_page_pinner()
|
D | slub.c | 1807 int err; in init_cache_random_seq() local 1813 err = cache_random_seq_create(s, count, GFP_KERNEL); in init_cache_random_seq() 1814 if (err) { in init_cache_random_seq() 1817 return err; in init_cache_random_seq() 4910 int err; in __kmem_cache_create() local 4912 err = kmem_cache_open(s, flags); in __kmem_cache_create() 4913 if (err) in __kmem_cache_create() 4914 return err; in __kmem_cache_create() 4920 err = sysfs_slab_add(s); in __kmem_cache_create() 4921 if (err) { in __kmem_cache_create() [all …]
|
D | vmalloc.c | 293 int err; in vmap_range_noflush() local 303 err = vmap_p4d_range(pgd, addr, next, phys_addr, prot, in vmap_range_noflush() 305 if (err) in vmap_range_noflush() 312 return err; in vmap_range_noflush() 318 int err; in ioremap_page_range() local 321 err = vmap_range_noflush(addr, end, phys_addr, prot, in ioremap_page_range() 325 if (IS_ENABLED(CONFIG_ARCH_HAS_IOREMAP_PHYS_HOOKS) && !err) in ioremap_page_range() 328 return err; in ioremap_page_range() 553 int err = 0; in vmap_small_pages_range_noflush() local 563 err = vmap_pages_p4d_range(pgd, addr, next, prot, pages, &nr, &mask); in vmap_small_pages_range_noflush() [all …]
|
/mm/damon/ |
D | reclaim.c | 296 int err; in damon_reclaim_turn() local 299 err = damon_stop(&ctx, 1); in damon_reclaim_turn() 300 if (!err) in damon_reclaim_turn() 302 return err; in damon_reclaim_turn() 305 err = damon_set_attrs(ctx, sample_interval, aggr_interval, 0, in damon_reclaim_turn() 307 if (err) in damon_reclaim_turn() 308 return err; in damon_reclaim_turn() 325 err = -ENOMEM; in damon_reclaim_turn() 328 err = damon_set_schemes(ctx, &scheme, 1); in damon_reclaim_turn() 329 if (err) in damon_reclaim_turn() [all …]
|
D | core.c | 382 int err = -EBUSY; in __damon_start() local 386 err = 0; in __damon_start() 390 err = PTR_ERR(ctx->kdamond); in __damon_start() 396 return err; in __damon_start() 415 int err = 0; in damon_start() local 424 err = __damon_start(ctxs[i]); in damon_start() 425 if (err) in damon_start() 431 return err; in damon_start() 476 int i, err = 0; in damon_stop() local 480 err = __damon_stop(ctxs[i]); in damon_stop() [all …]
|