Home
last modified time | relevance | path

Searched refs:err (Results 1 – 25 of 49) sorted by relevance

12

/mm/
Dpagewalk.c27 int err = 0; in walk_pte_range_inner() local
30 err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); in walk_pte_range_inner()
31 if (err) in walk_pte_range_inner()
38 return err; in walk_pte_range_inner()
45 int err = 0; in walk_pte_range() local
50 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range()
54 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range()
58 return err; in walk_pte_range()
67 int err = 0; in walk_pmd_range() local
76 err = ops->pte_hole(addr, next, depth, walk); in walk_pmd_range()
[all …]
Duserfaultfd.c289 ssize_t err; in __mcopy_atomic_hugetlb() local
319 err = -EINVAL; in __mcopy_atomic_hugetlb()
329 err = -ENOENT; in __mcopy_atomic_hugetlb()
334 err = -EINVAL; in __mcopy_atomic_hugetlb()
344 err = -ENOMEM; in __mcopy_atomic_hugetlb()
365 err = -ENOMEM; in __mcopy_atomic_hugetlb()
375 err = -EEXIST; in __mcopy_atomic_hugetlb()
381 err = hugetlb_mcopy_atomic_pte(dst_mm, dst_pte, dst_vma, in __mcopy_atomic_hugetlb()
390 if (unlikely(err == -ENOENT)) { in __mcopy_atomic_hugetlb()
394 err = copy_huge_page_from_user(page, in __mcopy_atomic_hugetlb()
[all …]
Dpage_reporting.c118 int err = 0; in page_reporting_cycle() local
125 return err; in page_reporting_cycle()
188 err = prdev->report(prdev, sgl, PAGE_REPORTING_CAPACITY); in page_reporting_cycle()
200 page_reporting_drain(prdev, sgl, PAGE_REPORTING_CAPACITY, !err); in page_reporting_cycle()
209 if (err) in page_reporting_cycle()
219 return err; in page_reporting_cycle()
228 int err = 0; in page_reporting_process_zone() local
239 return err; in page_reporting_process_zone()
248 err = page_reporting_cycle(prdev, zone, order, mt, in page_reporting_process_zone()
250 if (err) in page_reporting_process_zone()
[all …]
Dksm.c849 int err = 0; in unmerge_ksm_pages() local
851 for (addr = start; addr < end && !err; addr += PAGE_SIZE) { in unmerge_ksm_pages()
855 err = -ERESTARTSYS; in unmerge_ksm_pages()
857 err = break_ksm(vma, addr); in unmerge_ksm_pages()
859 return err; in unmerge_ksm_pages()
880 int err; in remove_stable_node() local
895 err = -EBUSY; in remove_stable_node()
907 err = 0; in remove_stable_node()
912 return err; in remove_stable_node()
944 int err = 0; in remove_all_stable_nodes() local
[all …]
Dmempolicy.c750 int err; in queue_pages_range() local
760 err = walk_page_range(mm, start, end, &queue_pages_walk_ops, &qp); in queue_pages_range()
764 err = -EFAULT; in queue_pages_range()
766 return err; in queue_pages_range()
776 int err; in vma_replace_policy() local
791 err = vma->vm_ops->set_policy(vma, new); in vma_replace_policy()
792 if (err) in vma_replace_policy()
809 return err; in vma_replace_policy()
818 int err = 0; in mbind_range() local
848 err = split_vma(vma->vm_mm, vma, vmstart, 1); in mbind_range()
[all …]
Dslab_common.c240 int err; in create_cache() local
245 err = -ENOMEM; in create_cache()
257 err = __kmem_cache_create(s, flags); in create_cache()
258 if (err) in create_cache()
264 if (err) in create_cache()
265 return ERR_PTR(err); in create_cache()
310 int err; in kmem_cache_create_usercopy() local
327 err = kmem_cache_sanity_check(name, size); in kmem_cache_create_usercopy()
328 if (err) { in kmem_cache_create_usercopy()
334 err = -EINVAL; in kmem_cache_create_usercopy()
[all …]
Dmigrate.c1587 int err; in do_move_pages_to_node() local
1593 err = migrate_pages(pagelist, alloc_migration_target, NULL, in do_move_pages_to_node()
1595 if (err) in do_move_pages_to_node()
1597 return err; in do_move_pages_to_node()
1615 int err; in add_page_for_migration() local
1618 err = -EFAULT; in add_page_for_migration()
1627 err = PTR_ERR(page); in add_page_for_migration()
1631 err = -ENOENT; in add_page_for_migration()
1635 err = 0; in add_page_for_migration()
1639 err = -EACCES; in add_page_for_migration()
[all …]
Dfilemap.c657 int err = 0; in filemap_write_and_wait_range() local
660 err = __filemap_fdatawrite_range(mapping, lstart, lend, in filemap_write_and_wait_range()
668 if (err != -EIO) { in filemap_write_and_wait_range()
671 if (!err) in filemap_write_and_wait_range()
672 err = err2; in filemap_write_and_wait_range()
678 err = filemap_check_errors(mapping); in filemap_write_and_wait_range()
680 return err; in filemap_write_and_wait_range()
684 void __filemap_set_wb_err(struct address_space *mapping, int err) in __filemap_set_wb_err() argument
686 errseq_t eseq = errseq_set(&mapping->wb_err, err); in __filemap_set_wb_err()
718 int err = 0; in file_check_and_advance_wb_err() local
[all …]
Dpercpu-vm.c96 goto err; in pcpu_alloc_pages()
101 err: in pcpu_alloc_pages()
217 int i, err; in pcpu_map_pages() local
220 err = __pcpu_map_pages(pcpu_chunk_addr(chunk, cpu, page_start), in pcpu_map_pages()
223 if (err < 0) in pcpu_map_pages()
224 goto err; in pcpu_map_pages()
231 err: in pcpu_map_pages()
239 return err; in pcpu_map_pages()
Dcma.c288 goto err; in cma_declare_contiguous_nid()
309 goto err; in cma_declare_contiguous_nid()
324 goto err; in cma_declare_contiguous_nid()
332 goto err; in cma_declare_contiguous_nid()
371 goto err; in cma_declare_contiguous_nid()
393 err: in cma_declare_contiguous_nid()
521 if (info.err) { in cma_alloc()
522 if (info.err & ACR_ERR_ISOLATE) in cma_alloc()
524 if (info.err & ACR_ERR_MIGRATE) in cma_alloc()
526 if (info.err & ACR_ERR_TEST) in cma_alloc()
Dcma_sysfs.c81 int i, err; in cma_sysfs_init() local
93 err = -ENOMEM; in cma_sysfs_init()
100 err = kobject_init_and_add(&cma_kobj->kobj, &cma_ktype, in cma_sysfs_init()
102 if (err) { in cma_sysfs_init()
116 return err; in cma_sysfs_init()
Dmemory.c1107 int err; in copy_pmd_range() local
1109 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range()
1111 if (err == -ENOMEM) in copy_pmd_range()
1113 if (!err) in copy_pmd_range()
1143 int err; in copy_pud_range() local
1146 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range()
1148 if (err == -ENOMEM) in copy_pud_range()
1150 if (!err) in copy_pud_range()
1765 int err; in insert_page_in_batch_locked() local
1769 err = validate_page_before_insert(page); in insert_page_in_batch_locked()
[all …]
Dkhugepaged.c135 int err; in scan_sleep_millisecs_store() local
137 err = kstrtoul(buf, 10, &msecs); in scan_sleep_millisecs_store()
138 if (err || msecs > UINT_MAX) in scan_sleep_millisecs_store()
163 int err; in alloc_sleep_millisecs_store() local
165 err = kstrtoul(buf, 10, &msecs); in alloc_sleep_millisecs_store()
166 if (err || msecs > UINT_MAX) in alloc_sleep_millisecs_store()
189 int err; in pages_to_scan_store() local
192 err = kstrtoul(buf, 10, &pages); in pages_to_scan_store()
193 if (err || !pages || pages > UINT_MAX) in pages_to_scan_store()
257 int err; in khugepaged_max_ptes_none_store() local
[all …]
Dswap_state.c220 int err; in add_to_swap() local
240 err = add_to_swap_cache(page, entry, in add_to_swap()
242 if (err) in add_to_swap()
463 int err; in __read_swap_cache_async() local
501 err = swapcache_prepare(entry); in __read_swap_cache_async()
502 if (!err) in __read_swap_cache_async()
506 if (err != -EEXIST) in __read_swap_cache_async()
938 int err; in swap_init_sysfs() local
946 err = sysfs_create_group(swap_kobj, &swap_attr_group); in swap_init_sysfs()
947 if (err) { in swap_init_sysfs()
[all …]
Dpage_idle.c216 int err; in page_idle_init() local
218 err = sysfs_create_group(mm_kobj, &page_idle_attr_group); in page_idle_init()
219 if (err) { in page_idle_init()
221 return err; in page_idle_init()
Dswapfile.c177 int err = 0; in discard_swap() local
184 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap()
186 if (err) in discard_swap()
187 return err; in discard_swap()
195 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap()
197 if (err) in discard_swap()
202 return err; /* That will often be -EOPNOTSUPP */ in discard_swap()
2605 int err, found = 0; in SYSCALL_DEFINE1() local
2619 err = PTR_ERR(victim); in SYSCALL_DEFINE1()
2634 err = -EINVAL; in SYSCALL_DEFINE1()
[all …]
Dhwpoison-inject.c19 int err; in hwpoison_inject() local
45 err = hwpoison_filter(hpage); in hwpoison_inject()
46 if (err) in hwpoison_inject()
Dmincore.c188 int err; in do_mincore() local
199 err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); in do_mincore()
200 if (err < 0) in do_mincore()
201 return err; in do_mincore()
Dmremap.c553 int err; in move_vma() local
570 err = ksm_madvise(vma, old_addr, old_addr + old_len, in move_vma()
572 if (err) in move_vma()
573 return err; in move_vma()
592 err = -ENOMEM; in move_vma()
594 err = vma->vm_ops->mremap(new_vma); in move_vma()
597 if (unlikely(err)) { in move_vma()
610 new_addr = err; in move_vma()
645 if (unlikely(!err && (flags & MREMAP_DONTUNMAP))) { in move_vma()
Dbacking-dev.c237 int err; in default_bdi_init() local
244 err = bdi_init(&noop_backing_dev_info); in default_bdi_init()
246 return err; in default_bdi_init()
283 int i, err; in wb_init() local
308 err = fprop_local_init_percpu(&wb->completions, gfp); in wb_init()
309 if (err) in wb_init()
313 err = percpu_counter_init(&wb->stat[i], 0, gfp); in wb_init()
314 if (err) in wb_init()
327 return err; in wb_init()
Dioremap.c228 int err; in ioremap_page_range() local
238 err = ioremap_p4d_range(pgd, addr, next, phys_addr, prot, in ioremap_page_range()
240 if (err) in ioremap_page_range()
249 return err; in ioremap_page_range()
Dslub.c1689 int err; in init_cache_random_seq() local
1695 err = cache_random_seq_create(s, count, GFP_KERNEL); in init_cache_random_seq()
1696 if (err) { in init_cache_random_seq()
1699 return err; in init_cache_random_seq()
4510 int err; in __kmem_cache_create() local
4512 err = kmem_cache_open(s, flags); in __kmem_cache_create()
4513 if (err) in __kmem_cache_create()
4514 return err; in __kmem_cache_create()
4520 err = sysfs_slab_add(s); in __kmem_cache_create()
4521 if (err) { in __kmem_cache_create()
[all …]
Dpage_owner.c408 goto err; in print_page_owner()
422 goto err; in print_page_owner()
427 goto err; in print_page_owner()
434 goto err; in print_page_owner()
439 goto err; in print_page_owner()
447 err: in print_page_owner()
/mm/damon/
Dreclaim.c296 int err; in damon_reclaim_turn() local
299 err = damon_stop(&ctx, 1); in damon_reclaim_turn()
300 if (!err) in damon_reclaim_turn()
302 return err; in damon_reclaim_turn()
305 err = damon_set_attrs(ctx, sample_interval, aggr_interval, 0, in damon_reclaim_turn()
307 if (err) in damon_reclaim_turn()
308 return err; in damon_reclaim_turn()
325 err = -ENOMEM; in damon_reclaim_turn()
328 err = damon_set_schemes(ctx, &scheme, 1); in damon_reclaim_turn()
329 if (err) in damon_reclaim_turn()
[all …]
Dcore.c383 int err = -EBUSY; in __damon_start() local
387 err = 0; in __damon_start()
391 err = PTR_ERR(ctx->kdamond); in __damon_start()
397 return err; in __damon_start()
416 int err = 0; in damon_start() local
425 err = __damon_start(ctxs[i]); in damon_start()
426 if (err) in damon_start()
432 return err; in damon_start()
468 int i, err = 0; in damon_stop() local
472 err = __damon_stop(ctxs[i]); in damon_stop()
[all …]

12