Home
last modified time | relevance | path

Searched refs:err (Results 1 – 25 of 48) sorted by relevance

12

/mm/
Dpagewalk.c27 int err = 0; in walk_pte_range_inner() local
30 err = ops->pte_entry(pte, addr, addr + PAGE_SIZE, walk); in walk_pte_range_inner()
31 if (err) in walk_pte_range_inner()
38 return err; in walk_pte_range_inner()
45 int err = 0; in walk_pte_range() local
50 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range()
54 err = walk_pte_range_inner(pte, addr, end, walk); in walk_pte_range()
58 return err; in walk_pte_range()
65 int err = 0; in walk_hugepd_range() local
81 err = ops->pte_entry(pte, addr, addr + page_size, walk); in walk_hugepd_range()
[all …]
Duserfaultfd.c296 ssize_t err; in __mcopy_atomic_hugetlb() local
326 err = -EINVAL; in __mcopy_atomic_hugetlb()
336 err = -ENOENT; in __mcopy_atomic_hugetlb()
341 err = -EINVAL; in __mcopy_atomic_hugetlb()
351 err = -ENOMEM; in __mcopy_atomic_hugetlb()
372 err = -ENOMEM; in __mcopy_atomic_hugetlb()
382 err = -EEXIST; in __mcopy_atomic_hugetlb()
388 err = hugetlb_mcopy_atomic_pte(dst_mm, dst_pte, dst_vma, in __mcopy_atomic_hugetlb()
396 if (unlikely(err == -ENOENT)) { in __mcopy_atomic_hugetlb()
400 err = copy_huge_page_from_user(page, in __mcopy_atomic_hugetlb()
[all …]
Dpage_reporting.c123 int err = 0; in page_reporting_cycle() local
130 return err; in page_reporting_cycle()
193 err = prdev->report(prdev, sgl, PAGE_REPORTING_CAPACITY); in page_reporting_cycle()
205 page_reporting_drain(prdev, sgl, PAGE_REPORTING_CAPACITY, !err); in page_reporting_cycle()
214 if (err) in page_reporting_cycle()
224 return err; in page_reporting_cycle()
233 int err = 0; in page_reporting_process_zone() local
244 return err; in page_reporting_process_zone()
253 err = page_reporting_cycle(prdev, zone, order, mt, in page_reporting_process_zone()
255 if (err) in page_reporting_process_zone()
[all …]
Dmempolicy.c735 int err; in queue_pages_range() local
745 err = walk_page_range(mm, start, end, &queue_pages_walk_ops, &qp); in queue_pages_range()
749 err = -EFAULT; in queue_pages_range()
751 return err; in queue_pages_range()
761 int err; in vma_replace_policy() local
775 err = vma->vm_ops->set_policy(vma, new); in vma_replace_policy()
776 if (err) in vma_replace_policy()
787 return err; in vma_replace_policy()
796 int err = 0; in mbind_range() local
826 err = split_vma(vma->vm_mm, vma, vmstart, 1); in mbind_range()
[all …]
Dksm.c842 int err = 0; in unmerge_ksm_pages() local
844 for (addr = start; addr < end && !err; addr += PAGE_SIZE) { in unmerge_ksm_pages()
848 err = -ERESTARTSYS; in unmerge_ksm_pages()
850 err = break_ksm(vma, addr); in unmerge_ksm_pages()
852 return err; in unmerge_ksm_pages()
873 int err; in remove_stable_node() local
888 err = -EBUSY; in remove_stable_node()
900 err = 0; in remove_stable_node()
905 return err; in remove_stable_node()
937 int err = 0; in remove_all_stable_nodes() local
[all …]
Dsecretmem.c59 int err; in secretmem_fault() local
75 err = set_direct_map_invalid_noflush(page); in secretmem_fault()
76 if (err) { in secretmem_fault()
78 ret = vmf_error(err); in secretmem_fault()
83 err = add_to_page_cache_lru(page, mapping, offset, gfp); in secretmem_fault()
84 if (unlikely(err)) { in secretmem_fault()
92 if (err == -EEXIST) in secretmem_fault()
95 ret = vmf_error(err); in secretmem_fault()
232 int fd, err; in SYSCALL_DEFINE1() local
251 err = PTR_ERR(file); in SYSCALL_DEFINE1()
[all …]
Dslab_common.c248 int err; in create_cache() local
253 err = -ENOMEM; in create_cache()
265 err = __kmem_cache_create(s, flags); in create_cache()
266 if (err) in create_cache()
272 if (err) in create_cache()
273 return ERR_PTR(err); in create_cache()
318 int err; in kmem_cache_create_usercopy() local
332 err = kmem_cache_sanity_check(name, size); in kmem_cache_create_usercopy()
333 if (err) { in kmem_cache_create_usercopy()
339 err = -EINVAL; in kmem_cache_create_usercopy()
[all …]
Dmigrate.c1679 int err; in do_move_pages_to_node() local
1685 err = migrate_pages(pagelist, alloc_migration_target, NULL, in do_move_pages_to_node()
1687 if (err) in do_move_pages_to_node()
1689 return err; in do_move_pages_to_node()
1707 int err; in add_page_for_migration() local
1710 err = -EFAULT; in add_page_for_migration()
1719 err = PTR_ERR(page); in add_page_for_migration()
1723 err = -ENOENT; in add_page_for_migration()
1727 err = 0; in add_page_for_migration()
1731 err = -EACCES; in add_page_for_migration()
[all …]
Dfilemap.c705 int err = 0; in filemap_write_and_wait_range() local
708 err = __filemap_fdatawrite_range(mapping, lstart, lend, in filemap_write_and_wait_range()
716 if (err != -EIO) { in filemap_write_and_wait_range()
719 if (!err) in filemap_write_and_wait_range()
720 err = err2; in filemap_write_and_wait_range()
726 err = filemap_check_errors(mapping); in filemap_write_and_wait_range()
728 return err; in filemap_write_and_wait_range()
732 void __filemap_set_wb_err(struct address_space *mapping, int err) in __filemap_set_wb_err() argument
734 errseq_t eseq = errseq_set(&mapping->wb_err, err); in __filemap_set_wb_err()
766 int err = 0; in file_check_and_advance_wb_err() local
[all …]
Dhwpoison-inject.c19 int err; in hwpoison_inject() local
45 err = hwpoison_filter(hpage); in hwpoison_inject()
46 if (err) in hwpoison_inject()
51 err = memory_failure(pfn, 0); in hwpoison_inject()
52 return (err == -EOPNOTSUPP) ? 0 : err; in hwpoison_inject()
Dpercpu-vm.c97 goto err; in pcpu_alloc_pages()
102 err: in pcpu_alloc_pages()
218 int i, err; in pcpu_map_pages() local
221 err = __pcpu_map_pages(pcpu_chunk_addr(chunk, cpu, page_start), in pcpu_map_pages()
224 if (err < 0) in pcpu_map_pages()
225 goto err; in pcpu_map_pages()
232 err: in pcpu_map_pages()
240 return err; in pcpu_map_pages()
Dcma_sysfs.c78 int i, err; in cma_sysfs_init() local
87 err = -ENOMEM; in cma_sysfs_init()
94 err = kobject_init_and_add(&cma_kobj->kobj, &cma_ktype, in cma_sysfs_init()
96 if (err) { in cma_sysfs_init()
110 return err; in cma_sysfs_init()
Dmemory.c1203 int err; in copy_pmd_range() local
1205 err = copy_huge_pmd(dst_mm, src_mm, dst_pmd, src_pmd, in copy_pmd_range()
1207 if (err == -ENOMEM) in copy_pmd_range()
1209 if (!err) in copy_pmd_range()
1239 int err; in copy_pud_range() local
1242 err = copy_huge_pud(dst_mm, src_mm, in copy_pud_range()
1244 if (err == -ENOMEM) in copy_pud_range()
1246 if (!err) in copy_pud_range()
1864 int err; in insert_page_in_batch_locked() local
1868 err = validate_page_before_insert(page); in insert_page_in_batch_locked()
[all …]
Dmremap.c630 int err = 0; in move_vma() local
642 err = vma->vm_ops->may_split(vma, old_addr); in move_vma()
643 if (!err && vma->vm_end != old_addr + old_len) in move_vma()
644 err = vma->vm_ops->may_split(vma, old_addr + old_len); in move_vma()
645 if (err) in move_vma()
646 return err; in move_vma()
656 err = ksm_madvise(vma, old_addr, old_addr + old_len, in move_vma()
658 if (err) in move_vma()
659 return err; in move_vma()
678 err = -ENOMEM; in move_vma()
[all …]
Dswap_state.c187 int err; in add_to_swap() local
207 err = add_to_swap_cache(page, entry, in add_to_swap()
209 if (err) in add_to_swap()
428 int err; in __read_swap_cache_async() local
466 err = swapcache_prepare(entry); in __read_swap_cache_async()
467 if (!err) in __read_swap_cache_async()
471 if (err != -EEXIST) in __read_swap_cache_async()
892 int err; in swap_init_sysfs() local
900 err = sysfs_create_group(swap_kobj, &swap_attr_group); in swap_init_sysfs()
901 if (err) { in swap_init_sysfs()
[all …]
Dkhugepaged.c137 int err; in scan_sleep_millisecs_store() local
139 err = kstrtouint(buf, 10, &msecs); in scan_sleep_millisecs_store()
140 if (err) in scan_sleep_millisecs_store()
165 int err; in alloc_sleep_millisecs_store() local
167 err = kstrtouint(buf, 10, &msecs); in alloc_sleep_millisecs_store()
168 if (err) in alloc_sleep_millisecs_store()
192 int err; in pages_to_scan_store() local
194 err = kstrtouint(buf, 10, &pages); in pages_to_scan_store()
195 if (err || !pages) in pages_to_scan_store()
259 int err; in khugepaged_max_ptes_none_store() local
[all …]
Dpage_idle.c212 int err; in page_idle_init() local
214 err = sysfs_create_group(mm_kobj, &page_idle_attr_group); in page_idle_init()
215 if (err) { in page_idle_init()
217 return err; in page_idle_init()
Dswapfile.c175 int err = 0; in discard_swap() local
182 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap()
184 if (err) in discard_swap()
185 return err; in discard_swap()
193 err = blkdev_issue_discard(si->bdev, start_block, in discard_swap()
195 if (err) in discard_swap()
200 return err; /* That will often be -EOPNOTSUPP */ in discard_swap()
2531 int err, found = 0; in SYSCALL_DEFINE1() local
2545 err = PTR_ERR(victim); in SYSCALL_DEFINE1()
2560 err = -EINVAL; in SYSCALL_DEFINE1()
[all …]
Dmincore.c189 int err; in do_mincore() local
200 err = walk_page_range(vma->vm_mm, addr, end, &mincore_walk_ops, vec); in do_mincore()
201 if (err < 0) in do_mincore()
202 return err; in do_mincore()
Dcma.c287 goto err; in cma_declare_contiguous_nid()
308 goto err; in cma_declare_contiguous_nid()
323 goto err; in cma_declare_contiguous_nid()
331 goto err; in cma_declare_contiguous_nid()
370 goto err; in cma_declare_contiguous_nid()
392 err: in cma_declare_contiguous_nid()
Dpage_pinner.c218 goto err; in print_page_pinner()
230 goto err; in print_page_pinner()
236 goto err; in print_page_pinner()
240 goto err; in print_page_pinner()
248 err: in print_page_pinner()
Dslub.c1807 int err; in init_cache_random_seq() local
1813 err = cache_random_seq_create(s, count, GFP_KERNEL); in init_cache_random_seq()
1814 if (err) { in init_cache_random_seq()
1817 return err; in init_cache_random_seq()
4910 int err; in __kmem_cache_create() local
4912 err = kmem_cache_open(s, flags); in __kmem_cache_create()
4913 if (err) in __kmem_cache_create()
4914 return err; in __kmem_cache_create()
4920 err = sysfs_slab_add(s); in __kmem_cache_create()
4921 if (err) { in __kmem_cache_create()
[all …]
Dvmalloc.c293 int err; in vmap_range_noflush() local
303 err = vmap_p4d_range(pgd, addr, next, phys_addr, prot, in vmap_range_noflush()
305 if (err) in vmap_range_noflush()
312 return err; in vmap_range_noflush()
318 int err; in ioremap_page_range() local
321 err = vmap_range_noflush(addr, end, phys_addr, prot, in ioremap_page_range()
325 if (IS_ENABLED(CONFIG_ARCH_HAS_IOREMAP_PHYS_HOOKS) && !err) in ioremap_page_range()
328 return err; in ioremap_page_range()
553 int err = 0; in vmap_small_pages_range_noflush() local
563 err = vmap_pages_p4d_range(pgd, addr, next, prot, pages, &nr, &mask); in vmap_small_pages_range_noflush()
[all …]
/mm/damon/
Dreclaim.c296 int err; in damon_reclaim_turn() local
299 err = damon_stop(&ctx, 1); in damon_reclaim_turn()
300 if (!err) in damon_reclaim_turn()
302 return err; in damon_reclaim_turn()
305 err = damon_set_attrs(ctx, sample_interval, aggr_interval, 0, in damon_reclaim_turn()
307 if (err) in damon_reclaim_turn()
308 return err; in damon_reclaim_turn()
325 err = -ENOMEM; in damon_reclaim_turn()
328 err = damon_set_schemes(ctx, &scheme, 1); in damon_reclaim_turn()
329 if (err) in damon_reclaim_turn()
[all …]
Dcore.c382 int err = -EBUSY; in __damon_start() local
386 err = 0; in __damon_start()
390 err = PTR_ERR(ctx->kdamond); in __damon_start()
396 return err; in __damon_start()
415 int err = 0; in damon_start() local
424 err = __damon_start(ctxs[i]); in damon_start()
425 if (err) in damon_start()
431 return err; in damon_start()
476 int i, err = 0; in damon_stop() local
480 err = __damon_stop(ctxs[i]); in damon_stop()
[all …]

12