/mm/ |
D | util.c | 57 size_t len; in kstrdup() local 63 len = strlen(s) + 1; in kstrdup() 64 buf = kmalloc_track_caller(len, gfp); in kstrdup() 66 memcpy(buf, s, len); in kstrdup() 103 size_t len; in kstrndup() local 109 len = strnlen(s, max); in kstrndup() 110 buf = kmalloc_track_caller(len+1, gfp); in kstrndup() 112 memcpy(buf, s, len); in kstrndup() 113 buf[len] = '\0'; in kstrndup() 128 void *kmemdup(const void *src, size_t len, gfp_t gfp) in kmemdup() argument [all …]
|
D | nommu.c | 723 unsigned long len) in find_vma_exact() argument 726 unsigned long end = addr + len; in find_vma_exact() 755 unsigned long len, in validate_mmap_request() argument 772 if (!len) in validate_mmap_request() 776 rlen = PAGE_ALIGN(len); in validate_mmap_request() 972 unsigned long len, in do_mmap_private() argument 1004 order = get_order(len); in do_mmap_private() 1006 point = len >> PAGE_SHIFT; in do_mmap_private() 1020 region->vm_end = region->vm_start + len; in do_mmap_private() 1024 vma->vm_end = region->vm_start + len; in do_mmap_private() [all …]
|
D | fadvise.c | 32 int generic_fadvise(struct file *file, loff_t offset, loff_t len, int advice) in generic_fadvise() argument 47 if (!mapping || len < 0) in generic_fadvise() 73 endbyte = (u64)offset + (u64)len; in generic_fadvise() 74 if (!len || endbyte < len) in generic_fadvise() 181 int vfs_fadvise(struct file *file, loff_t offset, loff_t len, int advice) in vfs_fadvise() argument 184 return file->f_op->fadvise(file, offset, len, advice); in vfs_fadvise() 186 return generic_fadvise(file, offset, len, advice); in vfs_fadvise() 192 int ksys_fadvise64_64(int fd, loff_t offset, loff_t len, int advice) in ksys_fadvise64_64() argument 200 ret = vfs_fadvise(f.file, offset, len, advice); in ksys_fadvise64_64() 206 SYSCALL_DEFINE4(fadvise64_64, int, fd, loff_t, offset, loff_t, len, int, advice) in SYSCALL_DEFINE4() argument [all …]
|
D | userfaultfd.c | 25 unsigned long len) in find_dst_vma() argument 38 dst_start + len > dst_vma->vm_end) in find_dst_vma() 292 unsigned long len, in __mcopy_atomic_hugetlb() argument 327 if (dst_start & (vma_hpagesize - 1) || len & (vma_hpagesize - 1)) in __mcopy_atomic_hugetlb() 337 dst_vma = find_dst_vma(dst_mm, dst_start, len); in __mcopy_atomic_hugetlb() 357 while (src_addr < src_start + len) { in __mcopy_atomic_hugetlb() 358 BUG_ON(dst_addr >= dst_start + len); in __mcopy_atomic_hugetlb() 443 unsigned long len, 495 unsigned long len, in __mcopy_atomic() argument 512 BUG_ON(len & ~PAGE_MASK); in __mcopy_atomic() [all …]
|
D | mmap.c | 65 #define arch_mmap_check(addr, len, flags) (0) argument 599 munmap_vma_range(struct mm_struct *mm, unsigned long start, unsigned long len, in munmap_vma_range() argument 604 while (find_vma_links(mm, start, start + len, pprev, link, parent)) in munmap_vma_range() 605 if (do_munmap(mm, start, len, uf)) in munmap_vma_range() 1362 unsigned long len) in mlock_future_check() argument 1368 locked = len >> PAGE_SHIFT; in mlock_future_check() 1398 unsigned long pgoff, unsigned long len) in file_mmap_ok() argument 1402 if (maxsize && len > maxsize) in file_mmap_ok() 1404 maxsize -= len; in file_mmap_ok() 1414 unsigned long len, unsigned long prot, in do_mmap() argument [all …]
|
D | mlock.c | 559 static int apply_vma_lock_flags(unsigned long start, size_t len, in apply_vma_lock_flags() argument 567 VM_BUG_ON(len != PAGE_ALIGN(len)); in apply_vma_lock_flags() 568 end = start + len; in apply_vma_lock_flags() 616 unsigned long start, size_t len) in count_mm_mlocked_page_nr() argument 631 if (start + len <= vma->vm_start) in count_mm_mlocked_page_nr() 636 if (start + len < vma->vm_end) { in count_mm_mlocked_page_nr() 637 count += start + len - vma->vm_start; in count_mm_mlocked_page_nr() 647 static __must_check int do_mlock(unsigned long start, size_t len, vm_flags_t flags) in do_mlock() argument 658 len = PAGE_ALIGN(len + (offset_in_page(start))); in do_mlock() 663 locked = len >> PAGE_SHIFT; in do_mlock() [all …]
|
D | process_vm_access.c | 29 size_t len, in process_vm_rw_pages() argument 34 while (len && iov_iter_count(iter)) { in process_vm_rw_pages() 39 if (copy > len) in process_vm_rw_pages() 40 copy = len; in process_vm_rw_pages() 47 len -= copied; in process_vm_rw_pages() 71 unsigned long len, in process_vm_rw_single_vec() argument 87 if (len == 0) in process_vm_rw_single_vec() 89 nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1; in process_vm_rw_single_vec() 114 if (bytes > len) in process_vm_rw_single_vec() 115 bytes = len; in process_vm_rw_single_vec() [all …]
|
D | usercopy.c | 34 static noinline int check_stack_object(const void *obj, unsigned long len) in check_stack_object() argument 41 if (obj + len <= stack || stackend <= obj) in check_stack_object() 49 if (obj < stack || stackend < obj + len) in check_stack_object() 53 ret = arch_within_stack_frames(stack, stackend, obj, len); in check_stack_object() 73 unsigned long offset, unsigned long len) in usercopy_warn() argument 80 offset, len); in usercopy_warn() 85 unsigned long len) in usercopy_abort() argument 92 offset, len); in usercopy_abort()
|
D | memfd.c | 274 long len; in SYSCALL_DEFINE2() local 287 len = strnlen_user(uname, MFD_NAME_MAX_LEN + 1); in SYSCALL_DEFINE2() 288 if (len <= 0) in SYSCALL_DEFINE2() 290 if (len > MFD_NAME_MAX_LEN + 1) in SYSCALL_DEFINE2() 293 name = kmalloc(len + MFD_NAME_PREFIX_LEN, GFP_KERNEL); in SYSCALL_DEFINE2() 298 if (copy_from_user(&name[MFD_NAME_PREFIX_LEN], uname, len)) { in SYSCALL_DEFINE2() 304 if (name[len + MFD_NAME_PREFIX_LEN - 1]) { in SYSCALL_DEFINE2()
|
D | maccess.c | 17 #define copy_from_kernel_nofault_loop(dst, src, len, type, err_label) \ argument 18 while (len >= sizeof(type)) { \ 22 len -= sizeof(type); \ 51 #define copy_to_kernel_nofault_loop(dst, src, len, type, err_label) \ argument 52 while (len >= sizeof(type)) { \ 56 len -= sizeof(type); \
|
D | shmem.c | 2127 unsigned long uaddr, unsigned long len, in shmem_get_unmapped_area() argument 2138 if (len > TASK_SIZE) in shmem_get_unmapped_area() 2142 addr = get_area(file, uaddr, len, pgoff, flags); in shmem_get_unmapped_area() 2150 if (addr > TASK_SIZE - len) in shmem_get_unmapped_area() 2155 if (len < HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2188 if (offset && offset + len < 2 * HPAGE_PMD_SIZE) in shmem_get_unmapped_area() 2193 inflated_len = len + HPAGE_PMD_SIZE - PAGE_SIZE; in shmem_get_unmapped_area() 2196 if (inflated_len < len) in shmem_get_unmapped_area() 2210 if (inflated_addr > TASK_SIZE - len) in shmem_get_unmapped_area() 2461 loff_t pos, unsigned len, unsigned flags, in shmem_write_begin() argument [all …]
|
D | msync.c | 32 SYSCALL_DEFINE3(msync, unsigned long, start, size_t, len, int, flags) in SYSCALL_DEFINE3() argument 49 len = (len + ~PAGE_MASK) & PAGE_MASK; in SYSCALL_DEFINE3() 50 end = start + len; in SYSCALL_DEFINE3()
|
D | mprotect.c | 528 static int do_mprotect_pkey(unsigned long start, size_t len, in do_mprotect_pkey() argument 547 if (!len) in do_mprotect_pkey() 549 len = PAGE_ALIGN(len); in do_mprotect_pkey() 550 end = start + len; in do_mprotect_pkey() 668 SYSCALL_DEFINE3(mprotect, unsigned long, start, size_t, len, in SYSCALL_DEFINE3() argument 671 return do_mprotect_pkey(start, len, prot, -1); in SYSCALL_DEFINE3() 676 SYSCALL_DEFINE4(pkey_mprotect, unsigned long, start, size_t, len, in SYSCALL_DEFINE4() argument 679 return do_mprotect_pkey(start, len, prot, pkey); in SYSCALL_DEFINE4()
|
D | cleancache.c | 148 int len = 0, maxlen = CLEANCACHE_KEY_MAX; in cleancache_get_key() local 155 len = (*fhfn)(inode, &key->u.fh[0], &maxlen, NULL); in cleancache_get_key() 156 if (len <= FILEID_ROOT || len == FILEID_INVALID) in cleancache_get_key()
|
D | mincore.c | 230 SYSCALL_DEFINE3(mincore, unsigned long, start, size_t, len, in SYSCALL_DEFINE3() argument 244 if (!access_ok((void __user *) start, len)) in SYSCALL_DEFINE3() 248 pages = len >> PAGE_SHIFT; in SYSCALL_DEFINE3() 249 pages += (offset_in_page(len)) != 0; in SYSCALL_DEFINE3()
|
D | madvise.c | 1254 unsigned long len; in madvise_set_anon_name() local 1258 len = (len_in + ~PAGE_MASK) & PAGE_MASK; in madvise_set_anon_name() 1261 if (len_in && !len) in madvise_set_anon_name() 1264 end = start + len; in madvise_set_anon_name() 1350 size_t len; in do_madvise() local 1360 len = PAGE_ALIGN(len_in); in do_madvise() 1363 if (len_in && !len) in do_madvise() 1366 end = start + len; in do_madvise()
|
D | slub.c | 1577 size_t len; in kmem_cache_flags() local 1590 len = strlen(name); in kmem_cache_flags() 1610 cmplen = max_t(size_t, len, (end - iter)); in kmem_cache_flags() 5254 int len = 0; in show_slab_objects() local 5344 len += sysfs_emit_at(buf, len, "%lu", total); in show_slab_objects() 5348 len += sysfs_emit_at(buf, len, " N%d=%lu", in show_slab_objects() 5352 len += sysfs_emit_at(buf, len, "\n"); in show_slab_objects() 5355 return len; in show_slab_objects() 5491 int len = 0; in slabs_cpu_partial_show() local 5504 len += sysfs_emit_at(buf, len, "%d(%d)", objects, pages); in slabs_cpu_partial_show() [all …]
|
D | mremap.c | 143 unsigned long len = old_end - old_addr; in move_ptes() local 204 flush_tlb_range(vma, old_end - len, old_end); in move_ptes() 532 unsigned long new_addr, unsigned long len, in move_page_tables() argument 540 if (!len) in move_page_tables() 543 old_end = old_addr + len; in move_page_tables() 613 return len + old_addr - old_end; /* how much done */ in move_page_tables()
|
D | kmemleak.c | 266 size_t len, bool ascii) in warn_or_seq_hex_dump() argument 270 buf, len, ascii); in warn_or_seq_hex_dump() 273 rowsize, groupsize, buf, len, ascii); in warn_or_seq_hex_dump() 286 size_t len; in hex_dump_object() local 289 len = min_t(size_t, object->size, HEX_MAX_LINES * HEX_ROW_SIZE); in hex_dump_object() 291 warn_or_seq_printf(seq, " hex dump (first %zu bytes):\n", len); in hex_dump_object() 294 HEX_GROUP_SIZE, kasan_reset_tag((void *)ptr), len, HEX_ASCII); in hex_dump_object()
|
D | secretmem.c | 123 unsigned long len = vma->vm_end - vma->vm_start; in secretmem_mmap() local 128 if (mlock_future_check(vma->vm_mm, vma->vm_flags | VM_LOCKED, len)) in secretmem_mmap()
|
D | mempolicy.c | 1252 static long do_mbind(unsigned long start, unsigned long len, in do_mbind() argument 1274 len = (len + PAGE_SIZE - 1) & PAGE_MASK; in do_mbind() 1275 end = start + len; in do_mbind() 1297 start, start + len, mode, mode_flags, in do_mbind() 1462 static long kernel_mbind(unsigned long start, unsigned long len, in kernel_mbind() argument 1480 return do_mbind(start, len, lmode, mode_flags, &nodes, flags); in kernel_mbind() 1483 SYSCALL_DEFINE6(mbind, unsigned long, start, unsigned long, len, in SYSCALL_DEFINE6() argument 1487 return kernel_mbind(start, len, mode, nmask, maxnode, flags); in SYSCALL_DEFINE6()
|
/mm/damon/ |
D | dbgfs.c | 100 static ssize_t sprint_schemes(struct damon_ctx *c, char *buf, ssize_t len) in sprint_schemes() argument 107 rc = scnprintf(&buf[written], len - written, in sprint_schemes() 136 ssize_t len; in dbgfs_schemes_read() local 143 len = sprint_schemes(ctx, kbuf, count); in dbgfs_schemes_read() 145 if (len < 0) in dbgfs_schemes_read() 147 len = simple_read_from_buffer(buf, count, ppos, kbuf, len); in dbgfs_schemes_read() 151 return len; in dbgfs_schemes_read() 184 static struct damos **str_to_schemes(const char *str, ssize_t len, in str_to_schemes() argument 200 while (pos < len && *nr_schemes < max_nr_schemes) { in str_to_schemes() 283 static ssize_t sprint_target_ids(struct damon_ctx *ctx, char *buf, ssize_t len) in sprint_target_ids() argument [all …]
|
/mm/kasan/ |
D | shadow.c | 42 void *memset(void *addr, int c, size_t len) in memset() argument 44 if (!kasan_check_range((unsigned long)addr, len, true, _RET_IP_)) in memset() 47 return __memset(addr, c, len); in memset() 52 void *memmove(void *dest, const void *src, size_t len) in memmove() argument 54 if (!kasan_check_range((unsigned long)src, len, false, _RET_IP_) || in memmove() 55 !kasan_check_range((unsigned long)dest, len, true, _RET_IP_)) in memmove() 58 return __memmove(dest, src, len); in memmove() 63 void *memcpy(void *dest, const void *src, size_t len) in memcpy() argument 65 if (!kasan_check_range((unsigned long)src, len, false, _RET_IP_) || in memcpy() 66 !kasan_check_range((unsigned long)dest, len, true, _RET_IP_)) in memcpy() [all …]
|
/mm/kfence/ |
D | kfence_test.c | 45 static void probe_console(void *ignore, const char *buf, size_t len) in probe_console() argument 53 if (strnstr(buf, "BUG: KFENCE: ", len) && strnstr(buf, "test_", len)) { in probe_console() 60 strscpy(observed.lines[0], buf, min(len + 1, sizeof(observed.lines[0]))); in probe_console() 62 } else if (nlines == 1 && (strnstr(buf, "at 0x", len) || strnstr(buf, "of 0x", len))) { in probe_console() 63 strscpy(observed.lines[nlines++], buf, min(len + 1, sizeof(observed.lines[0]))); in probe_console()
|
D | report.c | 74 int len = scnprintf(buf, sizeof(buf), "%ps", (void *)stack_entries[skipnr]); in get_stack_skipnr() local 78 !strncmp(buf, ARCH_FUNC_PREFIX "__slab_free", len)) { in get_stack_skipnr()
|