/fs/proc/ |
D | task_nommu.c | 37 size += region->vm_end - region->vm_start; in task_mem() 39 size = vma->vm_end - vma->vm_start; in task_mem() 92 vsize += vma->vm_end - vma->vm_start; in task_vsize() 114 size += region->vm_end - region->vm_start; in task_statm() 138 return vma->vm_start <= mm->start_stack && in is_stack() 167 vma->vm_start, in nommu_vma_show()
|
D | task_mmu.c | 227 *ppos = next ? next->vm_start : -1UL; in m_next() 296 return vma->vm_start <= vma->vm_mm->start_stack && in is_stack() 339 start = vma->vm_start; in show_map_vma() 366 if (vma->vm_start <= mm->brk && in show_map_vma() 904 SEQ_PUT_DEC("Size: ", vma->vm_end - vma->vm_start); in show_smap() 1006 if (vma->vm_start >= last_vma_end) in show_smaps_rollup() 1017 show_vma_header_prefix(m, priv->mm->mmap ? priv->mm->mmap->vm_start : 0, in show_smaps_rollup() 1421 hole_end = min(end, vma->vm_start); in pagemap_pte_hole() 1966 pol = __get_vma_policy(vma, vma->vm_start); in show_numa_map() 1974 seq_printf(m, "%08lx %s", vma->vm_start, buffer); in show_numa_map() [all …]
|
D | vmcore.c | 556 size_t size = vma->vm_end - vma->vm_start; in mmap_vmcore() 580 if (remap_pfn_range(vma, vma->vm_start, pfn, tsz, in mmap_vmcore() 613 if (vmcoredd_mmap_dumps(vma, vma->vm_start + len, in mmap_vmcore() 630 if (remap_vmalloc_range_partial(vma, vma->vm_start + len, in mmap_vmcore() 649 if (vmcore_remap_oldmem_pfn(vma, vma->vm_start + len, in mmap_vmcore() 664 do_munmap(vma->vm_mm, vma->vm_start, len, NULL); in mmap_vmcore()
|
D | nommu.c | 51 region->vm_start, in nommu_region_show()
|
D | base.c | 2138 unsigned long vm_start, vm_end; in map_files_d_revalidate() local 2157 if (!dname_to_vma_addr(dentry, &vm_start, &vm_end)) { in map_files_d_revalidate() 2160 exact_vma_exists = !!find_exact_vma(mm, vm_start, in map_files_d_revalidate() 2189 unsigned long vm_start, vm_end; in map_files_get_link() local 2205 rc = dname_to_vma_addr(dentry, &vm_start, &vm_end); in map_files_get_link() 2214 vma = find_exact_vma(mm, vm_start, vm_end); in map_files_get_link() 2286 unsigned long vm_start, vm_end; in proc_map_files_lookup() local 2302 if (dname_to_vma_addr(dentry, &vm_start, &vm_end)) in proc_map_files_lookup() 2314 vma = find_exact_vma(mm, vm_start, vm_end); in proc_map_files_lookup() 2400 p->start = vma->vm_start; in proc_map_files_readdir()
|
/fs/ |
D | userfaultfd.c | 837 for ( ; vma && vma->vm_start < end; vma = vma->vm_next) { in userfaultfd_unmap_prep() 917 prev = vma_merge(mm, prev, vma->vm_start, vma->vm_end, in userfaultfd_release() 1372 if (vma->vm_start >= end) in userfaultfd_register() 1391 for (cur = vma; cur && cur->vm_start < end; cur = cur->vm_next) { in userfaultfd_register() 1422 end > cur->vm_start) { in userfaultfd_register() 1453 if (vma->vm_start < start) in userfaultfd_register() 1475 if (vma->vm_start > start) in userfaultfd_register() 1476 start = vma->vm_start; in userfaultfd_register() 1489 if (vma->vm_start < start) { in userfaultfd_register() 1517 } while (vma && vma->vm_start < end); in userfaultfd_register() [all …]
|
D | exec.c | 274 vma->vm_start = vma->vm_end - PAGE_SIZE; in __bprm_mm_init() 687 unsigned long old_start = vma->vm_start; in shift_arg_pages() 724 vma->vm_next ? vma->vm_next->vm_start : USER_PGTABLES_CEILING); in shift_arg_pages() 733 vma->vm_next ? vma->vm_next->vm_start : USER_PGTABLES_CEILING); in shift_arg_pages() 774 if (vma->vm_end - vma->vm_start > stack_base) in setup_arg_pages() 779 stack_shift = vma->vm_start - stack_base; in setup_arg_pages() 787 unlikely(vma->vm_end - vma->vm_start >= stack_top - mmap_min_addr)) in setup_arg_pages() 817 ret = mprotect_fixup(vma, &prev, vma->vm_start, vma->vm_end, in setup_arg_pages() 839 stack_size = vma->vm_end - vma->vm_start; in setup_arg_pages() 847 stack_base = vma->vm_start + rlim_stack; in setup_arg_pages() [all …]
|
D | coredump.c | 1054 return vma->vm_end - vma->vm_start; in vma_dump_size() 1118 m->start = vma->vm_start; in dump_vma_snapshot()
|
D | dax.c | 798 address = vma->vm_start + ((pgoff - vma->vm_pgoff) << PAGE_SHIFT); in pgoff_address() 799 VM_BUG_ON_VMA(address < vma->vm_start || address >= vma->vm_end, vma); in pgoff_address() 1526 if (pmd_addr < vma->vm_start) in dax_iomap_pmd_fault()
|
D | aio.c | 347 ctx->user_id = ctx->mmap_base = vma->vm_start; in aio_ring_mremap()
|
D | binfmt_elf.c | 1673 *start_end_ofs++ = vma->vm_start; in fill_files_note()
|
/fs/cramfs/ |
D | inode.c | 412 ret = remap_pfn_range(vma, vma->vm_start, address >> PAGE_SHIFT, in cramfs_physmem_mmap() 426 vmf = vmf_insert_mixed(vma, vma->vm_start + off, pfn); in cramfs_physmem_mmap() 436 address, pages, vma_pages(vma), vma->vm_start, in cramfs_physmem_mmap()
|
/fs/coda/ |
D | file.c | 142 count = vma->vm_end - vma->vm_start; in coda_file_mmap()
|
/fs/hugetlbfs/ |
D | inode.c | 170 vma_len = (loff_t)(vma->vm_end - vma->vm_start); in hugetlbfs_file_mmap() 446 + vma->vm_start; in hugetlb_vmdelete_list() 451 unmap_hugepage_range(vma, vma->vm_start + v_offset, v_end, in hugetlb_vmdelete_list()
|
/fs/9p/ |
D | vfs_file.c | 618 (vma->vm_end - vma->vm_start - 1), in v9fs_mmap_vm_close()
|