/fs/proc/ |
D | task_mmu.c | 84 static void vma_stop(struct proc_maps_private *priv, struct vm_area_struct *vma) in vma_stop() argument 86 if (vma && vma != priv->tail_vma) { in vma_stop() 87 struct mm_struct *mm = vma->vm_mm; in vma_stop() 98 struct vm_area_struct *vma, *tail_vma = NULL; in m_start() local 127 vma = find_vma(mm, last_addr); in m_start() 128 if (last_addr && vma) { in m_start() 129 vma = vma->vm_next; in m_start() 137 vma = NULL; in m_start() 139 vma = mm->mmap; in m_start() 140 while (l-- && vma) in m_start() [all …]
|
D | task_nommu.c | 18 struct vm_area_struct *vma; in task_mem() local 25 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_mem() 27 bytes += kobjsize(vma); in task_mem() 29 region = vma->vm_region; in task_mem() 34 size = vma->vm_end - vma->vm_start; in task_mem() 38 vma->vm_flags & VM_MAYSHARE) { in task_mem() 43 slack = region->vm_end - vma->vm_end; in task_mem() 80 struct vm_area_struct *vma; in task_vsize() local 86 vma = rb_entry(p, struct vm_area_struct, vm_rb); in task_vsize() 87 vsize += vma->vm_end - vma->vm_start; in task_vsize() [all …]
|
D | mmu.c | 19 struct vm_struct *vma; in get_vmalloc_info() local 35 for (vma = vmlist; vma; vma = vma->next) { in get_vmalloc_info() 36 unsigned long addr = (unsigned long) vma->addr; in get_vmalloc_info() 46 vmi->used += vma->size; in get_vmalloc_info() 52 prev_end = vma->size + addr; in get_vmalloc_info()
|
D | inode.c | 293 static int proc_reg_mmap(struct file *file, struct vm_area_struct *vma) in proc_reg_mmap() argument 309 rv = mmap(file, vma); in proc_reg_mmap()
|
/fs/ |
D | exec.c | 179 ret = expand_stack_downwards(bprm->vma, pos); in get_arg_page() 190 unsigned long size = bprm->vma->vm_end - bprm->vma->vm_start; in get_arg_page() 233 flush_cache_page(bprm->vma, pos, page_to_pfn(page)); in flush_arg_page() 239 struct vm_area_struct *vma = NULL; in __bprm_mm_init() local 242 bprm->vma = vma = kmem_cache_zalloc(vm_area_cachep, GFP_KERNEL); in __bprm_mm_init() 243 if (!vma) in __bprm_mm_init() 247 vma->vm_mm = mm; in __bprm_mm_init() 255 vma->vm_end = STACK_TOP_MAX; in __bprm_mm_init() 256 vma->vm_start = vma->vm_end - PAGE_SIZE; in __bprm_mm_init() 257 vma->vm_flags = VM_STACK_FLAGS; in __bprm_mm_init() [all …]
|
D | binfmt_elf_fdpic.c | 1220 static int maydump(struct vm_area_struct *vma, unsigned long mm_flags) in maydump() argument 1225 if (vma->vm_flags & (VM_IO | VM_RESERVED)) { in maydump() 1226 kdcore("%08lx: %08lx: no (IO)", vma->vm_start, vma->vm_flags); in maydump() 1233 if (!(vma->vm_flags & VM_READ)) { in maydump() 1234 kdcore("%08lx: %08lx: no (!read)", vma->vm_start, vma->vm_flags); in maydump() 1239 if (vma->vm_flags & VM_SHARED) { in maydump() 1240 if (vma->vm_file->f_path.dentry->d_inode->i_nlink == 0) { in maydump() 1242 kdcore("%08lx: %08lx: %s (share)", vma->vm_start, in maydump() 1243 vma->vm_flags, dump_ok ? "yes" : "no"); in maydump() 1248 kdcore("%08lx: %08lx: %s (share)", vma->vm_start, in maydump() [all …]
|
D | binfmt_elf.c | 163 struct vm_area_struct *vma; in create_elf_tables() local 281 vma = find_extend_vma(current->mm, bprm->p); in create_elf_tables() 282 if (!vma) in create_elf_tables() 1168 static unsigned long vma_dump_size(struct vm_area_struct *vma, in vma_dump_size() argument 1174 if (vma->vm_flags & VM_ALWAYSDUMP) in vma_dump_size() 1178 if (vma->vm_flags & VM_HUGETLB) { in vma_dump_size() 1179 if ((vma->vm_flags & VM_SHARED) && FILTER(HUGETLB_SHARED)) in vma_dump_size() 1181 if (!(vma->vm_flags & VM_SHARED) && FILTER(HUGETLB_PRIVATE)) in vma_dump_size() 1186 if (vma->vm_flags & (VM_IO | VM_RESERVED)) in vma_dump_size() 1190 if (vma->vm_flags & VM_SHARED) { in vma_dump_size() [all …]
|
D | bad_inode.c | 76 static int bad_file_mmap(struct file *file, struct vm_area_struct *vma) in bad_file_mmap() argument
|
/fs/hugetlbfs/ |
D | inode.c | 80 static int hugetlbfs_file_mmap(struct file *file, struct vm_area_struct *vma) in hugetlbfs_file_mmap() argument 95 vma->vm_flags |= VM_HUGETLB | VM_RESERVED; in hugetlbfs_file_mmap() 96 vma->vm_ops = &hugetlb_vm_ops; in hugetlbfs_file_mmap() 98 if (vma->vm_pgoff & ~(huge_page_mask(h) >> PAGE_SHIFT)) in hugetlbfs_file_mmap() 101 vma_len = (loff_t)(vma->vm_end - vma->vm_start); in hugetlbfs_file_mmap() 107 len = vma_len + ((loff_t)vma->vm_pgoff << PAGE_SHIFT); in hugetlbfs_file_mmap() 110 vma->vm_pgoff >> huge_page_order(h), in hugetlbfs_file_mmap() 111 len >> huge_page_shift(h), vma, in hugetlbfs_file_mmap() 112 vma->vm_flags)) in hugetlbfs_file_mmap() 116 hugetlb_prefault_arch_hook(vma->vm_mm); in hugetlbfs_file_mmap() [all …]
|
/fs/ncpfs/ |
D | mmap.c | 105 int ncp_mmap(struct file *file, struct vm_area_struct *vma) in ncp_mmap() argument 115 if (vma->vm_flags & VM_SHARED) in ncp_mmap() 119 if (((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff in ncp_mmap() 123 vma->vm_ops = &ncp_file_mmap; in ncp_mmap()
|
/fs/ocfs2/ |
D | mmap.c | 157 static int ocfs2_page_mkwrite(struct vm_area_struct *vma, struct page *page) in ocfs2_page_mkwrite() argument 159 struct inode *inode = vma->vm_file->f_path.dentry->d_inode; in ocfs2_page_mkwrite() 208 int ocfs2_mmap(struct file *file, struct vm_area_struct *vma) in ocfs2_mmap() argument 220 vma->vm_ops = &ocfs2_file_vm_ops; in ocfs2_mmap() 221 vma->vm_flags |= VM_CAN_NONLINEAR; in ocfs2_mmap()
|
D | mmap.h | 4 int ocfs2_mmap(struct file *file, struct vm_area_struct *vma);
|
/fs/ramfs/ |
D | file-nommu.c | 146 struct vm_area_struct *vma; in ramfs_nommu_check_mappings() local 150 vma_prio_tree_foreach(vma, &iter, &inode->i_mapping->i_mmap, in ramfs_nommu_check_mappings() 156 if (vma->vm_flags & VM_SHARED) in ramfs_nommu_check_mappings() 300 int ramfs_nommu_mmap(struct file *file, struct vm_area_struct *vma) in ramfs_nommu_mmap() argument 302 if (!(vma->vm_flags & VM_SHARED)) in ramfs_nommu_mmap() 306 vma->vm_ops = &generic_file_vm_ops; in ramfs_nommu_mmap()
|
/fs/xfs/linux-2.6/ |
D | xfs_file.c | 219 struct vm_area_struct *vma) in xfs_file_mmap() argument 221 vma->vm_ops = &xfs_file_vm_ops; in xfs_file_mmap() 222 vma->vm_flags |= VM_CAN_NONLINEAR; in xfs_file_mmap() 236 struct vm_area_struct *vma, in xfs_vm_page_mkwrite() argument 239 return block_page_mkwrite(vma, page, xfs_get_blocks); in xfs_vm_page_mkwrite()
|
/fs/ext4/ |
D | file.c | 131 static int ext4_file_mmap(struct file *file, struct vm_area_struct *vma) in ext4_file_mmap() argument 138 vma->vm_ops = &ext4_file_vm_ops; in ext4_file_mmap() 139 vma->vm_flags |= VM_CAN_NONLINEAR; in ext4_file_mmap()
|
/fs/smbfs/ |
D | file.c | 245 smb_file_mmap(struct file * file, struct vm_area_struct * vma) in smb_file_mmap() argument 251 DENTRY_PATH(dentry), vma->vm_start, vma->vm_end); in smb_file_mmap() 259 status = generic_file_mmap(file, vma); in smb_file_mmap()
|
/fs/nfs/ |
D | file.c | 298 nfs_file_mmap(struct file * file, struct vm_area_struct * vma) in nfs_file_mmap() argument 309 vma->vm_ops = &nfs_file_vm_ops; in nfs_file_mmap() 310 vma->vm_flags |= VM_CAN_NONLINEAR; in nfs_file_mmap() 454 static int nfs_vm_page_mkwrite(struct vm_area_struct *vma, struct page *page) in nfs_vm_page_mkwrite() argument 456 struct file *filp = vma->vm_file; in nfs_vm_page_mkwrite()
|
/fs/reiserfs/ |
D | file.c | 113 static int reiserfs_file_mmap(struct file *file, struct vm_area_struct *vma) in reiserfs_file_mmap() argument 122 return generic_file_mmap(file, vma); in reiserfs_file_mmap()
|
/fs/coda/ |
D | file.c | 92 coda_file_mmap(struct file *coda_file, struct vm_area_struct *vma) in coda_file_mmap() argument 122 return host_file->f_op->mmap(host_file, vma); in coda_file_mmap()
|
/fs/sysfs/ |
D | bin.c | 171 static int mmap(struct file *file, struct vm_area_struct *vma) in mmap() argument 187 rc = attr->mmap(kobj, attr, vma); in mmap()
|
/fs/gfs2/ |
D | ops_file.c | 339 static int gfs2_page_mkwrite(struct vm_area_struct *vma, struct page *page) in gfs2_page_mkwrite() argument 341 struct inode *inode = vma->vm_file->f_path.dentry->d_inode; in gfs2_page_mkwrite() 429 static int gfs2_mmap(struct file *file, struct vm_area_struct *vma) in gfs2_mmap() argument 442 vma->vm_ops = &gfs2_vm_ops; in gfs2_mmap()
|
/fs/ubifs/ |
D | file.c | 1437 static int ubifs_vm_page_mkwrite(struct vm_area_struct *vma, struct page *page) in ubifs_vm_page_mkwrite() argument 1439 struct inode *inode = vma->vm_file->f_path.dentry->d_inode; in ubifs_vm_page_mkwrite() 1531 static int ubifs_file_mmap(struct file *file, struct vm_area_struct *vma) in ubifs_file_mmap() argument 1536 err = generic_file_mmap(file, vma); in ubifs_file_mmap() 1539 vma->vm_ops = &ubifs_file_vm_ops; in ubifs_file_mmap()
|
/fs/fuse/ |
D | file.c | 1217 static void fuse_vma_close(struct vm_area_struct *vma) in fuse_vma_close() argument 1219 filemap_write_and_wait(vma->vm_file->f_mapping); in fuse_vma_close() 1237 static int fuse_page_mkwrite(struct vm_area_struct *vma, struct page *page) in fuse_page_mkwrite() argument 1243 struct inode *inode = vma->vm_file->f_mapping->host; in fuse_page_mkwrite() 1255 static int fuse_file_mmap(struct file *file, struct vm_area_struct *vma) in fuse_file_mmap() argument 1257 if ((vma->vm_flags & VM_SHARED) && (vma->vm_flags & VM_MAYWRITE)) { in fuse_file_mmap() 1272 vma->vm_ops = &fuse_file_vm_ops; in fuse_file_mmap()
|
/fs/xfs/ |
D | xfs_mount.h | 96 #define XFS_SEND_MMAP(mp, vma,fl) \ argument 97 (*(mp)->m_dm_ops->xfs_send_mmap)(vma,fl)
|
/fs/btrfs/ |
D | file.c | 1276 static int btrfs_file_mmap(struct file *filp, struct vm_area_struct *vma) in btrfs_file_mmap() argument 1278 vma->vm_ops = &btrfs_file_vm_ops; in btrfs_file_mmap()
|