Lines Matching refs:vm_flags
110 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument
112 pgprot_t ret = __pgprot(pgprot_val(protection_map[vm_flags & in vm_get_page_prot()
114 pgprot_val(arch_vm_get_page_prot(vm_flags))); in vm_get_page_prot()
120 static pgprot_t vm_pgprot_modify(pgprot_t oldprot, unsigned long vm_flags) in vm_pgprot_modify() argument
122 return pgprot_modify(oldprot, vm_get_page_prot(vm_flags)); in vm_pgprot_modify()
128 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local
131 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot()
133 vm_flags &= ~VM_SHARED; in vma_set_page_prot()
134 vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags); in vma_set_page_prot()
146 if (vma->vm_flags & VM_DENYWRITE) in __remove_shared_vm_struct()
148 if (vma->vm_flags & VM_SHARED) in __remove_shared_vm_struct()
704 if (vma->vm_flags & VM_DENYWRITE) in __vma_link_file()
706 if (vma->vm_flags & VM_SHARED) in __vma_link_file()
1084 struct file *file, unsigned long vm_flags, in is_mergeable_vma() argument
1096 if ((vma->vm_flags ^ vm_flags) & ~VM_SOFTDIRTY) in is_mergeable_vma()
1135 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, in can_vma_merge_before() argument
1141 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx, anon_name) && in can_vma_merge_before()
1157 can_vma_merge_after(struct vm_area_struct *vma, unsigned long vm_flags, in can_vma_merge_after() argument
1163 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx, anon_name) && in can_vma_merge_after()
1218 unsigned long end, unsigned long vm_flags, in __vma_merge() argument
1232 if (vm_flags & VM_SPECIAL) in __vma_merge()
1250 can_vma_merge_after(prev, vm_flags, in __vma_merge()
1259 can_vma_merge_before(next, vm_flags, in __vma_merge()
1276 khugepaged_enter_vma_merge(prev, vm_flags); in __vma_merge()
1285 can_vma_merge_before(next, vm_flags, in __vma_merge()
1306 khugepaged_enter_vma_merge(area, vm_flags); in __vma_merge()
1331 !((a->vm_flags ^ b->vm_flags) & ~(VM_ACCESS_FLAGS | VM_SOFTDIRTY)) && in anon_vma_compatible()
1476 vm_flags_t vm_flags; in do_mmap() local
1538 vm_flags = calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(flags) | in do_mmap()
1545 if (mlock_future_check(mm, vm_flags, len)) in do_mmap()
1591 vm_flags |= VM_SHARED | VM_MAYSHARE; in do_mmap()
1593 vm_flags &= ~(VM_MAYWRITE | VM_SHARED); in do_mmap()
1599 if (vm_flags & VM_EXEC) in do_mmap()
1601 vm_flags &= ~VM_MAYEXEC; in do_mmap()
1606 if (vm_flags & (VM_GROWSDOWN|VM_GROWSUP)) in do_mmap()
1616 if (vm_flags & (VM_GROWSDOWN|VM_GROWSUP)) in do_mmap()
1622 vm_flags |= VM_SHARED | VM_MAYSHARE; in do_mmap()
1642 vm_flags |= VM_NORESERVE; in do_mmap()
1646 vm_flags |= VM_NORESERVE; in do_mmap()
1649 addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); in do_mmap()
1651 ((vm_flags & VM_LOCKED) || in do_mmap()
1746 vm_flags_t vm_flags = vma->vm_flags; in vma_wants_writenotify() local
1750 if ((vm_flags & (VM_WRITE|VM_SHARED)) != ((VM_WRITE|VM_SHARED))) in vma_wants_writenotify()
1760 pgprot_val(vm_pgprot_modify(vm_page_prot, vm_flags))) in vma_wants_writenotify()
1767 if (IS_ENABLED(CONFIG_MEM_SOFT_DIRTY) && !(vm_flags & VM_SOFTDIRTY) && in vma_wants_writenotify()
1772 if (vm_flags & VM_PFNMAP) in vma_wants_writenotify()
1784 static inline int accountable_mapping(struct file *file, vm_flags_t vm_flags) in accountable_mapping() argument
1793 return (vm_flags & (VM_NORESERVE | VM_SHARED | VM_WRITE)) == VM_WRITE; in accountable_mapping()
1797 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, in mmap_region() argument
1807 if (!may_expand_vm(mm, vm_flags, len >> PAGE_SHIFT)) { in mmap_region()
1816 if (!may_expand_vm(mm, vm_flags, in mmap_region()
1827 if (accountable_mapping(file, vm_flags)) { in mmap_region()
1831 vm_flags |= VM_ACCOUNT; in mmap_region()
1837 vma = vma_merge(mm, prev, addr, addr + len, vm_flags, in mmap_region()
1855 vma->vm_flags = vm_flags; in mmap_region()
1856 vma->vm_page_prot = vm_get_page_prot(vm_flags); in mmap_region()
1860 if (vm_flags & VM_DENYWRITE) { in mmap_region()
1865 if (vm_flags & VM_SHARED) { in mmap_region()
1895 if (unlikely(vm_flags != vma->vm_flags && prev)) { in mmap_region()
1896 merge = vma_merge(mm, prev, vma->vm_start, vma->vm_end, vma->vm_flags, in mmap_region()
1908 vm_flags = vma->vm_flags; in mmap_region()
1913 vm_flags = vma->vm_flags; in mmap_region()
1914 } else if (vm_flags & VM_SHARED) { in mmap_region()
1923 if (!arch_validate_flags(vma->vm_flags)) { in mmap_region()
1935 if (vm_flags & VM_SHARED) in mmap_region()
1937 if (vm_flags & VM_DENYWRITE) in mmap_region()
1945 vm_stat_account(mm, vm_flags, len >> PAGE_SHIFT); in mmap_region()
1946 if (vm_flags & VM_LOCKED) { in mmap_region()
1947 if ((vm_flags & VM_SPECIAL) || vma_is_dax(vma) || in mmap_region()
1950 WRITE_ONCE(vma->vm_flags, in mmap_region()
1951 vma->vm_flags & VM_LOCKED_CLEAR_MASK); in mmap_region()
1966 WRITE_ONCE(vma->vm_flags, vma->vm_flags | VM_SOFTDIRTY); in mmap_region()
1984 if (vm_flags & VM_SHARED) in mmap_region()
1987 if (vm_flags & VM_DENYWRITE) in mmap_region()
2481 if (!may_expand_vm(mm, vma->vm_flags, grow)) in acct_stack_growth()
2489 if (vma->vm_flags & VM_LOCKED) { in acct_stack_growth()
2500 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : in acct_stack_growth()
2527 if (!(vma->vm_flags & VM_GROWSUP)) in expand_upwards()
2545 if (!(next->vm_flags & VM_GROWSUP)) in expand_upwards()
2584 if (vma->vm_flags & VM_LOCKED) in expand_upwards()
2586 vm_stat_account(mm, vma->vm_flags, grow); in expand_upwards()
2601 khugepaged_enter_vma_merge(vma, vma->vm_flags); in expand_upwards()
2624 if (prev && !(prev->vm_flags & VM_GROWSDOWN) && in expand_downwards()
2664 if (vma->vm_flags & VM_LOCKED) in expand_downwards()
2666 vm_stat_account(mm, vma->vm_flags, grow); in expand_downwards()
2679 khugepaged_enter_vma_merge(vma, vma->vm_flags); in expand_downwards()
2718 if (prev->vm_flags & VM_LOCKED) in find_extend_vma()
2740 if (!(vma->vm_flags & VM_GROWSDOWN)) in find_extend_vma()
2745 if (vma->vm_flags & VM_LOCKED) in find_extend_vma()
2768 if (vma->vm_flags & VM_ACCOUNT) in remove_vma_list()
2770 vm_stat_account(mm, vma->vm_flags, -nrpages); in remove_vma_list()
2804 if ((cur_vma->vm_flags & (VM_PFNMAP|VM_MIXEDMAP)) != 0) { in unmap_region()
2850 if (vma && (vma->vm_flags & VM_GROWSDOWN)) in detach_vmas_to_be_unmapped()
2852 if (prev && (prev->vm_flags & VM_GROWSUP)) in detach_vmas_to_be_unmapped()
3025 if (tmp->vm_flags & VM_LOCKED) { in __do_munmap()
3127 if (!vma || !(vma->vm_flags & VM_SHARED)) in SYSCALL_DEFINE5()
3144 if (next->vm_flags != vma->vm_flags) in SYSCALL_DEFINE5()
3155 prot |= vma->vm_flags & VM_READ ? PROT_READ : 0; in SYSCALL_DEFINE5()
3156 prot |= vma->vm_flags & VM_WRITE ? PROT_WRITE : 0; in SYSCALL_DEFINE5()
3157 prot |= vma->vm_flags & VM_EXEC ? PROT_EXEC : 0; in SYSCALL_DEFINE5()
3161 if (vma->vm_flags & VM_LOCKED) { in SYSCALL_DEFINE5()
3253 vma->vm_flags = flags; in do_brk_flags()
3262 vma->vm_flags |= VM_SOFTDIRTY; in do_brk_flags()
3335 if (vma->vm_flags & VM_LOCKED) in exit_mmap()
3361 if (vma->vm_flags & VM_ACCOUNT) in exit_mmap()
3383 if ((vma->vm_flags & VM_ACCOUNT) && in insert_vm_struct()
3445 new_vma = __vma_merge(mm, prev, addr, addr + len, vma->vm_flags, in copy_vma()
3623 unsigned long vm_flags, void *priv, in __install_special_mapping() argument
3636 vma->vm_flags = vm_flags | mm->def_flags | VM_DONTEXPAND | VM_SOFTDIRTY; in __install_special_mapping()
3637 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in __install_special_mapping()
3646 vm_stat_account(mm, vma->vm_flags, len >> PAGE_SHIFT); in __install_special_mapping()
3677 unsigned long vm_flags, const struct vm_special_mapping *spec) in _install_special_mapping() argument
3679 return __install_special_mapping(mm, addr, len, vm_flags, (void *)spec, in _install_special_mapping()
3685 unsigned long vm_flags, struct page **pages) in install_special_mapping() argument
3688 mm, addr, len, vm_flags, (void *)pages, in install_special_mapping()