Searched refs:vm_page_prot (Results 1 – 13 of 13) sorted by relevance
/mm/ |
D | mmap.c | 129 pgprot_t vm_page_prot; in vma_set_page_prot() local 131 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot() 132 if (vma_wants_writenotify(vma, vm_page_prot)) { in vma_set_page_prot() 134 vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags); in vma_set_page_prot() 137 WRITE_ONCE(vma->vm_page_prot, vm_page_prot); in vma_set_page_prot() 1744 int vma_wants_writenotify(struct vm_area_struct *vma, pgprot_t vm_page_prot) in vma_wants_writenotify() argument 1759 if (pgprot_val(vm_page_prot) != in vma_wants_writenotify() 1760 pgprot_val(vm_pgprot_modify(vm_page_prot, vm_flags))) in vma_wants_writenotify() 1856 vma->vm_page_prot = vm_get_page_prot(vm_flags); in mmap_region() 3254 vma->vm_page_prot = vm_get_page_prot(flags); in do_brk_flags() [all …]
|
D | userfaultfd.c | 71 _dst_pte = mk_pte(page, dst_vma->vm_page_prot); in mfill_atomic_install_pte() 195 dst_vma->vm_page_prot)); in mfill_zeropage_pte()
|
D | mprotect.c | 485 dirty_accountable = vma_wants_writenotify(vma, vma->vm_page_prot); in mprotect_fixup() 488 change_protection(vma, start, end, vma->vm_page_prot, in mprotect_fixup()
|
D | debug.c | 210 (unsigned long)pgprot_val(vma->vm_page_prot), in dump_vma()
|
D | huge_memory.c | 635 entry = mk_huge_pmd(page, vma->vm_page_prot); in __do_huge_pmd_anonymous_page() 702 entry = mk_pmd(zero_page, vma->vm_page_prot); in set_huge_zero_page() 1546 pmd = pmd_modify(pmd, vma->vm_page_prot); in do_huge_pmd_numa_page() 2015 entry = pfn_pte(my_zero_pfn(haddr), vma->vm_page_prot); in __split_huge_zero_page_pmd() 2156 entry = mk_pte(page + i, READ_ONCE(vma->vm_page_prot)); in __split_huge_pmd_locked() 3010 pmde = pmd_mkold(mk_huge_pmd(new, vma->vm_page_prot)); in remove_migration_pmd()
|
D | memory.c | 902 pte = mk_pte(new_page, dst_vma->vm_page_prot); in copy_present_page() 1862 return insert_pages(vma, addr, pages, num, vma->vm_page_prot); in vm_insert_pages() 1919 return insert_page(vma, addr, page, vma->vm_page_prot); in vm_insert_page() 2130 return vmf_insert_pfn_prot(vma, addr, pfn, vma->vm_page_prot); in vmf_insert_pfn() 2230 return __vm_insert_mixed(vma, addr, pfn, vma->vm_page_prot, false); in vmf_insert_mixed() 2242 return __vm_insert_mixed(vma, addr, pfn, vma->vm_page_prot, true); in vmf_insert_mixed_mkwrite() 2463 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); in vm_iomap_memory() 4860 .vma_page_prot = vma->vm_page_prot, in __handle_mm_fault() 5057 vmf.vma_page_prot = READ_ONCE(vmf.vma->vm_page_prot); in ___handle_speculative_fault()
|
D | migrate.c | 214 pte = pte_mkold(mk_pte(new, READ_ONCE(vma->vm_page_prot))); in remove_migration_pte() 2180 entry = mk_huge_pmd(new_page, vma->vm_page_prot); in migrate_misplaced_transhuge_page() 2234 entry = pmd_modify(entry, vma->vm_page_prot); in migrate_misplaced_transhuge_page() 2918 entry = mk_pte(page, vma->vm_page_prot); in migrate_vma_insert_page()
|
D | khugepaged.c | 1023 .vma_page_prot = vma->vm_page_prot, in __collapse_huge_page_swapin() 1200 _pmd = mk_huge_pmd(new_page, vma->vm_page_prot); in collapse_huge_page()
|
D | mremap.c | 197 pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr); in move_ptes()
|
D | ksm.c | 1158 newpte = mk_pte(kpage, vma->vm_page_prot); in replace_page() 1161 vma->vm_page_prot)); in replace_page()
|
D | nommu.c | 1639 return io_remap_pfn_range(vma, vma->vm_start, pfn, vm_len, vma->vm_page_prot); in vm_iomap_memory()
|
D | hugetlb.c | 3767 vma->vm_page_prot))); in make_huge_pte() 3770 vma->vm_page_prot)); in make_huge_pte()
|
D | swapfile.c | 1961 pte_mkold(mk_pte(page, vma->vm_page_prot))); in unuse_pte()
|