/arch/powerpc/include/asm/ |
D | mman.h | 27 static inline pgprot_t arch_vm_get_page_prot(unsigned long vm_flags) in arch_vm_get_page_prot() argument 29 return (vm_flags & VM_SAO) ? __pgprot(_PAGE_SAO) : __pgprot(0); in arch_vm_get_page_prot() 31 #define arch_vm_get_page_prot(vm_flags) arch_vm_get_page_prot(vm_flags) argument
|
/arch/powerpc/platforms/cell/ |
D | spu_fault.c | 55 if (!(vma->vm_flags & VM_GROWSDOWN)) in spu_handle_mm_fault() 63 if (!(vma->vm_flags & VM_WRITE)) in spu_handle_mm_fault() 68 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in spu_handle_mm_fault()
|
/arch/hexagon/mm/ |
D | vm_fault.c | 77 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 89 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 93 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 97 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
/arch/m68k/mm/ |
D | fault.c | 97 if (vma->vm_flags & VM_IO) in do_page_fault() 101 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 126 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 133 if (!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE))) in do_page_fault()
|
/arch/alpha/mm/ |
D | fault.c | 126 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 136 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 140 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) in do_page_fault() 143 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
/arch/score/include/asm/ |
D | cacheflush.h | 31 if (vma->vm_flags & VM_EXEC) { in flush_icache_page() 44 if ((vma->vm_flags & VM_EXEC)) \
|
/arch/arm64/mm/ |
D | flush.c | 33 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 41 if (vma->vm_flags & VM_EXEC) { in flush_ptrace_access()
|
D | fault.c | 156 unsigned int mm_flags, unsigned long vm_flags, in __do_page_fault() argument 178 if (!(vma->vm_flags & vm_flags)) { in __do_page_fault() 186 if (vma->vm_flags & VM_GROWSDOWN && !expand_stack(vma, addr)) in __do_page_fault() 198 unsigned long vm_flags = VM_READ | VM_WRITE; in do_page_fault() local 202 vm_flags = VM_EXEC; in do_page_fault() 204 vm_flags = VM_WRITE; in do_page_fault() 244 fault = __do_page_fault(mm, addr, mm_flags, vm_flags, tsk); in do_page_fault()
|
/arch/unicore32/mm/ |
D | flush.c | 26 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 39 if (vma->vm_flags & VM_EXEC) { in flush_ptrace_access()
|
/arch/avr32/mm/ |
D | fault.c | 97 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 112 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 117 if (!(vma->vm_flags & (VM_READ | VM_WRITE | VM_EXEC))) in do_page_fault() 122 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
/arch/mips/mm/ |
D | fault.c | 104 if (!(vma->vm_flags & VM_GROWSDOWN)) in __do_page_fault() 116 if (!(vma->vm_flags & VM_WRITE)) in __do_page_fault() 120 if (address == regs->cp0_epc && !(vma->vm_flags & VM_EXEC)) { in __do_page_fault() 130 if ((!(vma->vm_flags & VM_READ)) && (regs->cp0_epc != address)) { in __do_page_fault() 144 if (!(vma->vm_flags & (VM_READ | VM_WRITE | VM_EXEC))) in __do_page_fault()
|
/arch/tile/include/asm/ |
D | tlbflush.h | 53 if (!vma || (vma != FLUSH_NONEXEC && (vma->vm_flags & VM_EXEC))) in local_flush_tlb_page() 67 if (!vma || (vma != FLUSH_NONEXEC && (vma->vm_flags & VM_EXEC))) in local_flush_tlb_pages()
|
/arch/powerpc/mm/ |
D | fault.c | 318 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 403 if (!(vma->vm_flags & VM_EXEC) && in do_page_fault() 405 !(vma->vm_flags & (VM_READ | VM_WRITE)))) in do_page_fault() 409 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 416 if (!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE))) in do_page_fault()
|
/arch/ia64/mm/ |
D | fault.c | 149 if (((isr >> IA64_ISR_R_BIT) & 1UL) && (!(vma->vm_flags & (VM_READ | VM_WRITE)))) in ia64_do_page_fault() 152 if ((vma->vm_flags & mask) != mask) in ia64_do_page_fault() 202 if (!(prev_vma && (prev_vma->vm_flags & VM_GROWSUP) && (address == prev_vma->vm_end))) { in ia64_do_page_fault() 205 if (!(vma->vm_flags & VM_GROWSDOWN)) in ia64_do_page_fault()
|
/arch/cris/mm/ |
D | fault.c | 127 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 153 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 156 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 159 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_page_fault()
|
/arch/xtensa/mm/ |
D | fault.c | 83 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 96 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 100 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 103 if (!(vma->vm_flags & (VM_READ | VM_WRITE))) in do_page_fault()
|
D | cache.c | 182 && (vma->vm_flags & VM_EXEC) != 0) { in update_mmu_cache() 225 if ((vma->vm_flags & VM_EXEC) != 0) in copy_to_user_page() 228 } else if ((vma->vm_flags & VM_EXEC) != 0) { in copy_to_user_page()
|
D | tlb.c | 103 if (vma->vm_flags & VM_EXEC) in flush_tlb_range() 135 if (vma->vm_flags & VM_EXEC) in flush_tlb_page()
|
/arch/sparc/mm/ |
D | fault_32.c | 219 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_sparc_fault() 230 if (!(vma->vm_flags & VM_WRITE)) in do_sparc_fault() 234 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_sparc_fault() 396 if (!(vma->vm_flags & VM_GROWSDOWN)) in force_user_fault() 403 if (!(vma->vm_flags & VM_WRITE)) in force_user_fault() 406 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in force_user_fault()
|
/arch/arm/mm/ |
D | flush.c | 81 if (vma->vm_flags & VM_EXEC) in flush_cache_range() 97 if (vma->vm_flags & VM_EXEC && icache_is_vivt_asid_tagged()) in flush_cache_page() 130 if (vma->vm_flags & VM_EXEC) { in flush_ptrace_access() 221 if (!(mpnt->vm_flags & VM_MAYSHARE)) in __flush_dcache_aliases()
|
/arch/alpha/include/asm/ |
D | tlbflush.h | 43 if (vma->vm_flags & VM_EXEC) { in ev4_flush_tlb_current_page() 55 if (vma->vm_flags & VM_EXEC) in ev5_flush_tlb_current_page()
|
/arch/arc/mm/ |
D | fault.c | 98 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 116 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 119 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in do_page_fault()
|
/arch/score/mm/ |
D | fault.c | 84 if (!(vma->vm_flags & VM_GROWSDOWN)) in do_page_fault() 96 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 99 if (!(vma->vm_flags & (VM_READ | VM_WRITE | VM_EXEC))) in do_page_fault()
|
/arch/microblaze/mm/ |
D | fault.c | 154 if (unlikely(!(vma->vm_flags & VM_GROWSDOWN))) in do_page_fault() 200 if (unlikely(!(vma->vm_flags & VM_WRITE))) in do_page_fault() 207 if (unlikely(!(vma->vm_flags & (VM_READ | VM_EXEC)))) in do_page_fault()
|
/arch/tile/kernel/ |
D | tlb.c | 57 int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0; in flush_tlb_page_mm() 73 int cache = (vma->vm_flags & VM_EXEC) ? HV_FLUSH_EVICT_L1I : 0; in flush_tlb_range()
|