/arch/s390/mm/ |
D | fault.c | 216 static noinline void do_fault_error(struct pt_regs *regs, int fault) in do_fault_error() argument 220 switch (fault) { in do_fault_error() 226 si_code = (fault == VM_FAULT_BADMAP) ? in do_fault_error() 235 if (fault & VM_FAULT_OOM) { in do_fault_error() 240 } else if (fault & VM_FAULT_SIGBUS) { in do_fault_error() 271 int fault; in do_exception() local 285 fault = VM_FAULT_BADCONTEXT; in do_exception() 301 fault = VM_FAULT_BADMAP; in do_exception() 305 fault = VM_FAULT_OOM; in do_exception() 312 fault = VM_FAULT_BADMAP; in do_exception() [all …]
|
/arch/arm/lib/ |
D | uaccess.S | 35 USER( TUSER( strb) r3, [r0], #1) @ May fault 37 USER( TUSER( strgeb) r3, [r0], #1) @ May fault 39 USER( TUSER( strgtb) r3, [r0], #1) @ May fault 62 USER( TUSER( str) r3, [r0], #4) @ May fault 77 stmia r0!, {r3 - r6} @ Shouldnt fault 80 stmia r0!, {r3 - r6} @ Shouldnt fault 85 stmgeia r0!, {r3 - r6} @ Shouldnt fault 88 stmneia r0!, {r3 - r4} @ Shouldnt fault 91 TUSER( strne) r3, [r0], #4 @ Shouldnt fault 98 USER( TUSER( strb) r3, [r0], #1) @ May fault [all …]
|
/arch/unicore32/mm/ |
D | fault.c | 174 int fault; in __do_pf() local 177 fault = VM_FAULT_BADMAP; in __do_pf() 189 fault = VM_FAULT_BADACCESS; in __do_pf() 197 fault = handle_mm_fault(mm, vma, addr & PAGE_MASK, in __do_pf() 199 if (unlikely(fault & VM_FAULT_ERROR)) in __do_pf() 200 return fault; in __do_pf() 201 if (fault & VM_FAULT_MAJOR) in __do_pf() 205 return fault; in __do_pf() 211 return fault; in __do_pf() 218 int fault, sig, code; in do_pf() local [all …]
|
/arch/powerpc/mm/ |
D | fault.c | 146 static int mm_fault_error(struct pt_regs *regs, unsigned long addr, int fault) in mm_fault_error() argument 158 if (!(fault & VM_FAULT_RETRY)) in mm_fault_error() 167 if (!(fault & VM_FAULT_ERROR)) in mm_fault_error() 171 if (fault & VM_FAULT_OOM) in mm_fault_error() 177 if (fault & VM_FAULT_SIGBUS) in mm_fault_error() 208 int fault; in do_page_fault() local 420 fault = handle_mm_fault(mm, vma, address, flags); in do_page_fault() 421 if (unlikely(fault & (VM_FAULT_RETRY|VM_FAULT_ERROR))) { in do_page_fault() 422 int rc = mm_fault_error(regs, address, fault); in do_page_fault() 433 if (fault & VM_FAULT_MAJOR) { in do_page_fault() [all …]
|
/arch/m68k/mm/ |
D | fault.c | 75 int write, fault; in do_page_fault() local 142 fault = handle_mm_fault(mm, vma, address, write ? FAULT_FLAG_WRITE : 0); in do_page_fault() 144 printk("handle_mm_fault returns %d\n",fault); in do_page_fault() 146 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 147 if (fault & VM_FAULT_OOM) in do_page_fault() 149 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 153 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/hexagon/mm/ |
D | vm_fault.c | 54 int fault; in do_page_fault() local 99 fault = handle_mm_fault(mm, vma, address, (cause > 0)); in do_page_fault() 102 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault() 103 if (fault & VM_FAULT_MAJOR) in do_page_fault() 118 if (fault & VM_FAULT_OOM) { in do_page_fault() 126 if (fault & VM_FAULT_SIGBUS) { in do_page_fault()
|
/arch/mn10300/kernel/ |
D | gdb-stub.c | 517 goto fault; in gdbstub_single_step() 523 goto fault; in gdbstub_single_step() 525 goto fault; in gdbstub_single_step() 528 goto fault; in gdbstub_single_step() 534 goto fault; in gdbstub_single_step() 537 goto fault; in gdbstub_single_step() 544 goto fault; in gdbstub_single_step() 552 goto fault; in gdbstub_single_step() 554 goto fault; in gdbstub_single_step() 564 goto fault; in gdbstub_single_step() [all …]
|
/arch/s390/lib/ |
D | uaccess_pt.c | 53 goto fault; in __user_copy_pt() 56 goto fault; in __user_copy_pt() 59 goto fault; in __user_copy_pt() 78 fault: in __user_copy_pt() 99 goto fault; in __dat_user_addr() 102 goto fault; in __dat_user_addr() 107 fault: in __dat_user_addr() 179 goto fault; in strnlen_user_pt() 182 goto fault; in strnlen_user_pt() 195 fault: in strnlen_user_pt() [all …]
|
/arch/arm/mm/ |
D | fault.c | 228 int fault; in __do_page_fault() local 231 fault = VM_FAULT_BADMAP; in __do_page_fault() 243 fault = VM_FAULT_BADACCESS; in __do_page_fault() 255 return fault; in __do_page_fault() 263 int fault, sig, code; in do_page_fault() local 309 fault = __do_page_fault(mm, addr, fsr, flags, tsk); in do_page_fault() 315 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 325 if (!(fault & VM_FAULT_ERROR) && flags & FAULT_FLAG_ALLOW_RETRY) { in do_page_fault() 326 if (fault & VM_FAULT_MAJOR) { in do_page_fault() 335 if (fault & VM_FAULT_RETRY) { in do_page_fault() [all …]
|
D | alignment.c | 219 goto fault; \ 240 goto fault; \ 272 goto fault; \ 314 goto fault; \ 375 fault: in do_alignment_ldrhstrh() 431 fault: in do_alignment_ldrdstrd() 462 fault: in do_alignment_ldrstr() 560 fault: in do_alignment_ldmstm() 753 unsigned int fault; in do_alignment() local 765 fault = probe_kernel_address(ptr, tinstr); in do_alignment() [all …]
|
/arch/mips/kernel/ |
D | unaligned.c | 189 goto fault; in emulate_load_store_insn() 221 goto fault; in emulate_load_store_insn() 260 goto fault; in emulate_load_store_insn() 290 goto fault; in emulate_load_store_insn() 324 goto fault; in emulate_load_store_insn() 364 goto fault; in emulate_load_store_insn() 406 goto fault; in emulate_load_store_insn() 448 goto fault; in emulate_load_store_insn() 479 goto fault; in emulate_load_store_insn() 518 goto fault; in emulate_load_store_insn() [all …]
|
/arch/parisc/mm/ |
D | fault.c | 177 int fault; in do_page_fault() local 204 fault = handle_mm_fault(mm, vma, address, (acc_type & VM_WRITE) ? FAULT_FLAG_WRITE : 0); in do_page_fault() 205 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 211 if (fault & VM_FAULT_OOM) in do_page_fault() 213 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 217 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/microblaze/mm/ |
D | fault.c | 94 int fault; in do_page_fault() local 213 fault = handle_mm_fault(mm, vma, address, is_write ? FAULT_FLAG_WRITE : 0); in do_page_fault() 214 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 215 if (fault & VM_FAULT_OOM) in do_page_fault() 217 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 221 if (unlikely(fault & VM_FAULT_MAJOR)) in do_page_fault()
|
/arch/mips/mm/ |
D | fault.c | 43 int fault; in do_page_fault() local 149 fault = handle_mm_fault(mm, vma, address, flags); in do_page_fault() 151 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 155 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 156 if (fault & VM_FAULT_OOM) in do_page_fault() 158 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 163 if (fault & VM_FAULT_MAJOR) { in do_page_fault() 172 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/alpha/mm/ |
D | fault.c | 90 int fault, si_code = SEGV_MAPERR; in do_page_fault() local 147 fault = handle_mm_fault(mm, vma, address, cause > 0 ? FAULT_FLAG_WRITE : 0); in do_page_fault() 149 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 150 if (fault & VM_FAULT_OOM) in do_page_fault() 152 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 156 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/score/mm/ |
D | fault.c | 51 int fault; in do_page_fault() local 109 fault = handle_mm_fault(mm, vma, address, write); in do_page_fault() 110 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 111 if (fault & VM_FAULT_OOM) in do_page_fault() 113 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 117 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/avr32/mm/ |
D | fault.c | 67 int fault; in do_page_fault() local 135 fault = handle_mm_fault(mm, vma, address, writeaccess ? FAULT_FLAG_WRITE : 0); in do_page_fault() 136 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 137 if (fault & VM_FAULT_OOM) in do_page_fault() 139 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 143 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/openrisc/mm/ |
D | fault.c | 56 int fault; in do_page_fault() local 162 fault = handle_mm_fault(mm, vma, address, write_acc); in do_page_fault() 163 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 164 if (fault & VM_FAULT_OOM) in do_page_fault() 166 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 171 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/cris/mm/ |
D | fault.c | 60 int fault; in do_page_fault() local 166 fault = handle_mm_fault(mm, vma, address, (writeaccess & 1) ? FAULT_FLAG_WRITE : 0); in do_page_fault() 167 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 168 if (fault & VM_FAULT_OOM) in do_page_fault() 170 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 174 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/xtensa/mm/ |
D | fault.c | 46 int fault; in do_page_fault() local 107 fault = handle_mm_fault(mm, vma, address, is_write ? FAULT_FLAG_WRITE : 0); in do_page_fault() 108 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 109 if (fault & VM_FAULT_OOM) in do_page_fault() 111 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 115 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/m32r/mm/ |
D | Makefile | 6 obj-y := init.o fault.o mmu.o extable.o ioremap.o cache.o page.o 8 obj-y := init.o fault-nommu.o mmu.o extable.o ioremap-nommu.o cache.o page.o
|
/arch/ia64/mm/ |
D | fault.c | 83 int fault; in ia64_do_page_fault() local 156 fault = handle_mm_fault(mm, vma, address, (mask & VM_WRITE) ? FAULT_FLAG_WRITE : 0); in ia64_do_page_fault() 157 if (unlikely(fault & VM_FAULT_ERROR)) { in ia64_do_page_fault() 163 if (fault & VM_FAULT_OOM) { in ia64_do_page_fault() 165 } else if (fault & VM_FAULT_SIGBUS) { in ia64_do_page_fault() 171 if (fault & VM_FAULT_MAJOR) in ia64_do_page_fault()
|
/arch/frv/mm/ |
D | fault.c | 42 int fault; in do_page_fault() local 165 fault = handle_mm_fault(mm, vma, ear0, write ? FAULT_FLAG_WRITE : 0); in do_page_fault() 166 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 167 if (fault & VM_FAULT_OOM) in do_page_fault() 169 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 173 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/x86/mm/ |
D | fault.c | 166 struct task_struct *tsk, int fault) in force_sig_info_fault() argument 175 if (fault & VM_FAULT_HWPOISON_LARGE) in force_sig_info_fault() 176 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in force_sig_info_fault() 177 if (fault & VM_FAULT_HWPOISON) in force_sig_info_fault() 825 unsigned int fault) in do_sigbus() argument 848 if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { in do_sigbus() 855 force_sig_info_fault(SIGBUS, code, address, tsk, fault); in do_sigbus() 860 unsigned long address, unsigned int fault) in mm_fault_error() argument 867 if (!(fault & VM_FAULT_RETRY)) in mm_fault_error() 873 if (!(fault & VM_FAULT_ERROR)) in mm_fault_error() [all …]
|
/arch/um/kernel/ |
D | trap.c | 66 int fault; in handle_page_fault() local 68 fault = handle_mm_fault(mm, vma, address, is_write ? FAULT_FLAG_WRITE : 0); in handle_page_fault() 69 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault() 70 if (fault & VM_FAULT_OOM) { in handle_page_fault() 72 } else if (fault & VM_FAULT_SIGBUS) { in handle_page_fault() 78 if (fault & VM_FAULT_MAJOR) in handle_page_fault()
|