/arch/nios2/kernel/ |
D | misaligned.c | 72 unsigned int fault; in handle_unaligned_c() local 85 fault = 0; in handle_unaligned_c() 98 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 99 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() 111 fault |= __put_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 112 fault |= __put_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() 116 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 117 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() 133 fault |= __put_user(d0, (u8 *)(addr+0)); in handle_unaligned_c() 134 fault |= __put_user(d1, (u8 *)(addr+1)); in handle_unaligned_c() [all …]
|
/arch/arc/kernel/ |
D | unaligned.c | 54 goto fault; \ 69 goto fault; \ 96 goto fault; \ 129 goto fault; \ 163 fault: state->fault = 1; in fixup_load() 183 goto fault; in fixup_store() 195 fault: state->fault = 1; in fixup_store() 229 if (state.fault) in misaligned_fixup() 230 goto fault; in misaligned_fixup() 234 goto fault; in misaligned_fixup() [all …]
|
/arch/s390/mm/ |
D | fault.c | 336 static noinline void do_fault_error(struct pt_regs *regs, int access, int fault) in do_fault_error() argument 340 switch (fault) { in do_fault_error() 348 si_code = (fault == VM_FAULT_BADMAP) ? in do_fault_error() 362 if (fault & VM_FAULT_OOM) { in do_fault_error() 367 } else if (fault & VM_FAULT_SIGSEGV) { in do_fault_error() 373 } else if (fault & VM_FAULT_SIGBUS) { in do_fault_error() 407 int fault; in do_exception() local 427 fault = VM_FAULT_BADCONTEXT; in do_exception() 449 fault = VM_FAULT_BADMAP; in do_exception() 458 fault = VM_FAULT_BADMAP; in do_exception() [all …]
|
/arch/m68k/mm/ |
D | fault.c | 75 int fault; in do_page_fault() local 140 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 141 pr_debug("handle_mm_fault returns %d\n", fault); in do_page_fault() 143 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 146 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 147 if (fault & VM_FAULT_OOM) in do_page_fault() 149 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 151 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 162 if (fault & VM_FAULT_MAJOR) in do_page_fault() 166 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/parisc/mm/ |
D | fault.c | 265 int fault = 0; in do_page_fault() local 304 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 306 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 309 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 315 if (fault & VM_FAULT_OOM) in do_page_fault() 317 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 319 else if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| in do_page_fault() 325 if (fault & VM_FAULT_MAJOR) in do_page_fault() 329 if (fault & VM_FAULT_RETRY) { in do_page_fault() 392 if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { in do_page_fault() [all …]
|
/arch/arc/mm/ |
D | fault.c | 69 int fault, ret; in do_page_fault() local 140 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 144 if ((fault & VM_FAULT_ERROR) && !(fault & VM_FAULT_RETRY)) in do_page_fault() 152 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault() 155 if (fault & VM_FAULT_MAJOR) { in do_page_fault() 165 if (fault & VM_FAULT_RETRY) { in do_page_fault() 177 if (fault & VM_FAULT_OOM) in do_page_fault() 179 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 181 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
|
/arch/hexagon/mm/ |
D | vm_fault.c | 55 int fault; in do_page_fault() local 105 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 107 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 111 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault() 113 if (fault & VM_FAULT_MAJOR) in do_page_fault() 117 if (fault & VM_FAULT_RETRY) { in do_page_fault() 134 if (fault & VM_FAULT_OOM) { in do_page_fault() 142 if (fault & VM_FAULT_SIGBUS) { in do_page_fault()
|
/arch/powerpc/mm/ |
D | fault.c | 154 unsigned int fault) in do_sigbus() argument 168 if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { in do_sigbus() 174 if (fault & VM_FAULT_HWPOISON_LARGE) in do_sigbus() 175 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in do_sigbus() 176 if (fault & VM_FAULT_HWPOISON) in do_sigbus() 184 static int mm_fault_error(struct pt_regs *regs, unsigned long addr, int fault) in mm_fault_error() argument 194 if (fault & VM_FAULT_OOM) { in mm_fault_error() 203 if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| in mm_fault_error() 205 return do_sigbus(regs, addr, fault); in mm_fault_error() 206 else if (fault & VM_FAULT_SIGSEGV) in mm_fault_error() [all …]
|
/arch/nios2/mm/ |
D | fault.c | 50 int fault; in do_page_fault() local 135 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 137 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 140 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 141 if (fault & VM_FAULT_OOM) in do_page_fault() 143 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 145 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 156 if (fault & VM_FAULT_MAJOR) in do_page_fault() 160 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/alpha/mm/ |
D | fault.c | 90 int fault, si_code = SEGV_MAPERR; in do_page_fault() local 151 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 153 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 156 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 157 if (fault & VM_FAULT_OOM) in do_page_fault() 159 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 161 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 167 if (fault & VM_FAULT_MAJOR) in do_page_fault() 171 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/microblaze/mm/ |
D | fault.c | 94 int fault; in do_page_fault() local 219 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 221 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 224 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 225 if (fault & VM_FAULT_OOM) in do_page_fault() 227 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 229 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 235 if (unlikely(fault & VM_FAULT_MAJOR)) in do_page_fault() 239 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/mips/kernel/ |
D | unaligned.c | 950 goto fault; in emulate_load_store_insn() 959 goto fault; in emulate_load_store_insn() 985 goto fault; in emulate_load_store_insn() 998 goto fault; in emulate_load_store_insn() 1011 goto fault; in emulate_load_store_insn() 1026 goto fault; in emulate_load_store_insn() 1039 goto fault; in emulate_load_store_insn() 1064 goto fault; in emulate_load_store_insn() 1083 goto fault; in emulate_load_store_insn() 1102 goto fault; in emulate_load_store_insn() [all …]
|
/arch/openrisc/mm/ |
D | fault.c | 56 int fault; in do_page_fault() local 166 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 168 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 171 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 172 if (fault & VM_FAULT_OOM) in do_page_fault() 174 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 176 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 183 if (fault & VM_FAULT_MAJOR) in do_page_fault() 187 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/unicore32/mm/ |
D | fault.c | 174 int fault; in __do_pf() local 177 fault = VM_FAULT_BADMAP; in __do_pf() 189 fault = VM_FAULT_BADACCESS; in __do_pf() 197 fault = handle_mm_fault(vma, addr & PAGE_MASK, flags); in __do_pf() 198 return fault; in __do_pf() 204 return fault; in __do_pf() 211 int fault, sig, code; in do_pf() local 254 fault = __do_pf(mm, addr, fsr, flags, tsk); in do_pf() 260 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_pf() 263 if (!(fault & VM_FAULT_ERROR) && (flags & FAULT_FLAG_ALLOW_RETRY)) { in do_pf() [all …]
|
/arch/metag/mm/ |
D | fault.c | 57 int fault; in do_page_fault() local 138 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 140 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 143 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 144 if (fault & VM_FAULT_OOM) in do_page_fault() 146 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 148 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 153 if (fault & VM_FAULT_MAJOR) in do_page_fault() 157 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/cris/mm/ |
D | fault.c | 62 int fault; in do_page_fault() local 173 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 175 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 178 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 179 if (fault & VM_FAULT_OOM) in do_page_fault() 181 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 183 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 189 if (fault & VM_FAULT_MAJOR) in do_page_fault() 193 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/xtensa/mm/ |
D | fault.c | 47 int fault; in do_page_fault() local 113 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 115 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 118 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 119 if (fault & VM_FAULT_OOM) in do_page_fault() 121 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 123 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 128 if (fault & VM_FAULT_MAJOR) in do_page_fault() 132 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/mn10300/kernel/ |
D | gdb-stub.c | 518 goto fault; in gdbstub_single_step() 524 goto fault; in gdbstub_single_step() 526 goto fault; in gdbstub_single_step() 529 goto fault; in gdbstub_single_step() 535 goto fault; in gdbstub_single_step() 538 goto fault; in gdbstub_single_step() 545 goto fault; in gdbstub_single_step() 553 goto fault; in gdbstub_single_step() 555 goto fault; in gdbstub_single_step() 565 goto fault; in gdbstub_single_step() [all …]
|
/arch/ia64/mm/ |
D | fault.c | 90 int fault; in ia64_do_page_fault() local 165 fault = handle_mm_fault(vma, address, flags); in ia64_do_page_fault() 167 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in ia64_do_page_fault() 170 if (unlikely(fault & VM_FAULT_ERROR)) { in ia64_do_page_fault() 176 if (fault & VM_FAULT_OOM) { in ia64_do_page_fault() 178 } else if (fault & VM_FAULT_SIGSEGV) { in ia64_do_page_fault() 180 } else if (fault & VM_FAULT_SIGBUS) { in ia64_do_page_fault() 188 if (fault & VM_FAULT_MAJOR) in ia64_do_page_fault() 192 if (fault & VM_FAULT_RETRY) { in ia64_do_page_fault()
|
/arch/score/mm/ |
D | fault.c | 53 int fault; in do_page_fault() local 114 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 115 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 116 if (fault & VM_FAULT_OOM) in do_page_fault() 118 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 120 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 124 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/sh/mm/ |
D | fault.c | 323 unsigned long address, unsigned int fault) in mm_fault_error() argument 330 if (!(fault & VM_FAULT_RETRY)) in mm_fault_error() 337 if (!(fault & VM_FAULT_ERROR)) in mm_fault_error() 340 if (fault & VM_FAULT_OOM) { in mm_fault_error() 356 if (fault & VM_FAULT_SIGBUS) in mm_fault_error() 358 else if (fault & VM_FAULT_SIGSEGV) in mm_fault_error() 406 int fault; in do_page_fault() local 491 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 493 if (unlikely(fault & (VM_FAULT_RETRY | VM_FAULT_ERROR))) in do_page_fault() 494 if (mm_fault_error(regs, error_code, address, fault)) in do_page_fault() [all …]
|
/arch/mips/mm/ |
D | fault.c | 46 int fault; in __do_page_fault() local 155 fault = handle_mm_fault(vma, address, flags); in __do_page_fault() 157 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in __do_page_fault() 161 if (unlikely(fault & VM_FAULT_ERROR)) { in __do_page_fault() 162 if (fault & VM_FAULT_OOM) in __do_page_fault() 164 else if (fault & VM_FAULT_SIGSEGV) in __do_page_fault() 166 else if (fault & VM_FAULT_SIGBUS) in __do_page_fault() 171 if (fault & VM_FAULT_MAJOR) { in __do_page_fault() 180 if (fault & VM_FAULT_RETRY) { in __do_page_fault()
|
/arch/mn10300/mm/ |
D | fault.c | 126 int fault; in do_page_fault() local 257 fault = handle_mm_fault(vma, address, flags); in do_page_fault() 259 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 262 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 263 if (fault & VM_FAULT_OOM) in do_page_fault() 265 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault() 267 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 272 if (fault & VM_FAULT_MAJOR) in do_page_fault() 276 if (fault & VM_FAULT_RETRY) { in do_page_fault()
|
/arch/sparc/kernel/ |
D | wuf.S | 258 LEON_PI( lda [%l4] ASI_LEON_MMUREGS, %g0) ! clear fault status 259 SUN_PI_( lda [%l4] ASI_M_MMUREGS, %g0) ! clear fault status 268 or %l5, 0x2, %l5 ! turn on no-fault bit 284 andn %twin_tmp1, 0x2, %twin_tmp1 ! clear no-fault bit 289 LEON_PI(lda [%twin_tmp2] ASI_LEON_MMUREGS, %g0) ! read fault address 290 SUN_PI_(lda [%twin_tmp2] ASI_M_MMUREGS, %g0) ! read fault address 293 LEON_PI(lda [%twin_tmp2] ASI_LEON_MMUREGS, %twin_tmp2) ! read fault status 294 SUN_PI_(lda [%twin_tmp2] ASI_M_MMUREGS, %twin_tmp2) ! read fault status 295 andcc %twin_tmp2, 0x2, %g0 ! did fault occur?
|
/arch/arm/mm/ |
D | alignment.c | 225 goto fault; \ 246 goto fault; \ 278 goto fault; \ 320 goto fault; \ 387 fault: in do_alignment_ldrhstrh() 450 fault: in do_alignment_ldrdstrd() 486 fault: in do_alignment_ldrstr() 586 fault: in do_alignment_ldmstm() 774 int fault; in alignment_get_arm() local 777 fault = get_user(instr, ip); in alignment_get_arm() [all …]
|