Home
last modified time | relevance | path

Searched refs:fault (Results 1 – 25 of 147) sorted by relevance

123456

/arch/nios2/kernel/
Dmisaligned.c72 unsigned int fault; in handle_unaligned_c() local
85 fault = 0; in handle_unaligned_c()
98 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c()
99 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c()
111 fault |= __put_user(d0, (u8 *)(addr+0)); in handle_unaligned_c()
112 fault |= __put_user(d1, (u8 *)(addr+1)); in handle_unaligned_c()
116 fault |= __get_user(d0, (u8 *)(addr+0)); in handle_unaligned_c()
117 fault |= __get_user(d1, (u8 *)(addr+1)); in handle_unaligned_c()
133 fault |= __put_user(d0, (u8 *)(addr+0)); in handle_unaligned_c()
134 fault |= __put_user(d1, (u8 *)(addr+1)); in handle_unaligned_c()
[all …]
/arch/arc/kernel/
Dunaligned.c54 goto fault; \
69 goto fault; \
96 goto fault; \
129 goto fault; \
163 fault: state->fault = 1; in fixup_load()
183 goto fault; in fixup_store()
195 fault: state->fault = 1; in fixup_store()
229 if (state.fault) in misaligned_fixup()
230 goto fault; in misaligned_fixup()
234 goto fault; in misaligned_fixup()
[all …]
/arch/s390/mm/
Dfault.c336 static noinline void do_fault_error(struct pt_regs *regs, int access, int fault) in do_fault_error() argument
340 switch (fault) { in do_fault_error()
348 si_code = (fault == VM_FAULT_BADMAP) ? in do_fault_error()
362 if (fault & VM_FAULT_OOM) { in do_fault_error()
367 } else if (fault & VM_FAULT_SIGSEGV) { in do_fault_error()
373 } else if (fault & VM_FAULT_SIGBUS) { in do_fault_error()
407 int fault; in do_exception() local
427 fault = VM_FAULT_BADCONTEXT; in do_exception()
449 fault = VM_FAULT_BADMAP; in do_exception()
458 fault = VM_FAULT_BADMAP; in do_exception()
[all …]
/arch/m68k/mm/
Dfault.c75 int fault; in do_page_fault() local
140 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
141 pr_debug("handle_mm_fault returns %d\n", fault); in do_page_fault()
143 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
146 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
147 if (fault & VM_FAULT_OOM) in do_page_fault()
149 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
151 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
162 if (fault & VM_FAULT_MAJOR) in do_page_fault()
166 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/parisc/mm/
Dfault.c265 int fault = 0; in do_page_fault() local
304 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
306 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
309 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
315 if (fault & VM_FAULT_OOM) in do_page_fault()
317 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
319 else if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| in do_page_fault()
325 if (fault & VM_FAULT_MAJOR) in do_page_fault()
329 if (fault & VM_FAULT_RETRY) { in do_page_fault()
392 if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { in do_page_fault()
[all …]
/arch/arc/mm/
Dfault.c69 int fault, ret; in do_page_fault() local
140 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
144 if ((fault & VM_FAULT_ERROR) && !(fault & VM_FAULT_RETRY)) in do_page_fault()
152 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault()
155 if (fault & VM_FAULT_MAJOR) { in do_page_fault()
165 if (fault & VM_FAULT_RETRY) { in do_page_fault()
177 if (fault & VM_FAULT_OOM) in do_page_fault()
179 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
181 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
/arch/hexagon/mm/
Dvm_fault.c55 int fault; in do_page_fault() local
105 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
107 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
111 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault()
113 if (fault & VM_FAULT_MAJOR) in do_page_fault()
117 if (fault & VM_FAULT_RETRY) { in do_page_fault()
134 if (fault & VM_FAULT_OOM) { in do_page_fault()
142 if (fault & VM_FAULT_SIGBUS) { in do_page_fault()
/arch/powerpc/mm/
Dfault.c154 unsigned int fault) in do_sigbus() argument
168 if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { in do_sigbus()
174 if (fault & VM_FAULT_HWPOISON_LARGE) in do_sigbus()
175 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in do_sigbus()
176 if (fault & VM_FAULT_HWPOISON) in do_sigbus()
184 static int mm_fault_error(struct pt_regs *regs, unsigned long addr, int fault) in mm_fault_error() argument
194 if (fault & VM_FAULT_OOM) { in mm_fault_error()
203 if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON| in mm_fault_error()
205 return do_sigbus(regs, addr, fault); in mm_fault_error()
206 else if (fault & VM_FAULT_SIGSEGV) in mm_fault_error()
[all …]
/arch/nios2/mm/
Dfault.c50 int fault; in do_page_fault() local
135 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
137 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
140 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
141 if (fault & VM_FAULT_OOM) in do_page_fault()
143 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
145 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
156 if (fault & VM_FAULT_MAJOR) in do_page_fault()
160 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/alpha/mm/
Dfault.c90 int fault, si_code = SEGV_MAPERR; in do_page_fault() local
151 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
153 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
156 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
157 if (fault & VM_FAULT_OOM) in do_page_fault()
159 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
161 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
167 if (fault & VM_FAULT_MAJOR) in do_page_fault()
171 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/microblaze/mm/
Dfault.c94 int fault; in do_page_fault() local
219 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
221 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
224 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
225 if (fault & VM_FAULT_OOM) in do_page_fault()
227 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
229 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
235 if (unlikely(fault & VM_FAULT_MAJOR)) in do_page_fault()
239 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/mips/kernel/
Dunaligned.c950 goto fault; in emulate_load_store_insn()
959 goto fault; in emulate_load_store_insn()
985 goto fault; in emulate_load_store_insn()
998 goto fault; in emulate_load_store_insn()
1011 goto fault; in emulate_load_store_insn()
1026 goto fault; in emulate_load_store_insn()
1039 goto fault; in emulate_load_store_insn()
1064 goto fault; in emulate_load_store_insn()
1083 goto fault; in emulate_load_store_insn()
1102 goto fault; in emulate_load_store_insn()
[all …]
/arch/openrisc/mm/
Dfault.c56 int fault; in do_page_fault() local
166 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
168 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
171 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
172 if (fault & VM_FAULT_OOM) in do_page_fault()
174 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
176 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
183 if (fault & VM_FAULT_MAJOR) in do_page_fault()
187 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/unicore32/mm/
Dfault.c174 int fault; in __do_pf() local
177 fault = VM_FAULT_BADMAP; in __do_pf()
189 fault = VM_FAULT_BADACCESS; in __do_pf()
197 fault = handle_mm_fault(vma, addr & PAGE_MASK, flags); in __do_pf()
198 return fault; in __do_pf()
204 return fault; in __do_pf()
211 int fault, sig, code; in do_pf() local
254 fault = __do_pf(mm, addr, fsr, flags, tsk); in do_pf()
260 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_pf()
263 if (!(fault & VM_FAULT_ERROR) && (flags & FAULT_FLAG_ALLOW_RETRY)) { in do_pf()
[all …]
/arch/metag/mm/
Dfault.c57 int fault; in do_page_fault() local
138 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
140 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
143 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
144 if (fault & VM_FAULT_OOM) in do_page_fault()
146 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
148 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
153 if (fault & VM_FAULT_MAJOR) in do_page_fault()
157 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/cris/mm/
Dfault.c62 int fault; in do_page_fault() local
173 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
175 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
178 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
179 if (fault & VM_FAULT_OOM) in do_page_fault()
181 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
183 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
189 if (fault & VM_FAULT_MAJOR) in do_page_fault()
193 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/xtensa/mm/
Dfault.c47 int fault; in do_page_fault() local
113 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
115 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
118 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
119 if (fault & VM_FAULT_OOM) in do_page_fault()
121 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
123 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
128 if (fault & VM_FAULT_MAJOR) in do_page_fault()
132 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/mn10300/kernel/
Dgdb-stub.c518 goto fault; in gdbstub_single_step()
524 goto fault; in gdbstub_single_step()
526 goto fault; in gdbstub_single_step()
529 goto fault; in gdbstub_single_step()
535 goto fault; in gdbstub_single_step()
538 goto fault; in gdbstub_single_step()
545 goto fault; in gdbstub_single_step()
553 goto fault; in gdbstub_single_step()
555 goto fault; in gdbstub_single_step()
565 goto fault; in gdbstub_single_step()
[all …]
/arch/ia64/mm/
Dfault.c90 int fault; in ia64_do_page_fault() local
165 fault = handle_mm_fault(vma, address, flags); in ia64_do_page_fault()
167 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in ia64_do_page_fault()
170 if (unlikely(fault & VM_FAULT_ERROR)) { in ia64_do_page_fault()
176 if (fault & VM_FAULT_OOM) { in ia64_do_page_fault()
178 } else if (fault & VM_FAULT_SIGSEGV) { in ia64_do_page_fault()
180 } else if (fault & VM_FAULT_SIGBUS) { in ia64_do_page_fault()
188 if (fault & VM_FAULT_MAJOR) in ia64_do_page_fault()
192 if (fault & VM_FAULT_RETRY) { in ia64_do_page_fault()
/arch/score/mm/
Dfault.c53 int fault; in do_page_fault() local
114 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
115 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
116 if (fault & VM_FAULT_OOM) in do_page_fault()
118 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
120 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
124 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/sh/mm/
Dfault.c323 unsigned long address, unsigned int fault) in mm_fault_error() argument
330 if (!(fault & VM_FAULT_RETRY)) in mm_fault_error()
337 if (!(fault & VM_FAULT_ERROR)) in mm_fault_error()
340 if (fault & VM_FAULT_OOM) { in mm_fault_error()
356 if (fault & VM_FAULT_SIGBUS) in mm_fault_error()
358 else if (fault & VM_FAULT_SIGSEGV) in mm_fault_error()
406 int fault; in do_page_fault() local
491 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
493 if (unlikely(fault & (VM_FAULT_RETRY | VM_FAULT_ERROR))) in do_page_fault()
494 if (mm_fault_error(regs, error_code, address, fault)) in do_page_fault()
[all …]
/arch/mips/mm/
Dfault.c46 int fault; in __do_page_fault() local
155 fault = handle_mm_fault(vma, address, flags); in __do_page_fault()
157 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in __do_page_fault()
161 if (unlikely(fault & VM_FAULT_ERROR)) { in __do_page_fault()
162 if (fault & VM_FAULT_OOM) in __do_page_fault()
164 else if (fault & VM_FAULT_SIGSEGV) in __do_page_fault()
166 else if (fault & VM_FAULT_SIGBUS) in __do_page_fault()
171 if (fault & VM_FAULT_MAJOR) { in __do_page_fault()
180 if (fault & VM_FAULT_RETRY) { in __do_page_fault()
/arch/mn10300/mm/
Dfault.c126 int fault; in do_page_fault() local
257 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
259 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
262 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
263 if (fault & VM_FAULT_OOM) in do_page_fault()
265 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
267 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
272 if (fault & VM_FAULT_MAJOR) in do_page_fault()
276 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/sparc/kernel/
Dwuf.S258 LEON_PI( lda [%l4] ASI_LEON_MMUREGS, %g0) ! clear fault status
259 SUN_PI_( lda [%l4] ASI_M_MMUREGS, %g0) ! clear fault status
268 or %l5, 0x2, %l5 ! turn on no-fault bit
284 andn %twin_tmp1, 0x2, %twin_tmp1 ! clear no-fault bit
289 LEON_PI(lda [%twin_tmp2] ASI_LEON_MMUREGS, %g0) ! read fault address
290 SUN_PI_(lda [%twin_tmp2] ASI_M_MMUREGS, %g0) ! read fault address
293 LEON_PI(lda [%twin_tmp2] ASI_LEON_MMUREGS, %twin_tmp2) ! read fault status
294 SUN_PI_(lda [%twin_tmp2] ASI_M_MMUREGS, %twin_tmp2) ! read fault status
295 andcc %twin_tmp2, 0x2, %g0 ! did fault occur?
/arch/arm/mm/
Dalignment.c225 goto fault; \
246 goto fault; \
278 goto fault; \
320 goto fault; \
387 fault: in do_alignment_ldrhstrh()
450 fault: in do_alignment_ldrdstrd()
486 fault: in do_alignment_ldrstr()
586 fault: in do_alignment_ldmstm()
774 int fault; in alignment_get_arm() local
777 fault = get_user(instr, ip); in alignment_get_arm()
[all …]

123456