Home
last modified time | relevance | path

Searched refs:fault (Results 1 – 25 of 110) sorted by relevance

12345

/arch/s390/mm/
Dfault.c216 static noinline void do_fault_error(struct pt_regs *regs, int fault) in do_fault_error() argument
220 switch (fault) { in do_fault_error()
226 si_code = (fault == VM_FAULT_BADMAP) ? in do_fault_error()
235 if (fault & VM_FAULT_OOM) { in do_fault_error()
240 } else if (fault & VM_FAULT_SIGBUS) { in do_fault_error()
271 int fault; in do_exception() local
285 fault = VM_FAULT_BADCONTEXT; in do_exception()
301 fault = VM_FAULT_BADMAP; in do_exception()
305 fault = VM_FAULT_OOM; in do_exception()
312 fault = VM_FAULT_BADMAP; in do_exception()
[all …]
/arch/arm/lib/
Duaccess.S35 USER( TUSER( strb) r3, [r0], #1) @ May fault
37 USER( TUSER( strgeb) r3, [r0], #1) @ May fault
39 USER( TUSER( strgtb) r3, [r0], #1) @ May fault
62 USER( TUSER( str) r3, [r0], #4) @ May fault
77 stmia r0!, {r3 - r6} @ Shouldnt fault
80 stmia r0!, {r3 - r6} @ Shouldnt fault
85 stmgeia r0!, {r3 - r6} @ Shouldnt fault
88 stmneia r0!, {r3 - r4} @ Shouldnt fault
91 TUSER( strne) r3, [r0], #4 @ Shouldnt fault
98 USER( TUSER( strb) r3, [r0], #1) @ May fault
[all …]
/arch/unicore32/mm/
Dfault.c174 int fault; in __do_pf() local
177 fault = VM_FAULT_BADMAP; in __do_pf()
189 fault = VM_FAULT_BADACCESS; in __do_pf()
197 fault = handle_mm_fault(mm, vma, addr & PAGE_MASK, in __do_pf()
199 if (unlikely(fault & VM_FAULT_ERROR)) in __do_pf()
200 return fault; in __do_pf()
201 if (fault & VM_FAULT_MAJOR) in __do_pf()
205 return fault; in __do_pf()
211 return fault; in __do_pf()
218 int fault, sig, code; in do_pf() local
[all …]
/arch/powerpc/mm/
Dfault.c146 static int mm_fault_error(struct pt_regs *regs, unsigned long addr, int fault) in mm_fault_error() argument
158 if (!(fault & VM_FAULT_RETRY)) in mm_fault_error()
167 if (!(fault & VM_FAULT_ERROR)) in mm_fault_error()
171 if (fault & VM_FAULT_OOM) in mm_fault_error()
177 if (fault & VM_FAULT_SIGBUS) in mm_fault_error()
208 int fault; in do_page_fault() local
420 fault = handle_mm_fault(mm, vma, address, flags); in do_page_fault()
421 if (unlikely(fault & (VM_FAULT_RETRY|VM_FAULT_ERROR))) { in do_page_fault()
422 int rc = mm_fault_error(regs, address, fault); in do_page_fault()
433 if (fault & VM_FAULT_MAJOR) { in do_page_fault()
[all …]
/arch/m68k/mm/
Dfault.c75 int write, fault; in do_page_fault() local
142 fault = handle_mm_fault(mm, vma, address, write ? FAULT_FLAG_WRITE : 0); in do_page_fault()
144 printk("handle_mm_fault returns %d\n",fault); in do_page_fault()
146 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
147 if (fault & VM_FAULT_OOM) in do_page_fault()
149 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
153 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/hexagon/mm/
Dvm_fault.c54 int fault; in do_page_fault() local
99 fault = handle_mm_fault(mm, vma, address, (cause > 0)); in do_page_fault()
102 if (likely(!(fault & VM_FAULT_ERROR))) { in do_page_fault()
103 if (fault & VM_FAULT_MAJOR) in do_page_fault()
118 if (fault & VM_FAULT_OOM) { in do_page_fault()
126 if (fault & VM_FAULT_SIGBUS) { in do_page_fault()
/arch/mn10300/kernel/
Dgdb-stub.c517 goto fault; in gdbstub_single_step()
523 goto fault; in gdbstub_single_step()
525 goto fault; in gdbstub_single_step()
528 goto fault; in gdbstub_single_step()
534 goto fault; in gdbstub_single_step()
537 goto fault; in gdbstub_single_step()
544 goto fault; in gdbstub_single_step()
552 goto fault; in gdbstub_single_step()
554 goto fault; in gdbstub_single_step()
564 goto fault; in gdbstub_single_step()
[all …]
/arch/s390/lib/
Duaccess_pt.c53 goto fault; in __user_copy_pt()
56 goto fault; in __user_copy_pt()
59 goto fault; in __user_copy_pt()
78 fault: in __user_copy_pt()
99 goto fault; in __dat_user_addr()
102 goto fault; in __dat_user_addr()
107 fault: in __dat_user_addr()
179 goto fault; in strnlen_user_pt()
182 goto fault; in strnlen_user_pt()
195 fault: in strnlen_user_pt()
[all …]
/arch/arm/mm/
Dfault.c228 int fault; in __do_page_fault() local
231 fault = VM_FAULT_BADMAP; in __do_page_fault()
243 fault = VM_FAULT_BADACCESS; in __do_page_fault()
255 return fault; in __do_page_fault()
263 int fault, sig, code; in do_page_fault() local
309 fault = __do_page_fault(mm, addr, fsr, flags, tsk); in do_page_fault()
315 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
325 if (!(fault & VM_FAULT_ERROR) && flags & FAULT_FLAG_ALLOW_RETRY) { in do_page_fault()
326 if (fault & VM_FAULT_MAJOR) { in do_page_fault()
335 if (fault & VM_FAULT_RETRY) { in do_page_fault()
[all …]
Dalignment.c219 goto fault; \
240 goto fault; \
272 goto fault; \
314 goto fault; \
375 fault: in do_alignment_ldrhstrh()
431 fault: in do_alignment_ldrdstrd()
462 fault: in do_alignment_ldrstr()
560 fault: in do_alignment_ldmstm()
753 unsigned int fault; in do_alignment() local
765 fault = probe_kernel_address(ptr, tinstr); in do_alignment()
[all …]
/arch/mips/kernel/
Dunaligned.c189 goto fault; in emulate_load_store_insn()
221 goto fault; in emulate_load_store_insn()
260 goto fault; in emulate_load_store_insn()
290 goto fault; in emulate_load_store_insn()
324 goto fault; in emulate_load_store_insn()
364 goto fault; in emulate_load_store_insn()
406 goto fault; in emulate_load_store_insn()
448 goto fault; in emulate_load_store_insn()
479 goto fault; in emulate_load_store_insn()
518 goto fault; in emulate_load_store_insn()
[all …]
/arch/parisc/mm/
Dfault.c177 int fault; in do_page_fault() local
204 fault = handle_mm_fault(mm, vma, address, (acc_type & VM_WRITE) ? FAULT_FLAG_WRITE : 0); in do_page_fault()
205 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
211 if (fault & VM_FAULT_OOM) in do_page_fault()
213 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
217 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/microblaze/mm/
Dfault.c94 int fault; in do_page_fault() local
213 fault = handle_mm_fault(mm, vma, address, is_write ? FAULT_FLAG_WRITE : 0); in do_page_fault()
214 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
215 if (fault & VM_FAULT_OOM) in do_page_fault()
217 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
221 if (unlikely(fault & VM_FAULT_MAJOR)) in do_page_fault()
/arch/mips/mm/
Dfault.c43 int fault; in do_page_fault() local
149 fault = handle_mm_fault(mm, vma, address, flags); in do_page_fault()
151 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
155 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
156 if (fault & VM_FAULT_OOM) in do_page_fault()
158 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
163 if (fault & VM_FAULT_MAJOR) { in do_page_fault()
172 if (fault & VM_FAULT_RETRY) { in do_page_fault()
/arch/alpha/mm/
Dfault.c90 int fault, si_code = SEGV_MAPERR; in do_page_fault() local
147 fault = handle_mm_fault(mm, vma, address, cause > 0 ? FAULT_FLAG_WRITE : 0); in do_page_fault()
149 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
150 if (fault & VM_FAULT_OOM) in do_page_fault()
152 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
156 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/score/mm/
Dfault.c51 int fault; in do_page_fault() local
109 fault = handle_mm_fault(mm, vma, address, write); in do_page_fault()
110 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
111 if (fault & VM_FAULT_OOM) in do_page_fault()
113 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
117 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/avr32/mm/
Dfault.c67 int fault; in do_page_fault() local
135 fault = handle_mm_fault(mm, vma, address, writeaccess ? FAULT_FLAG_WRITE : 0); in do_page_fault()
136 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
137 if (fault & VM_FAULT_OOM) in do_page_fault()
139 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
143 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/openrisc/mm/
Dfault.c56 int fault; in do_page_fault() local
162 fault = handle_mm_fault(mm, vma, address, write_acc); in do_page_fault()
163 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
164 if (fault & VM_FAULT_OOM) in do_page_fault()
166 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
171 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/cris/mm/
Dfault.c60 int fault; in do_page_fault() local
166 fault = handle_mm_fault(mm, vma, address, (writeaccess & 1) ? FAULT_FLAG_WRITE : 0); in do_page_fault()
167 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
168 if (fault & VM_FAULT_OOM) in do_page_fault()
170 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
174 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/xtensa/mm/
Dfault.c46 int fault; in do_page_fault() local
107 fault = handle_mm_fault(mm, vma, address, is_write ? FAULT_FLAG_WRITE : 0); in do_page_fault()
108 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
109 if (fault & VM_FAULT_OOM) in do_page_fault()
111 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
115 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/m32r/mm/
DMakefile6 obj-y := init.o fault.o mmu.o extable.o ioremap.o cache.o page.o
8 obj-y := init.o fault-nommu.o mmu.o extable.o ioremap-nommu.o cache.o page.o
/arch/ia64/mm/
Dfault.c83 int fault; in ia64_do_page_fault() local
156 fault = handle_mm_fault(mm, vma, address, (mask & VM_WRITE) ? FAULT_FLAG_WRITE : 0); in ia64_do_page_fault()
157 if (unlikely(fault & VM_FAULT_ERROR)) { in ia64_do_page_fault()
163 if (fault & VM_FAULT_OOM) { in ia64_do_page_fault()
165 } else if (fault & VM_FAULT_SIGBUS) { in ia64_do_page_fault()
171 if (fault & VM_FAULT_MAJOR) in ia64_do_page_fault()
/arch/frv/mm/
Dfault.c42 int fault; in do_page_fault() local
165 fault = handle_mm_fault(mm, vma, ear0, write ? FAULT_FLAG_WRITE : 0); in do_page_fault()
166 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
167 if (fault & VM_FAULT_OOM) in do_page_fault()
169 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
173 if (fault & VM_FAULT_MAJOR) in do_page_fault()
/arch/x86/mm/
Dfault.c166 struct task_struct *tsk, int fault) in force_sig_info_fault() argument
175 if (fault & VM_FAULT_HWPOISON_LARGE) in force_sig_info_fault()
176 lsb = hstate_index_to_shift(VM_FAULT_GET_HINDEX(fault)); in force_sig_info_fault()
177 if (fault & VM_FAULT_HWPOISON) in force_sig_info_fault()
825 unsigned int fault) in do_sigbus() argument
848 if (fault & (VM_FAULT_HWPOISON|VM_FAULT_HWPOISON_LARGE)) { in do_sigbus()
855 force_sig_info_fault(SIGBUS, code, address, tsk, fault); in do_sigbus()
860 unsigned long address, unsigned int fault) in mm_fault_error() argument
867 if (!(fault & VM_FAULT_RETRY)) in mm_fault_error()
873 if (!(fault & VM_FAULT_ERROR)) in mm_fault_error()
[all …]
/arch/um/kernel/
Dtrap.c66 int fault; in handle_page_fault() local
68 fault = handle_mm_fault(mm, vma, address, is_write ? FAULT_FLAG_WRITE : 0); in handle_page_fault()
69 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault()
70 if (fault & VM_FAULT_OOM) { in handle_page_fault()
72 } else if (fault & VM_FAULT_SIGBUS) { in handle_page_fault()
78 if (fault & VM_FAULT_MAJOR) in handle_page_fault()

12345