Lines Matching full:fault
2 * arch/microblaze/mm/fault.c
6 * Derived from "arch/ppc/mm/fault.c"
9 * Derived from "arch/i386/mm/fault.c"
71 /* Are we prepared to handle this fault? */ in bad_page_fault()
83 * The error_code parameter is ESR for a data fault,
84 * 0 for an instruction fault.
93 vm_fault_t fault; in do_page_fault() local
115 pr_emerg("Page fault in user mode with faulthandler_disabled(), mm = %p\n", in do_page_fault()
119 die("Weird page fault", regs, SIGSEGV); in do_page_fault()
128 * erroneous fault occurring in a code path which already holds mmap_sem in do_page_fault()
129 * we will deadlock attempting to validate the fault against the in do_page_fault()
172 /* get user regs even if this fault is in kernel mode */ in do_page_fault()
206 /* protection fault */ in do_page_fault()
214 * If for any reason at all we couldn't handle the fault, in do_page_fault()
216 * the fault. in do_page_fault()
218 fault = handle_mm_fault(vma, address, flags); in do_page_fault()
220 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault()
223 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault()
224 if (fault & VM_FAULT_OOM) in do_page_fault()
226 else if (fault & VM_FAULT_SIGSEGV) in do_page_fault()
228 else if (fault & VM_FAULT_SIGBUS) in do_page_fault()
234 if (unlikely(fault & VM_FAULT_MAJOR)) in do_page_fault()
238 if (fault & VM_FAULT_RETRY) { in do_page_fault()
279 * us unable to handle the page fault gracefully. in do_page_fault()