/arch/arm/lib/ |
D | uaccess.S | 34 USER( strbt r3, [r0], #1) @ May fault 36 USER( strgebt r3, [r0], #1) @ May fault 38 USER( strgtbt r3, [r0], #1) @ May fault 61 USER( strt r3, [r0], #4) @ May fault 76 stmia r0!, {r3 - r6} @ Shouldnt fault 79 stmia r0!, {r3 - r6} @ Shouldnt fault 84 stmgeia r0!, {r3 - r6} @ Shouldnt fault 87 stmneia r0!, {r3 - r4} @ Shouldnt fault 90 strnet r3, [r0], #4 @ Shouldnt fault 97 USER( strbt r3, [r0], #1) @ May fault [all …]
|
/arch/arm/mm/ |
D | fault.c | 182 int fault, mask; in __do_page_fault() local 185 fault = VM_FAULT_BADMAP; in __do_page_fault() 201 fault = VM_FAULT_BADACCESS; in __do_page_fault() 211 fault = handle_mm_fault(mm, vma, addr & PAGE_MASK, fsr & (1 << 11)); in __do_page_fault() 212 if (unlikely(fault & VM_FAULT_ERROR)) { in __do_page_fault() 213 if (fault & VM_FAULT_OOM) in __do_page_fault() 215 else if (fault & VM_FAULT_SIGBUS) in __do_page_fault() 216 return fault; in __do_page_fault() 219 if (fault & VM_FAULT_MAJOR) in __do_page_fault() 223 return fault; in __do_page_fault() [all …]
|
D | alignment.c | 178 goto fault; \ 199 goto fault; \ 229 goto fault; \ 265 goto fault; \ 326 fault: in do_alignment_ldrhstrh() 372 fault: in do_alignment_ldrdstrd() 403 fault: in do_alignment_ldrstr() 501 fault: in do_alignment_ldmstm() 631 unsigned int fault; in do_alignment() local 639 fault = __get_user(tinstr, (u16 *)(instrptr & ~1)); in do_alignment() [all …]
|
/arch/mn10300/kernel/ |
D | gdb-stub.c | 521 goto fault; in gdbstub_single_step() 527 goto fault; in gdbstub_single_step() 529 goto fault; in gdbstub_single_step() 532 goto fault; in gdbstub_single_step() 538 goto fault; in gdbstub_single_step() 541 goto fault; in gdbstub_single_step() 548 goto fault; in gdbstub_single_step() 556 goto fault; in gdbstub_single_step() 558 goto fault; in gdbstub_single_step() 568 goto fault; in gdbstub_single_step() [all …]
|
/arch/um/kernel/ |
D | uaccess.c | 24 unsigned long fault; in __do_copy_to_user() local 27 fault = __do_user_copy(to, from, n, fault_addr, fault_catcher, in __do_copy_to_user() 32 return n - (fault - (unsigned long) to); in __do_copy_to_user()
|
D | trap.c | 66 int fault; in handle_page_fault() local 68 fault = handle_mm_fault(mm, vma, address, is_write); in handle_page_fault() 69 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault() 70 if (fault & VM_FAULT_OOM) { in handle_page_fault() 72 } else if (fault & VM_FAULT_SIGBUS) { in handle_page_fault() 78 if (fault & VM_FAULT_MAJOR) in handle_page_fault()
|
/arch/m68k/mm/ |
D | fault.c | 90 int write, fault; in do_page_fault() local 158 fault = handle_mm_fault(mm, vma, address, write); in do_page_fault() 160 printk("handle_mm_fault returns %d\n",fault); in do_page_fault() 162 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 163 if (fault & VM_FAULT_OOM) in do_page_fault() 165 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 169 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/mips/kernel/ |
D | unaligned.c | 179 goto fault; in emulate_load_store_insn() 209 goto fault; in emulate_load_store_insn() 243 goto fault; in emulate_load_store_insn() 283 goto fault; in emulate_load_store_insn() 325 goto fault; in emulate_load_store_insn() 367 goto fault; in emulate_load_store_insn() 398 goto fault; in emulate_load_store_insn() 437 goto fault; in emulate_load_store_insn() 479 fault: in emulate_load_store_insn()
|
/arch/s390/lib/ |
D | uaccess_pt.c | 44 int fault; in __handle_fault() local 69 fault = handle_mm_fault(mm, vma, address, write_access); in __handle_fault() 70 if (unlikely(fault & VM_FAULT_ERROR)) { in __handle_fault() 71 if (fault & VM_FAULT_OOM) in __handle_fault() 73 else if (fault & VM_FAULT_SIGBUS) in __handle_fault() 77 if (fault & VM_FAULT_MAJOR) in __handle_fault() 119 goto fault; in __user_copy_pt() 138 fault: in __user_copy_pt() 160 goto fault; in __dat_user_addr() 166 fault: in __dat_user_addr() [all …]
|
/arch/parisc/mm/ |
D | fault.c | 178 int fault; in do_page_fault() local 205 fault = handle_mm_fault(mm, vma, address, (acc_type & VM_WRITE) != 0); in do_page_fault() 206 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 212 if (fault & VM_FAULT_OOM) in do_page_fault() 214 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 218 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/alpha/mm/ |
D | fault.c | 91 int fault, si_code = SEGV_MAPERR; in do_page_fault() local 149 fault = handle_mm_fault(mm, vma, address, cause > 0); in do_page_fault() 151 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 152 if (fault & VM_FAULT_OOM) in do_page_fault() 154 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 158 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/mips/mm/ |
D | fault.c | 42 int fault; in do_page_fault() local 105 fault = handle_mm_fault(mm, vma, address, write); in do_page_fault() 106 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 107 if (fault & VM_FAULT_OOM) in do_page_fault() 109 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 113 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/cris/mm/ |
D | fault.c | 60 int fault; in do_page_fault() local 166 fault = handle_mm_fault(mm, vma, address, writeaccess & 1); in do_page_fault() 167 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 168 if (fault & VM_FAULT_OOM) in do_page_fault() 170 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 174 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/avr32/mm/ |
D | fault.c | 67 int fault; in do_page_fault() local 136 fault = handle_mm_fault(mm, vma, address, writeaccess); in do_page_fault() 137 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 138 if (fault & VM_FAULT_OOM) in do_page_fault() 140 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 144 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/xtensa/mm/ |
D | fault.c | 47 int fault; in do_page_fault() local 109 fault = handle_mm_fault(mm, vma, address, is_write); in do_page_fault() 110 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 111 if (fault & VM_FAULT_OOM) in do_page_fault() 113 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 117 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/ia64/mm/ |
D | fault.c | 83 int fault; in ia64_do_page_fault() local 157 fault = handle_mm_fault(mm, vma, address, (mask & VM_WRITE) != 0); in ia64_do_page_fault() 158 if (unlikely(fault & VM_FAULT_ERROR)) { in ia64_do_page_fault() 164 if (fault & VM_FAULT_OOM) { in ia64_do_page_fault() 166 } else if (fault & VM_FAULT_SIGBUS) { in ia64_do_page_fault() 172 if (fault & VM_FAULT_MAJOR) in ia64_do_page_fault()
|
/arch/frv/mm/ |
D | fault.c | 43 int fault; in do_page_fault() local 166 fault = handle_mm_fault(mm, vma, ear0, write); in do_page_fault() 167 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 168 if (fault & VM_FAULT_OOM) in do_page_fault() 170 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 174 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/m32r/mm/ |
D | Makefile | 6 obj-y := init.o fault.o mmu.o extable.o ioremap.o cache.o page.o 8 obj-y := init.o fault-nommu.o mmu.o extable.o ioremap-nommu.o cache.o page.o
|
D | fault.c | 83 int fault; in do_page_fault() local 199 fault = handle_mm_fault(mm, vma, addr, write); in do_page_fault() 200 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 201 if (fault & VM_FAULT_OOM) in do_page_fault() 203 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 207 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/sh/mm/ |
D | fault_32.c | 37 int fault; in do_page_fault() local 136 fault = handle_mm_fault(mm, vma, address, writeaccess); in do_page_fault() 137 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 138 if (fault & VM_FAULT_OOM) in do_page_fault() 140 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 144 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
D | tlbflush_64.c | 99 int fault; in do_page_fault() local 190 fault = handle_mm_fault(mm, vma, address, writeaccess); in do_page_fault() 191 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 192 if (fault & VM_FAULT_OOM) in do_page_fault() 194 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 198 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/sparc/kernel/ |
D | wuf.S | 311 lda [%l4] ASI_M_MMUREGS, %g0 ! clear fault status 319 or %l5, 0x2, %l5 ! turn on no-fault bit 333 andn %twin_tmp1, 0x2, %twin_tmp1 ! clear no-fault bit 337 lda [%twin_tmp2] ASI_M_MMUREGS, %g0 ! read fault address 340 lda [%twin_tmp2] ASI_M_MMUREGS, %twin_tmp2 ! read fault status 341 andcc %twin_tmp2, 0x2, %g0 ! did fault occur?
|
/arch/mn10300/mm/ |
D | fault.c | 133 int write, fault; in do_page_fault() local 261 fault = handle_mm_fault(mm, vma, address, write); in do_page_fault() 262 if (unlikely(fault & VM_FAULT_ERROR)) { in do_page_fault() 263 if (fault & VM_FAULT_OOM) in do_page_fault() 265 else if (fault & VM_FAULT_SIGBUS) in do_page_fault() 269 if (fault & VM_FAULT_MAJOR) in do_page_fault()
|
/arch/um/sys-i386/ |
D | Makefile | 5 obj-y = bug.o bugs.o checksum.o delay.o fault.o ksyms.o ldt.o ptrace.o \ 13 USER_OBJS := bugs.o ptrace_user.o fault.o
|
/arch/sparc/mm/ |
D | fault_64.c | 257 int si_code, fault_code, fault; in do_sparc64_fault() local 401 fault = handle_mm_fault(mm, vma, address, (fault_code & FAULT_CODE_WRITE)); in do_sparc64_fault() 402 if (unlikely(fault & VM_FAULT_ERROR)) { in do_sparc64_fault() 403 if (fault & VM_FAULT_OOM) in do_sparc64_fault() 405 else if (fault & VM_FAULT_SIGBUS) in do_sparc64_fault() 409 if (fault & VM_FAULT_MAJOR) in do_sparc64_fault()
|