/arch/arm64/kernel/ |
D | entry-common.c | 116 static void noinstr el1_abort(struct pt_regs *regs, unsigned long esr) in el1_abort() argument 122 do_mem_abort(far, esr, regs); in el1_abort() 127 static void noinstr el1_pc(struct pt_regs *regs, unsigned long esr) in el1_pc() argument 133 do_sp_pc_abort(far, esr, regs); in el1_pc() 147 static void noinstr el1_inv(struct pt_regs *regs, unsigned long esr) in el1_inv() argument 151 bad_mode(regs, 0, esr); in el1_inv() 180 static void noinstr el1_dbg(struct pt_regs *regs, unsigned long esr) in el1_dbg() argument 185 do_debug_exception(far, esr, regs); in el1_dbg() 189 static void noinstr el1_fpac(struct pt_regs *regs, unsigned long esr) in el1_fpac() argument 193 do_ptrauth_fault(regs, esr); in el1_fpac() [all …]
|
D | traps.c | 157 unsigned int esr = tsk->thread.fault_code; in arm64_show_signal() local 167 if (esr) in arm64_show_signal() 168 pr_cont("%s, ESR 0x%08x, ", esr_get_class_string(esr), esr); in arm64_show_signal() 422 void do_ptrauth_fault(struct pt_regs *regs, unsigned int esr) in do_ptrauth_fault() argument 428 trace_android_rvh_do_ptrauth_fault(regs, esr, user_mode(regs)); in do_ptrauth_fault() 430 force_signal_inject(SIGILL, ILL_ILLOPN, regs->pc, esr); in do_ptrauth_fault() 454 static void user_cache_maint_handler(unsigned int esr, struct pt_regs *regs) in user_cache_maint_handler() argument 457 int rt = ESR_ELx_SYS64_ISS_RT(esr); in user_cache_maint_handler() 458 int crm = (esr & ESR_ELx_SYS64_ISS_CRM_MASK) >> ESR_ELx_SYS64_ISS_CRM_SHIFT; in user_cache_maint_handler() 494 static void ctr_read_handler(unsigned int esr, struct pt_regs *regs) in ctr_read_handler() argument [all …]
|
D | debug-monitors.c | 205 static int call_step_hook(struct pt_regs *regs, unsigned int esr) in call_step_hook() argument 218 retval = hook->fn(regs, esr); in call_step_hook() 241 static int single_step_handler(unsigned long unused, unsigned int esr, in single_step_handler() argument 253 if (!handler_found && call_step_hook(regs, esr) == DBG_HOOK_HANDLED) in single_step_handler() 303 static int call_break_hook(struct pt_regs *regs, unsigned int esr) in call_break_hook() argument 307 int (*fn)(struct pt_regs *regs, unsigned int esr) = NULL; in call_break_hook() 316 unsigned int comment = esr & ESR_ELx_BRK64_ISS_COMMENT_MASK; in call_break_hook() 322 return fn ? fn(regs, esr) : DBG_HOOK_ERROR; in call_break_hook() 326 static int brk_handler(unsigned long unused, unsigned int esr, in brk_handler() argument 329 if (call_break_hook(regs, esr) == DBG_HOOK_HANDLED) in brk_handler()
|
D | fpsimd.c | 937 void do_sve_acc(unsigned int esr, struct pt_regs *regs) in do_sve_acc() argument 964 void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs) in do_fpsimd_acc() argument 973 void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs) in do_fpsimd_exc() argument 977 if (esr & ESR_ELx_FP_EXC_TFV) { in do_fpsimd_exc() 978 if (esr & FPEXC_IOF) in do_fpsimd_exc() 980 else if (esr & FPEXC_DZF) in do_fpsimd_exc() 982 else if (esr & FPEXC_OFF) in do_fpsimd_exc() 984 else if (esr & FPEXC_UFF) in do_fpsimd_exc() 986 else if (esr & FPEXC_IXF) in do_fpsimd_exc()
|
D | kgdb.c | 236 static int kgdb_brk_fn(struct pt_regs *regs, unsigned int esr) in kgdb_brk_fn() argument 243 static int kgdb_compiled_brk_fn(struct pt_regs *regs, unsigned int esr) in NOKPROBE_SYMBOL() 252 static int kgdb_step_brk_fn(struct pt_regs *regs, unsigned int esr) in kgdb_step_brk_fn() argument
|
/arch/arm64/mm/ |
D | fault.c | 48 int (*fn)(unsigned long far, unsigned int esr, 58 static inline const struct fault_info *esr_to_fault_info(unsigned int esr) in esr_to_fault_info() argument 60 return fault_info + (esr & ESR_ELx_FSC); in esr_to_fault_info() 63 static inline const struct fault_info *esr_to_debug_fault_info(unsigned int esr) in esr_to_debug_fault_info() argument 65 return debug_fault_info + DBG_ESR_EVT(esr); in esr_to_debug_fault_info() 68 static void data_abort_decode(unsigned int esr) in data_abort_decode() argument 72 if (esr & ESR_ELx_ISV) { in data_abort_decode() 74 1U << ((esr & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT)); in data_abort_decode() 76 (esr & ESR_ELx_SSE) >> ESR_ELx_SSE_SHIFT, in data_abort_decode() 77 (esr & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT); in data_abort_decode() [all …]
|
/arch/arm64/include/asm/ |
D | exception.h | 19 unsigned int esr = ESR_ELx_EC_SERROR << ESR_ELx_EC_SHIFT; in disr_to_esr() local 22 esr |= (disr & DISR_EL1_ESR_MASK); in disr_to_esr() 24 esr |= (disr & ESR_ELx_ISS_MASK); in disr_to_esr() 26 return esr; in disr_to_esr() 35 void do_mem_abort(unsigned long far, unsigned int esr, struct pt_regs *regs); 38 asmlinkage void bad_mode(struct pt_regs *regs, int reason, unsigned int esr); 39 void do_debug_exception(unsigned long addr_if_watchpoint, unsigned int esr, 41 void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs); 42 void do_sve_acc(unsigned int esr, struct pt_regs *regs); 43 void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs); [all …]
|
D | traps.h | 60 static inline bool arm64_is_ras_serror(u32 esr) in arm64_is_ras_serror() argument 64 if (esr & ESR_ELx_IDS) in arm64_is_ras_serror() 80 static inline u32 arm64_ras_serror_get_severity(u32 esr) in arm64_ras_serror_get_severity() argument 82 u32 aet = esr & ESR_ELx_AET; in arm64_ras_serror_get_severity() 84 if (!arm64_is_ras_serror(esr)) { in arm64_ras_serror_get_severity() 93 if ((esr & ESR_ELx_FSC) != ESR_ELx_FSC_SERROR) { in arm64_ras_serror_get_severity() 101 bool arm64_is_fatal_ras_serror(struct pt_regs *regs, unsigned int esr); 102 void __noreturn arm64_serror_panic(struct pt_regs *regs, u32 esr);
|
D | esr.h | 73 #define ESR_ELx_EC(esr) (((esr) & ESR_ELx_EC_MASK) >> ESR_ELx_EC_SHIFT) argument 190 #define ESR_ELx_SYS64_ISS_RT(esr) \ argument 191 (((esr) & ESR_ELx_SYS64_ISS_RT_MASK) >> ESR_ELx_SYS64_ISS_RT_SHIFT) 328 static inline bool esr_is_data_abort(u32 esr) in esr_is_data_abort() argument 330 const u32 ec = ESR_ELx_EC(esr); in esr_is_data_abort() 335 const char *esr_get_class_string(u32 esr);
|
D | debug-monitors.h | 79 int (*fn)(struct pt_regs *regs, unsigned int esr); 90 int (*fn)(struct pt_regs *regs, unsigned int esr);
|
D | kvm_emulate.h | 229 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_get_condition() local 231 if (esr & ESR_ELx_CV) in kvm_vcpu_get_condition() 232 return (esr & ESR_ELx_COND_MASK) >> ESR_ELx_COND_SHIFT; in kvm_vcpu_get_condition() 360 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_sys_get_rt() local 361 return ESR_ELx_SYS64_ISS_RT(esr); in kvm_vcpu_sys_get_rt()
|
D | kvm_ras.h | 17 static inline int kvm_handle_guest_sea(phys_addr_t addr, unsigned int esr) in kvm_handle_guest_sea() argument
|
/arch/arm64/kvm/ |
D | inject_fault.c | 21 u32 esr = 0; in inject_abt64() local 34 esr |= ESR_ELx_IL; in inject_abt64() 41 esr |= (ESR_ELx_EC_IABT_LOW << ESR_ELx_EC_SHIFT); in inject_abt64() 43 esr |= (ESR_ELx_EC_IABT_CUR << ESR_ELx_EC_SHIFT); in inject_abt64() 46 esr |= ESR_ELx_EC_DABT_LOW << ESR_ELx_EC_SHIFT; in inject_abt64() 48 vcpu_write_sys_reg(vcpu, esr | ESR_ELx_FSC_EXTABT, ESR_EL1); in inject_abt64() 53 u32 esr = (ESR_ELx_EC_UNKNOWN << ESR_ELx_EC_SHIFT); in inject_undef64() local 64 esr |= ESR_ELx_IL; in inject_undef64() 66 vcpu_write_sys_reg(vcpu, esr, ESR_EL1); in inject_undef64() 163 void kvm_set_sei_esr(struct kvm_vcpu *vcpu, u64 esr) in kvm_set_sei_esr() argument [all …]
|
D | handle_exit.c | 29 static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u32 esr) in kvm_handle_guest_serror() argument 31 if (!arm64_is_ras_serror(esr) || arm64_is_fatal_ras_serror(NULL, esr)) in kvm_handle_guest_serror() 121 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_guest_debug() local 125 run->debug.arch.hsr = esr; in kvm_handle_guest_debug() 127 switch (ESR_ELx_EC(esr)) { in kvm_handle_guest_debug() 138 __func__, (unsigned int) esr); in kvm_handle_guest_debug() 148 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_unknown_ec() local 151 esr, esr_get_class_string(esr)); in kvm_handle_unknown_ec() 202 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_get_exit_handler() local 203 u8 esr_ec = ESR_ELx_EC(esr); in kvm_get_exit_handler()
|
D | sys_regs.c | 2183 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_cp_64() local 2185 int Rt2 = (esr >> 10) & 0x1f; in kvm_handle_cp_64() 2187 params.CRm = (esr >> 1) & 0xf; in kvm_handle_cp_64() 2188 params.is_write = ((esr & 1) == 0); in kvm_handle_cp_64() 2191 params.Op1 = (esr >> 16) & 0xf; in kvm_handle_cp_64() 2233 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_cp_32() local 2236 params.CRm = (esr >> 1) & 0xf; in kvm_handle_cp_32() 2238 params.is_write = ((esr & 1) == 0); in kvm_handle_cp_32() 2239 params.CRn = (esr >> 10) & 0xf; in kvm_handle_cp_32() 2241 params.Op1 = (esr >> 14) & 0x7; in kvm_handle_cp_32() [all …]
|
/arch/alpha/kernel/ |
D | core_lca.c | 338 mem_error(unsigned long esr, unsigned long ear) in mem_error() argument 341 ((esr & ESR_CEE) ? "Correctable" : in mem_error() 342 (esr & ESR_UEE) ? "Uncorrectable" : "A"), in mem_error() 343 (esr & ESR_WRE) ? "write" : "read", in mem_error() 344 (esr & ESR_SOR) ? "memory" : "b-cache", in mem_error() 346 if (esr & ESR_CTE) { in mem_error() 349 if (esr & ESR_MSE) { in mem_error() 352 if (esr & ESR_MHE) { in mem_error() 355 if (esr & ESR_NXM) { in mem_error() 435 if (el.s->esr & ESR_EAV) { in lca_machine_check() [all …]
|
/arch/sh/boards/mach-dreamcast/ |
D | irq.c | 90 __u32 esr = ESR_BASE + (LEVEL(irq) << 2); in mask_ack_systemasic_irq() local 92 outl((1 << EVENT_BIT(irq)), esr); in mask_ack_systemasic_irq() 107 __u32 emr, esr, status, level; in systemasic_irq_demux() local 124 esr = ESR_BASE + (level << 2); in systemasic_irq_demux() 127 status = inl(esr); in systemasic_irq_demux()
|
/arch/arm64/kvm/hyp/include/hyp/ |
D | switch.h | 163 static inline bool __get_fault_info(u64 esr, struct kvm_vcpu_fault_info *fault) in __get_fault_info() argument 180 if (!(esr & ESR_ELx_S1PTW) && in __get_fault_info() 182 (esr & ESR_ELx_FSC_TYPE) == FSC_PERM)) { in __get_fault_info() 197 u64 esr; in __populate_fault_info() local 199 esr = vcpu->arch.fault.esr_el2; in __populate_fault_info() 200 ec = ESR_ELx_EC(esr); in __populate_fault_info() 205 return __get_fault_info(esr, &vcpu->arch.fault); in __populate_fault_info() 347 static inline bool esr_is_ptrauth_trap(u32 esr) in esr_is_ptrauth_trap() argument 349 u32 ec = ESR_ELx_EC(esr); in esr_is_ptrauth_trap() 357 switch (esr_sys64_to_sysreg(esr)) { in esr_is_ptrauth_trap()
|
/arch/mips/include/asm/octeon/ |
D | cvmx-sli-defs.h | 107 __BITFIELD_FIELD(uint64_t esr:2, 119 __BITFIELD_FIELD(uint64_t esr:2,
|
/arch/arm64/kernel/probes/ |
D | uprobes.c | 169 unsigned int esr) in uprobe_breakpoint_handler() argument 178 unsigned int esr) in uprobe_single_step_handler() argument
|
/arch/arm64/kvm/hyp/nvhe/ |
D | mem_protect.c | 281 u64 esr, addr; in handle_host_mem_abort() local 284 esr = read_sysreg_el2(SYS_ESR); in handle_host_mem_abort() 285 if (!__get_fault_info(esr, &fault)) in handle_host_mem_abort()
|
D | hyp-main.c | 235 u64 esr = read_sysreg_el2(SYS_ESR); in handle_trap() local 237 switch (ESR_ELx_EC(esr)) { in handle_trap()
|
/arch/powerpc/include/uapi/asm/ |
D | kvm_para.h | 58 __u32 esr; member
|
/arch/microblaze/include/uapi/asm/ |
D | ptrace.h | 53 microblaze_reg_t esr; member
|
/arch/arm64/kvm/hyp/ |
D | vgic-v3-sr.c | 476 u32 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_get_group() local 477 u8 crm = (esr & ESR_ELx_SYS64_ISS_CRM_MASK) >> ESR_ELx_SYS64_ISS_CRM_SHIFT; in __vgic_v3_get_group() 1022 u32 esr; in __vgic_v3_perform_cpuif_access() local 1028 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_perform_cpuif_access() 1035 sysreg = esr_cp15_to_sysreg(esr); in __vgic_v3_perform_cpuif_access() 1037 sysreg = esr_sys64_to_sysreg(esr); in __vgic_v3_perform_cpuif_access() 1040 is_read = (esr & ESR_ELx_SYS64_ISS_DIR_MASK) == ESR_ELx_SYS64_ISS_DIR_READ; in __vgic_v3_perform_cpuif_access()
|