Home
last modified time | relevance | path

Searched refs:esr (Results 1 – 25 of 48) sorted by relevance

12

/arch/arm64/kernel/
Dentry-common.c116 static void noinstr el1_abort(struct pt_regs *regs, unsigned long esr) in el1_abort() argument
122 do_mem_abort(far, esr, regs); in el1_abort()
127 static void noinstr el1_pc(struct pt_regs *regs, unsigned long esr) in el1_pc() argument
133 do_sp_pc_abort(far, esr, regs); in el1_pc()
147 static void noinstr el1_inv(struct pt_regs *regs, unsigned long esr) in el1_inv() argument
151 bad_mode(regs, 0, esr); in el1_inv()
180 static void noinstr el1_dbg(struct pt_regs *regs, unsigned long esr) in el1_dbg() argument
185 do_debug_exception(far, esr, regs); in el1_dbg()
189 static void noinstr el1_fpac(struct pt_regs *regs, unsigned long esr) in el1_fpac() argument
193 do_ptrauth_fault(regs, esr); in el1_fpac()
[all …]
Dtraps.c157 unsigned int esr = tsk->thread.fault_code; in arm64_show_signal() local
167 if (esr) in arm64_show_signal()
168 pr_cont("%s, ESR 0x%08x, ", esr_get_class_string(esr), esr); in arm64_show_signal()
422 void do_ptrauth_fault(struct pt_regs *regs, unsigned int esr) in do_ptrauth_fault() argument
428 trace_android_rvh_do_ptrauth_fault(regs, esr, user_mode(regs)); in do_ptrauth_fault()
430 force_signal_inject(SIGILL, ILL_ILLOPN, regs->pc, esr); in do_ptrauth_fault()
454 static void user_cache_maint_handler(unsigned int esr, struct pt_regs *regs) in user_cache_maint_handler() argument
457 int rt = ESR_ELx_SYS64_ISS_RT(esr); in user_cache_maint_handler()
458 int crm = (esr & ESR_ELx_SYS64_ISS_CRM_MASK) >> ESR_ELx_SYS64_ISS_CRM_SHIFT; in user_cache_maint_handler()
494 static void ctr_read_handler(unsigned int esr, struct pt_regs *regs) in ctr_read_handler() argument
[all …]
Ddebug-monitors.c205 static int call_step_hook(struct pt_regs *regs, unsigned int esr) in call_step_hook() argument
218 retval = hook->fn(regs, esr); in call_step_hook()
241 static int single_step_handler(unsigned long unused, unsigned int esr, in single_step_handler() argument
253 if (!handler_found && call_step_hook(regs, esr) == DBG_HOOK_HANDLED) in single_step_handler()
303 static int call_break_hook(struct pt_regs *regs, unsigned int esr) in call_break_hook() argument
307 int (*fn)(struct pt_regs *regs, unsigned int esr) = NULL; in call_break_hook()
316 unsigned int comment = esr & ESR_ELx_BRK64_ISS_COMMENT_MASK; in call_break_hook()
322 return fn ? fn(regs, esr) : DBG_HOOK_ERROR; in call_break_hook()
326 static int brk_handler(unsigned long unused, unsigned int esr, in brk_handler() argument
329 if (call_break_hook(regs, esr) == DBG_HOOK_HANDLED) in brk_handler()
Dfpsimd.c937 void do_sve_acc(unsigned int esr, struct pt_regs *regs) in do_sve_acc() argument
964 void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs) in do_fpsimd_acc() argument
973 void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs) in do_fpsimd_exc() argument
977 if (esr & ESR_ELx_FP_EXC_TFV) { in do_fpsimd_exc()
978 if (esr & FPEXC_IOF) in do_fpsimd_exc()
980 else if (esr & FPEXC_DZF) in do_fpsimd_exc()
982 else if (esr & FPEXC_OFF) in do_fpsimd_exc()
984 else if (esr & FPEXC_UFF) in do_fpsimd_exc()
986 else if (esr & FPEXC_IXF) in do_fpsimd_exc()
Dkgdb.c236 static int kgdb_brk_fn(struct pt_regs *regs, unsigned int esr) in kgdb_brk_fn() argument
243 static int kgdb_compiled_brk_fn(struct pt_regs *regs, unsigned int esr) in NOKPROBE_SYMBOL()
252 static int kgdb_step_brk_fn(struct pt_regs *regs, unsigned int esr) in kgdb_step_brk_fn() argument
/arch/arm64/mm/
Dfault.c48 int (*fn)(unsigned long far, unsigned int esr,
58 static inline const struct fault_info *esr_to_fault_info(unsigned int esr) in esr_to_fault_info() argument
60 return fault_info + (esr & ESR_ELx_FSC); in esr_to_fault_info()
63 static inline const struct fault_info *esr_to_debug_fault_info(unsigned int esr) in esr_to_debug_fault_info() argument
65 return debug_fault_info + DBG_ESR_EVT(esr); in esr_to_debug_fault_info()
68 static void data_abort_decode(unsigned int esr) in data_abort_decode() argument
72 if (esr & ESR_ELx_ISV) { in data_abort_decode()
74 1U << ((esr & ESR_ELx_SAS) >> ESR_ELx_SAS_SHIFT)); in data_abort_decode()
76 (esr & ESR_ELx_SSE) >> ESR_ELx_SSE_SHIFT, in data_abort_decode()
77 (esr & ESR_ELx_SRT_MASK) >> ESR_ELx_SRT_SHIFT); in data_abort_decode()
[all …]
/arch/arm64/include/asm/
Dexception.h19 unsigned int esr = ESR_ELx_EC_SERROR << ESR_ELx_EC_SHIFT; in disr_to_esr() local
22 esr |= (disr & DISR_EL1_ESR_MASK); in disr_to_esr()
24 esr |= (disr & ESR_ELx_ISS_MASK); in disr_to_esr()
26 return esr; in disr_to_esr()
35 void do_mem_abort(unsigned long far, unsigned int esr, struct pt_regs *regs);
38 asmlinkage void bad_mode(struct pt_regs *regs, int reason, unsigned int esr);
39 void do_debug_exception(unsigned long addr_if_watchpoint, unsigned int esr,
41 void do_fpsimd_acc(unsigned int esr, struct pt_regs *regs);
42 void do_sve_acc(unsigned int esr, struct pt_regs *regs);
43 void do_fpsimd_exc(unsigned int esr, struct pt_regs *regs);
[all …]
Dtraps.h60 static inline bool arm64_is_ras_serror(u32 esr) in arm64_is_ras_serror() argument
64 if (esr & ESR_ELx_IDS) in arm64_is_ras_serror()
80 static inline u32 arm64_ras_serror_get_severity(u32 esr) in arm64_ras_serror_get_severity() argument
82 u32 aet = esr & ESR_ELx_AET; in arm64_ras_serror_get_severity()
84 if (!arm64_is_ras_serror(esr)) { in arm64_ras_serror_get_severity()
93 if ((esr & ESR_ELx_FSC) != ESR_ELx_FSC_SERROR) { in arm64_ras_serror_get_severity()
101 bool arm64_is_fatal_ras_serror(struct pt_regs *regs, unsigned int esr);
102 void __noreturn arm64_serror_panic(struct pt_regs *regs, u32 esr);
Desr.h73 #define ESR_ELx_EC(esr) (((esr) & ESR_ELx_EC_MASK) >> ESR_ELx_EC_SHIFT) argument
190 #define ESR_ELx_SYS64_ISS_RT(esr) \ argument
191 (((esr) & ESR_ELx_SYS64_ISS_RT_MASK) >> ESR_ELx_SYS64_ISS_RT_SHIFT)
328 static inline bool esr_is_data_abort(u32 esr) in esr_is_data_abort() argument
330 const u32 ec = ESR_ELx_EC(esr); in esr_is_data_abort()
335 const char *esr_get_class_string(u32 esr);
Ddebug-monitors.h79 int (*fn)(struct pt_regs *regs, unsigned int esr);
90 int (*fn)(struct pt_regs *regs, unsigned int esr);
Dkvm_emulate.h229 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_get_condition() local
231 if (esr & ESR_ELx_CV) in kvm_vcpu_get_condition()
232 return (esr & ESR_ELx_COND_MASK) >> ESR_ELx_COND_SHIFT; in kvm_vcpu_get_condition()
360 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_vcpu_sys_get_rt() local
361 return ESR_ELx_SYS64_ISS_RT(esr); in kvm_vcpu_sys_get_rt()
Dkvm_ras.h17 static inline int kvm_handle_guest_sea(phys_addr_t addr, unsigned int esr) in kvm_handle_guest_sea() argument
/arch/arm64/kvm/
Dinject_fault.c21 u32 esr = 0; in inject_abt64() local
34 esr |= ESR_ELx_IL; in inject_abt64()
41 esr |= (ESR_ELx_EC_IABT_LOW << ESR_ELx_EC_SHIFT); in inject_abt64()
43 esr |= (ESR_ELx_EC_IABT_CUR << ESR_ELx_EC_SHIFT); in inject_abt64()
46 esr |= ESR_ELx_EC_DABT_LOW << ESR_ELx_EC_SHIFT; in inject_abt64()
48 vcpu_write_sys_reg(vcpu, esr | ESR_ELx_FSC_EXTABT, ESR_EL1); in inject_abt64()
53 u32 esr = (ESR_ELx_EC_UNKNOWN << ESR_ELx_EC_SHIFT); in inject_undef64() local
64 esr |= ESR_ELx_IL; in inject_undef64()
66 vcpu_write_sys_reg(vcpu, esr, ESR_EL1); in inject_undef64()
163 void kvm_set_sei_esr(struct kvm_vcpu *vcpu, u64 esr) in kvm_set_sei_esr() argument
[all …]
Dhandle_exit.c29 static void kvm_handle_guest_serror(struct kvm_vcpu *vcpu, u32 esr) in kvm_handle_guest_serror() argument
31 if (!arm64_is_ras_serror(esr) || arm64_is_fatal_ras_serror(NULL, esr)) in kvm_handle_guest_serror()
121 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_guest_debug() local
125 run->debug.arch.hsr = esr; in kvm_handle_guest_debug()
127 switch (ESR_ELx_EC(esr)) { in kvm_handle_guest_debug()
138 __func__, (unsigned int) esr); in kvm_handle_guest_debug()
148 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_unknown_ec() local
151 esr, esr_get_class_string(esr)); in kvm_handle_unknown_ec()
202 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_get_exit_handler() local
203 u8 esr_ec = ESR_ELx_EC(esr); in kvm_get_exit_handler()
Dsys_regs.c2183 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_cp_64() local
2185 int Rt2 = (esr >> 10) & 0x1f; in kvm_handle_cp_64()
2187 params.CRm = (esr >> 1) & 0xf; in kvm_handle_cp_64()
2188 params.is_write = ((esr & 1) == 0); in kvm_handle_cp_64()
2191 params.Op1 = (esr >> 16) & 0xf; in kvm_handle_cp_64()
2233 u32 esr = kvm_vcpu_get_esr(vcpu); in kvm_handle_cp_32() local
2236 params.CRm = (esr >> 1) & 0xf; in kvm_handle_cp_32()
2238 params.is_write = ((esr & 1) == 0); in kvm_handle_cp_32()
2239 params.CRn = (esr >> 10) & 0xf; in kvm_handle_cp_32()
2241 params.Op1 = (esr >> 14) & 0x7; in kvm_handle_cp_32()
[all …]
/arch/alpha/kernel/
Dcore_lca.c338 mem_error(unsigned long esr, unsigned long ear) in mem_error() argument
341 ((esr & ESR_CEE) ? "Correctable" : in mem_error()
342 (esr & ESR_UEE) ? "Uncorrectable" : "A"), in mem_error()
343 (esr & ESR_WRE) ? "write" : "read", in mem_error()
344 (esr & ESR_SOR) ? "memory" : "b-cache", in mem_error()
346 if (esr & ESR_CTE) { in mem_error()
349 if (esr & ESR_MSE) { in mem_error()
352 if (esr & ESR_MHE) { in mem_error()
355 if (esr & ESR_NXM) { in mem_error()
435 if (el.s->esr & ESR_EAV) { in lca_machine_check()
[all …]
/arch/sh/boards/mach-dreamcast/
Dirq.c90 __u32 esr = ESR_BASE + (LEVEL(irq) << 2); in mask_ack_systemasic_irq() local
92 outl((1 << EVENT_BIT(irq)), esr); in mask_ack_systemasic_irq()
107 __u32 emr, esr, status, level; in systemasic_irq_demux() local
124 esr = ESR_BASE + (level << 2); in systemasic_irq_demux()
127 status = inl(esr); in systemasic_irq_demux()
/arch/arm64/kvm/hyp/include/hyp/
Dswitch.h163 static inline bool __get_fault_info(u64 esr, struct kvm_vcpu_fault_info *fault) in __get_fault_info() argument
180 if (!(esr & ESR_ELx_S1PTW) && in __get_fault_info()
182 (esr & ESR_ELx_FSC_TYPE) == FSC_PERM)) { in __get_fault_info()
197 u64 esr; in __populate_fault_info() local
199 esr = vcpu->arch.fault.esr_el2; in __populate_fault_info()
200 ec = ESR_ELx_EC(esr); in __populate_fault_info()
205 return __get_fault_info(esr, &vcpu->arch.fault); in __populate_fault_info()
347 static inline bool esr_is_ptrauth_trap(u32 esr) in esr_is_ptrauth_trap() argument
349 u32 ec = ESR_ELx_EC(esr); in esr_is_ptrauth_trap()
357 switch (esr_sys64_to_sysreg(esr)) { in esr_is_ptrauth_trap()
/arch/mips/include/asm/octeon/
Dcvmx-sli-defs.h107 __BITFIELD_FIELD(uint64_t esr:2,
119 __BITFIELD_FIELD(uint64_t esr:2,
/arch/arm64/kernel/probes/
Duprobes.c169 unsigned int esr) in uprobe_breakpoint_handler() argument
178 unsigned int esr) in uprobe_single_step_handler() argument
/arch/arm64/kvm/hyp/nvhe/
Dmem_protect.c281 u64 esr, addr; in handle_host_mem_abort() local
284 esr = read_sysreg_el2(SYS_ESR); in handle_host_mem_abort()
285 if (!__get_fault_info(esr, &fault)) in handle_host_mem_abort()
Dhyp-main.c235 u64 esr = read_sysreg_el2(SYS_ESR); in handle_trap() local
237 switch (ESR_ELx_EC(esr)) { in handle_trap()
/arch/powerpc/include/uapi/asm/
Dkvm_para.h58 __u32 esr; member
/arch/microblaze/include/uapi/asm/
Dptrace.h53 microblaze_reg_t esr; member
/arch/arm64/kvm/hyp/
Dvgic-v3-sr.c476 u32 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_get_group() local
477 u8 crm = (esr & ESR_ELx_SYS64_ISS_CRM_MASK) >> ESR_ELx_SYS64_ISS_CRM_SHIFT; in __vgic_v3_get_group()
1022 u32 esr; in __vgic_v3_perform_cpuif_access() local
1028 esr = kvm_vcpu_get_esr(vcpu); in __vgic_v3_perform_cpuif_access()
1035 sysreg = esr_cp15_to_sysreg(esr); in __vgic_v3_perform_cpuif_access()
1037 sysreg = esr_sys64_to_sysreg(esr); in __vgic_v3_perform_cpuif_access()
1040 is_read = (esr & ESR_ELx_SYS64_ISS_DIR_MASK) == ESR_ELx_SYS64_ISS_DIR_READ; in __vgic_v3_perform_cpuif_access()

12