/kernel/linux/linux-5.10/arch/powerpc/kernel/ |
D | head_32.h | 67 li r10, MSR_KERNEL & ~(MSR_IR | MSR_RI) /* can take DTLB miss */ 114 li r10, MSR_KERNEL & ~MSR_IR /* can take exceptions */ 116 li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR) /* can take exceptions */ 139 li r10, MSR_KERNEL & ~(MSR_IR | MSR_RI) /* can take DTLB miss */ 171 LOAD_REG_IMMEDIATE(r10, MSR_KERNEL & ~MSR_IR) /* can take exceptions */ 173 LOAD_REG_IMMEDIATE(r10, MSR_KERNEL & ~(MSR_IR|MSR_DR)) /* can take exceptions */
|
D | head_40x.S | 453 andi. r10,r9,MSR_IR|MSR_PR /* check supervisor + MMU off */ 596 lis r3,(MSR_KERNEL & ~(MSR_IR|MSR_DR))@h 597 ori r3,r3,(MSR_KERNEL & ~(MSR_IR|MSR_DR))@l
|
D | head_book3s_32.S | 205 ori r0,r0,MSR_DR|MSR_IR|MSR_RI 1031 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 1167 li r3, MSR_KERNEL & ~(MSR_IR | MSR_DR) 1190 li r3, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI) 1207 andi. r0,r3,MSR_DR|MSR_IR /* MMU enabled? */
|
D | paca.c | 224 new_paca->kernel_msr = MSR_KERNEL & ~(MSR_IR | MSR_DR); in initialise_paca()
|
D | head_8xx.S | 106 ori r0,r0,MSR_DR|MSR_IR 599 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 727 li r12, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI)
|
D | kvm_emul.S | 300 andi. r31, r31, MSR_DR | MSR_IR
|
D | entry_32.S | 592 LOAD_REG_IMMEDIATE(r11, MSR_KERNEL & ~(MSR_IR|MSR_DR)) 1084 li r10,MSR_IR; \ 1357 li r9,MSR_KERNEL & ~(MSR_IR|MSR_DR) 1365 li r0, MSR_KERNEL & ~MSR_IR /* can take DTLB miss */
|
D | kprobes.c | 280 (!(regs->msr & MSR_IR) || !(regs->msr & MSR_DR))) in kprobe_handler()
|
D | misc_64.S | 384 li r10,MSR_DR|MSR_IR
|
/kernel/linux/linux-5.10/arch/powerpc/kvm/ |
D | book3s_32_mmu.c | 356 if (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_32_esid_to_vsid() 365 switch (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_32_esid_to_vsid() 369 case MSR_IR: in kvmppc_mmu_book3s_32_esid_to_vsid() 375 case MSR_DR|MSR_IR: in kvmppc_mmu_book3s_32_esid_to_vsid()
|
D | book3s_64_mmu.c | 484 if (kvmppc_get_msr(vcpu) & MSR_IR) { in kvmppc_mmu_book3s_64_slbia() 583 if (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_64_esid_to_vsid() 596 switch (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_64_esid_to_vsid() 600 case MSR_IR: in kvmppc_mmu_book3s_64_esid_to_vsid() 606 case MSR_DR|MSR_IR: in kvmppc_mmu_book3s_64_esid_to_vsid()
|
D | book3s_hv_builtin.c | 203 if (kvm_is_radix(vcpu->kvm) && (mfmsr() & MSR_IR)) in kvmppc_h_random() 797 (msr & (MSR_IR|MSR_DR)) == (MSR_IR|MSR_DR) ) { in inject_interrupt() 798 new_msr |= MSR_IR | MSR_DR; in inject_interrupt()
|
D | book3s_rmhandlers.S | 151 li r6, MSR_IR | MSR_DR
|
D | book3s_pr.c | 69 return (msr & (MSR_IR|MSR_DR)) == MSR_DR; in kvmppc_is_split_real() 78 if ((msr & (MSR_IR|MSR_DR)) != MSR_DR) in kvmppc_fixup_split_real() 237 smsr |= MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_PR | MSR_EE; in kvmppc_recalc_shadow_msr() 531 if ((kvmppc_get_msr(vcpu) & (MSR_PR|MSR_IR|MSR_DR)) != in kvmppc_set_msr_pr() 532 (old_msr & (MSR_PR|MSR_IR|MSR_DR))) { in kvmppc_set_msr_pr() 713 bool ir = (kvmppc_get_msr(vcpu) & MSR_IR) ? true : false; in kvmppc_handle_pagefault() 734 switch (kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) { in kvmppc_handle_pagefault() 744 case MSR_IR: in kvmppc_handle_pagefault() 747 if ((kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) == MSR_DR) in kvmppc_handle_pagefault()
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/ |
D | opal-wrappers.S | 27 li r0,MSR_IR|MSR_DR|MSR_LE
|
D | subcore-asm.S | 31 li r5, MSR_IR|MSR_DR
|
D | idle.c | 389 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power7_idle_insn() 702 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power9_idle_stop() 941 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power10_idle_stop()
|
D | opal-call.c | 100 bool mmu = (msr & (MSR_IR|MSR_DR)); in opal_call()
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/pasemi/ |
D | powersave.S | 62 LOAD_REG_IMMEDIATE(r6,MSR_DR|MSR_IR|MSR_ME|MSR_EE)
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/82xx/ |
D | pq2.c | 29 mtmsr(mfmsr() & ~(MSR_ME | MSR_EE | MSR_IR | MSR_DR)); in pq2_restart()
|
/kernel/linux/linux-5.10/arch/powerpc/xmon/ |
D | xmon.c | 555 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) in xmon_core() 709 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { in xmon_core() 762 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) in xmon_bpt() 795 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) in xmon_break_match() 810 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) in xmon_iabr_match() 835 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { in xmon_fault_handler() 1206 if ((regs->msr & (MSR_64BIT|MSR_PR|MSR_IR)) == (MSR_64BIT|MSR_IR)) { in do_step()
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/pseries/ |
D | ras.c | 493 (MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR| in pSeries_system_reset_exception() 747 mtmsr(mfmsr() | MSR_IR | MSR_DR); in mce_handle_error()
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
D | reg.h | 108 #define MSR_IR __MASK(MSR_IR_LG) /* Instruction Relocate */ macro 137 #define __MSR (MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_ISF |MSR_HV) 150 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR)
|
D | reg_booke.h | 46 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR|MSR_CE)
|
/kernel/linux/linux-5.10/arch/powerpc/platforms/52xx/ |
D | lite5200_sleep.S | 207 ori r10, r10, MSR_DR | MSR_IR
|