Home
last modified time | relevance | path

Searched refs:MSR_DR (Results 1 – 25 of 39) sorted by relevance

12

/arch/powerpc/kernel/
Dmisc_64.S105 ori r0,r7,MSR_DR
106 xori r0,r0,MSR_DR
136 ori r0,r7,MSR_DR
137 xori r0,r0,MSR_DR
167 ori r0,r7,MSR_DR
168 xori r0,r0,MSR_DR
182 ori r0,r7,MSR_DR
183 xori r0,r0,MSR_DR
384 li r10,MSR_DR|MSR_IR
Dbtext.c57 if (!(mfmsr() & MSR_DR)) in rmci_maybe_on()
65 if (!(mfmsr() & MSR_DR)) in rmci_maybe_off()
Dhead_32.h116 li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR) /* can take exceptions */
173 LOAD_REG_IMMEDIATE(r10, MSR_KERNEL & ~(MSR_IR|MSR_DR)) /* can take exceptions */
Dmisc_32.S225 rlwinm r0,r7,0,~MSR_DR
243 rlwinm r0,r7,0,~MSR_DR
Dhead_book3s_32.S205 ori r0,r0,MSR_DR|MSR_IR|MSR_RI
1031 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR)
1167 li r3, MSR_KERNEL & ~(MSR_IR | MSR_DR)
1190 li r3, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI)
1207 andi. r0,r3,MSR_DR|MSR_IR /* MMU enabled? */
Dpaca.c224 new_paca->kernel_msr = MSR_KERNEL & ~(MSR_IR | MSR_DR); in initialise_paca()
Dkvm_emul.S300 andi. r31, r31, MSR_DR | MSR_IR
Dhead_8xx.S106 ori r0,r0,MSR_DR|MSR_IR
599 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR)
727 li r12, MSR_KERNEL & ~(MSR_IR | MSR_DR | MSR_RI)
Dhead_40x.S596 lis r3,(MSR_KERNEL & ~(MSR_IR|MSR_DR))@h
597 ori r3,r3,(MSR_KERNEL & ~(MSR_IR|MSR_DR))@l
/arch/powerpc/kvm/
Dbook3s_32_mmu.c356 if (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_32_esid_to_vsid()
365 switch (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_32_esid_to_vsid()
372 case MSR_DR: in kvmppc_mmu_book3s_32_esid_to_vsid()
375 case MSR_DR|MSR_IR: in kvmppc_mmu_book3s_32_esid_to_vsid()
Dbook3s_pr.c69 return (msr & (MSR_IR|MSR_DR)) == MSR_DR; in kvmppc_is_split_real()
78 if ((msr & (MSR_IR|MSR_DR)) != MSR_DR) in kvmppc_fixup_split_real()
237 smsr |= MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_PR | MSR_EE; in kvmppc_recalc_shadow_msr()
531 if ((kvmppc_get_msr(vcpu) & (MSR_PR|MSR_IR|MSR_DR)) != in kvmppc_set_msr_pr()
532 (old_msr & (MSR_PR|MSR_IR|MSR_DR))) { in kvmppc_set_msr_pr()
540 if (msr & MSR_DR) in kvmppc_set_msr_pr()
712 bool dr = (kvmppc_get_msr(vcpu) & MSR_DR) ? true : false; in kvmppc_handle_pagefault()
734 switch (kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) { in kvmppc_handle_pagefault()
738 case MSR_DR: in kvmppc_handle_pagefault()
747 if ((kvmppc_get_msr(vcpu) & (MSR_DR|MSR_IR)) == MSR_DR) in kvmppc_handle_pagefault()
Dbook3s_hv_builtin.c562 return !(mfmsr() & MSR_DR); in is_rm()
797 (msr & (MSR_IR|MSR_DR)) == (MSR_IR|MSR_DR) ) { in inject_interrupt()
798 new_msr |= MSR_IR | MSR_DR; in inject_interrupt()
Dbook3s_rmhandlers.S151 li r6, MSR_IR | MSR_DR
Dbook3s_64_mmu.c583 if (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_64_esid_to_vsid()
596 switch (msr & (MSR_DR|MSR_IR)) { in kvmppc_mmu_book3s_64_esid_to_vsid()
603 case MSR_DR: in kvmppc_mmu_book3s_64_esid_to_vsid()
606 case MSR_DR|MSR_IR: in kvmppc_mmu_book3s_64_esid_to_vsid()
Dbook3s.c429 int relocated = (kvmppc_get_msr(vcpu) & (data ? MSR_DR : MSR_IR)); in kvmppc_xlate()
443 if ((kvmppc_get_msr(vcpu) & (MSR_IR | MSR_DR)) == MSR_DR && in kvmppc_xlate()
Dbook3s_segment.S296 ori r11, r9, MSR_DR /* Enable paging for data */
/arch/powerpc/platforms/powernv/
Dopal-wrappers.S27 li r0,MSR_IR|MSR_DR|MSR_LE
Dsubcore-asm.S31 li r5, MSR_IR|MSR_DR
Didle.c389 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power7_idle_insn()
702 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power9_idle_stop()
941 WARN_ON_ONCE(mfmsr() & (MSR_IR|MSR_DR)); in power10_idle_stop()
/arch/powerpc/platforms/pasemi/
Dpowersave.S62 LOAD_REG_IMMEDIATE(r6,MSR_DR|MSR_IR|MSR_ME|MSR_EE)
/arch/powerpc/platforms/82xx/
Dpq2.c29 mtmsr(mfmsr() & ~(MSR_ME | MSR_EE | MSR_IR | MSR_DR)); in pq2_restart()
/arch/powerpc/platforms/pseries/
Dras.c493 (MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR| in pSeries_system_reset_exception()
494 MSR_ILE|MSR_HV|MSR_SF)) == (MSR_DR|MSR_SF)) { in pSeries_system_reset_exception()
747 mtmsr(mfmsr() | MSR_IR | MSR_DR); in mce_handle_error()
/arch/powerpc/platforms/powermac/
Dcache.S44 rlwinm r0,r0,0,~MSR_DR
194 rlwinm r0,r0,0,~MSR_DR
/arch/powerpc/include/asm/
Dreg.h109 #define MSR_DR __MASK(MSR_DR_LG) /* Data Relocate */ macro
137 #define __MSR (MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_ISF |MSR_HV)
150 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR)
/arch/powerpc/mm/
Dmem.c443 msr = msr0 & ~MSR_DR; in flush_dcache_icache_phys()

12