Searched refs:MSR_LE (Results 1 – 15 of 15) sorted by relevance
/arch/powerpc/kernel/ |
D | signal_64.c | 324 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_sigcontext() 434 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_tm_sigcontexts() 788 regs->msr &= ~MSR_LE; in handle_rt_signal64() 789 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal64()
|
D | align.c | 816 if (cpu_has_feature(CPU_FTR_PPC_LE) && (regs->msr & MSR_LE)) in fix_alignment() 821 if (cpu_has_feature(CPU_FTR_REAL_LE) && (regs->msr & MSR_LE)) in fix_alignment() 868 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) { in fix_alignment() 908 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) in fix_alignment()
|
D | signal_32.c | 688 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_user_regs() 799 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_tm_user_regs() 1057 regs->msr &= ~MSR_LE; in handle_rt_signal32() 1058 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal32() 1473 regs->msr &= ~MSR_LE; in handle_signal32()
|
D | idle_power7.S | 170 li r0,MSR_LE
|
D | process.c | 972 {MSR_LE, "LE"}, 1427 regs->msr &= ~MSR_LE; in set_endian() 1429 regs->msr |= MSR_LE; in set_endian() 1448 if (regs->msr & MSR_LE) { in get_endian()
|
D | entry_64.S | 1030 ori r9,r9,MSR_IR|MSR_DR|MSR_FE0|MSR_FE1|MSR_FP|MSR_RI|MSR_LE 1134 LOAD_REG_IMMEDIATE(r12, MSR_SF | MSR_ISF | MSR_LE)
|
D | exceptions-64s.S | 54 xori r12,r12,MSR_LE ; \
|
/arch/powerpc/include/asm/ |
D | kvm_book3s.h | 277 return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE); in kvmppc_need_byteswap()
|
D | reg.h | 103 #define MSR_LE __MASK(MSR_LE_LG) /* Little Endian */ macro 123 #define MSR_ (__MSR | MSR_LE) 124 #define MSR_IDLE (MSR_ME | MSR_SF | MSR_HV | MSR_LE)
|
/arch/powerpc/platforms/powernv/ |
D | opal-wrappers.S | 76 li r11,MSR_DR|MSR_IR|MSR_LE;\ 139 li r11,MSR_DR|MSR_IR|MSR_LE
|
/arch/powerpc/kvm/ |
D | book3s_pr.c | 325 smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE; in kvmppc_recalc_shadow_msr() 1334 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_get_one_reg_pr() 1350 vcpu->arch.intr_msr |= MSR_LE; in kvmppc_set_lpcr_pr() 1352 vcpu->arch.intr_msr &= ~MSR_LE; in kvmppc_set_lpcr_pr() 1434 vcpu->arch.shadow_msr = MSR_USER64 & ~MSR_LE; in kvmppc_core_vcpu_create_pr()
|
D | book3s_emulate.c | 102 if ((kvmppc_get_msr(vcpu) & MSR_LE) && in kvmppc_core_emulate_op_pr()
|
D | powerpc.c | 173 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_kvm_pv()
|
D | book3s_hv.c | 927 vcpu->arch.intr_msr |= MSR_LE; in kvmppc_set_lpcr() 929 vcpu->arch.intr_msr &= ~MSR_LE; in kvmppc_set_lpcr()
|
/arch/powerpc/lib/ |
D | sstep.c | 956 op->val = 0xffffffff & ~(MSR_ME | MSR_LE); in analyse_instr() 1895 if (regs->msr & MSR_LE) in emulate_step() 1941 if (regs->msr & MSR_LE) in emulate_step() 1981 regs->msr ^= MSR_LE; in emulate_step()
|