Searched refs:MSR_LE (Results 1 – 17 of 17) sorted by relevance
/arch/powerpc/boot/ |
D | ppc_asm.h | 67 #define MSR_LE 0x0000000000000001 macro
|
D | opal-calls.S | 47 li r11,MSR_LE
|
/arch/powerpc/kernel/ |
D | signal_64.c | 346 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_sigcontext() 470 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_tm_sigcontexts() 836 regs->msr &= ~MSR_LE; in handle_rt_signal64() 837 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal64()
|
D | align.c | 815 if (cpu_has_feature(CPU_FTR_PPC_LE) && (regs->msr & MSR_LE)) in fix_alignment() 820 if (cpu_has_feature(CPU_FTR_REAL_LE) && (regs->msr & MSR_LE)) in fix_alignment() 867 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) { in fix_alignment() 907 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) in fix_alignment()
|
D | syscalls.c | 130 current->thread.regs->msr ^= MSR_LE; in sys_switch_endian()
|
D | signal_32.c | 684 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_user_regs() 789 regs->msr = (regs->msr & ~MSR_LE) | (msr & MSR_LE); in restore_tm_user_regs() 1054 regs->msr &= ~MSR_LE; in handle_rt_signal32() 1055 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal32() 1488 regs->msr &= ~MSR_LE; in handle_signal32()
|
D | process.c | 1294 {MSR_LE, "LE"}, 1784 regs->msr &= ~MSR_LE; in set_endian() 1786 regs->msr |= MSR_LE; in set_endian() 1805 if (regs->msr & MSR_LE) { in get_endian()
|
D | entry_64.S | 1069 ori r9,r9,MSR_IR|MSR_DR|MSR_FE0|MSR_FE1|MSR_FP|MSR_RI|MSR_LE 1173 LOAD_REG_IMMEDIATE(r12, MSR_SF | MSR_ISF | MSR_LE)
|
D | exceptions-64s.S | 902 xori r12,r12,MSR_LE ; \
|
/arch/powerpc/platforms/powernv/ |
D | opal-wrappers.S | 75 li r11,MSR_DR|MSR_IR|MSR_LE;\ 137 li r11,MSR_DR|MSR_IR|MSR_LE 170 li r11,MSR_LE; \
|
/arch/powerpc/include/asm/ |
D | kvm_book3s.h | 311 return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE); in kvmppc_need_byteswap()
|
D | reg.h | 111 #define MSR_LE __MASK(MSR_LE_LG) /* Little Endian */ macro 132 #define MSR_ (__MSR | MSR_LE) 133 #define MSR_IDLE (MSR_ME | MSR_SF | MSR_HV | MSR_LE)
|
/arch/powerpc/kvm/ |
D | book3s_pr.c | 325 smsr &= MSR_FE0 | MSR_FE1 | MSR_SF | MSR_SE | MSR_BE | MSR_LE; in kvmppc_recalc_shadow_msr() 1378 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_get_one_reg_pr() 1394 vcpu->arch.intr_msr |= MSR_LE; in kvmppc_set_lpcr_pr() 1396 vcpu->arch.intr_msr &= ~MSR_LE; in kvmppc_set_lpcr_pr() 1481 vcpu->arch.shadow_msr = MSR_USER64 & ~MSR_LE; in kvmppc_core_vcpu_create_pr()
|
D | book3s_emulate.c | 103 if ((kvmppc_get_msr(vcpu) & MSR_LE) && in kvmppc_core_emulate_op_pr()
|
D | powerpc.c | 179 if (vcpu->arch.intr_msr & MSR_LE) in kvmppc_kvm_pv()
|
D | book3s_hv.c | 1115 vcpu->arch.intr_msr |= MSR_LE; in kvmppc_set_lpcr() 1117 vcpu->arch.intr_msr &= ~MSR_LE; in kvmppc_set_lpcr()
|
/arch/powerpc/lib/ |
D | sstep.c | 957 op->val = 0xffffffff & ~(MSR_ME | MSR_LE); in analyse_instr() 1901 if (regs->msr & MSR_LE) in emulate_step() 1948 if (regs->msr & MSR_LE) in emulate_step() 1988 regs->msr ^= MSR_LE; in emulate_step()
|