/arch/powerpc/kernel/ |
D | head_40x.S | 72 lis r0,MSR_KERNEL@h 73 ori r0,r0,MSR_KERNEL@l 211 EXC_XFER_TEMPLATE(hdlr, n+2, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 229 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, NOCOPY, transfer_to_handler_full, \ 233 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, NOCOPY, transfer_to_handler, \ 237 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, COPY_EE, transfer_to_handler_full, \ 241 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, COPY_EE, transfer_to_handler, \ 726 (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 748 (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), 862 lis r3,(MSR_KERNEL & ~(MSR_IR|MSR_DR))@h [all …]
|
D | head_booke.h | 212 EXC_XFER_TEMPLATE(hdlr, n+2, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 222 EXC_XFER_TEMPLATE(hdlr, n+4, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 240 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, NOCOPY, transfer_to_handler_full, \ 244 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, NOCOPY, transfer_to_handler, \ 248 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, COPY_EE, transfer_to_handler_full, \ 252 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, COPY_EE, transfer_to_handler, \ 319 …EXC_XFER_TEMPLATE(DebugException, 0x2008, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), NOCOPY, debug_tr… 372 …EXC_XFER_TEMPLATE(DebugException, 0x2002, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), NOCOPY, crit_tra…
|
D | entry_32.S | 42 #if MSR_KERNEL >= 0x10000 289 LOAD_MSR_KERNEL(r10,MSR_KERNEL) 361 LOAD_MSR_KERNEL(r10,MSR_KERNEL) /* doesn't include MSR_EE */ 803 lis r10,MSR_KERNEL@h 804 ori r10,r10,MSR_KERNEL@l 821 LOAD_MSR_KERNEL(r10,MSR_KERNEL) 999 LOAD_MSR_KERNEL(r10,MSR_KERNEL & ~MSR_RI) 1072 LOAD_MSR_KERNEL(r10,MSR_KERNEL); \ 1245 LOAD_MSR_KERNEL(r10,MSR_KERNEL) 1336 LOAD_MSR_KERNEL(r0,MSR_KERNEL) [all …]
|
D | fsl_booke_entry_mapping.S | 215 lis r7,MSR_KERNEL@h 216 ori r7,r7,MSR_KERNEL@l
|
D | paca.c | 159 new_paca->kernel_msr = MSR_KERNEL & ~(MSR_IR | MSR_DR); in initialise_paca()
|
D | head_8xx.S | 144 li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR); /* can take exceptions */ \ 179 li r10,MSR_KERNEL; \ 709 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 730 li r4,MSR_KERNEL
|
D | head_32.S | 284 li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR); /* can take exceptions */ \ 315 li r10,MSR_KERNEL; \ 871 li r4,MSR_KERNEL 980 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 1004 li r4,MSR_KERNEL
|
D | head_fsl_booke.S | 293 lis r3,MSR_KERNEL@h 294 ori r3,r3,MSR_KERNEL@l 1206 lis r4,MSR_KERNEL@h 1207 ori r4,r4,MSR_KERNEL@l
|
D | align.c | 868 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) { in fix_alignment() 908 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) in fix_alignment()
|
D | setup_64.c | 202 get_paca()->kernel_msr = MSR_KERNEL; in cpu_ready_for_interrupts()
|
D | head_44x.S | 226 lis r3,MSR_KERNEL@h 227 ori r3,r3,MSR_KERNEL@l
|
D | head_64.S | 665 LOAD_REG_IMMEDIATE(r4, MSR_KERNEL)
|
D | signal_64.c | 789 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal64()
|
D | exceptions-64e.S | 1352 lis r7,MSR_KERNEL@h 1353 ori r7,r7,MSR_KERNEL@l
|
D | signal_32.c | 1058 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal32()
|
/arch/microblaze/include/asm/ |
D | registers.h | 40 # define MSR_KERNEL (MSR_EE | MSR_VM) macro
|
/arch/powerpc/include/asm/ |
D | reg_booke.h | 45 #define MSR_KERNEL (MSR_ | MSR_64BIT) macro 49 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR|MSR_CE) macro 50 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE) 52 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_CE) macro 53 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
|
D | reg.h | 126 #define MSR_KERNEL (MSR_ | MSR_64BIT) macro 131 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR) macro 132 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
|
D | kvm_book3s.h | 277 return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE); in kvmppc_need_byteswap()
|
/arch/powerpc/platforms/85xx/ |
D | smp.c | 179 imsr1 = MSR_KERNEL; in wake_hw_thread()
|
/arch/microblaze/kernel/ |
D | head.S | 370 ori r4,r0,MSR_KERNEL
|
/arch/powerpc/lib/ |
D | sstep.c | 1985 regs->gpr[10] = MSR_KERNEL; in emulate_step() 1990 regs->msr = MSR_KERNEL; in emulate_step()
|