/arch/powerpc/kernel/ |
D | head_40x.S | 72 lis r0,MSR_KERNEL@h 73 ori r0,r0,MSR_KERNEL@l 211 EXC_XFER_TEMPLATE(hdlr, n+2, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 229 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, NOCOPY, transfer_to_handler_full, \ 233 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, NOCOPY, transfer_to_handler, \ 237 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, COPY_EE, transfer_to_handler_full, \ 241 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, COPY_EE, transfer_to_handler, \ 726 (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 748 (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), 862 lis r3,(MSR_KERNEL & ~(MSR_IR|MSR_DR))@h [all …]
|
D | head_booke.h | 224 EXC_XFER_TEMPLATE(hdlr, n+2, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 234 EXC_XFER_TEMPLATE(hdlr, n+4, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), \ 252 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, NOCOPY, transfer_to_handler_full, \ 256 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, NOCOPY, transfer_to_handler, \ 260 EXC_XFER_TEMPLATE(hdlr, n, MSR_KERNEL, COPY_EE, transfer_to_handler_full, \ 264 EXC_XFER_TEMPLATE(hdlr, n+1, MSR_KERNEL, COPY_EE, transfer_to_handler, \ 331 …EXC_XFER_TEMPLATE(DebugException, 0x2008, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), NOCOPY, debug_tr… 384 …EXC_XFER_TEMPLATE(DebugException, 0x2002, (MSR_KERNEL & ~(MSR_ME|MSR_DE|MSR_CE)), NOCOPY, crit_tra…
|
D | entry_32.S | 41 #if MSR_KERNEL >= 0x10000 288 LOAD_MSR_KERNEL(r10,MSR_KERNEL) 363 LOAD_MSR_KERNEL(r10,MSR_KERNEL) /* doesn't include MSR_EE */ 744 lis r10,MSR_KERNEL@h 745 ori r10,r10,MSR_KERNEL@l 762 LOAD_MSR_KERNEL(r10,MSR_KERNEL) 943 LOAD_MSR_KERNEL(r10,MSR_KERNEL & ~MSR_RI) 1019 LOAD_MSR_KERNEL(r10,MSR_KERNEL); \ 1192 LOAD_MSR_KERNEL(r10,MSR_KERNEL) 1283 LOAD_MSR_KERNEL(r0,MSR_KERNEL) [all …]
|
D | fsl_booke_entry_mapping.S | 218 lis r7,MSR_KERNEL@h 219 ori r7,r7,MSR_KERNEL@l
|
D | paca.c | 167 new_paca->kernel_msr = MSR_KERNEL & ~(MSR_IR | MSR_DR); in initialise_paca()
|
D | head_8xx.S | 168 li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR); /* can take exceptions */ \ 202 li r10,MSR_KERNEL; \ 703 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 724 li r4,MSR_KERNEL
|
D | head_32.S | 284 li r10,MSR_KERNEL & ~(MSR_IR|MSR_DR); /* can take exceptions */ \ 315 li r10,MSR_KERNEL; \ 871 li r4,MSR_KERNEL 980 li r3,MSR_KERNEL & ~(MSR_IR|MSR_DR) 1004 li r4,MSR_KERNEL
|
D | head_fsl_booke.S | 293 lis r3,MSR_KERNEL@h 294 ori r3,r3,MSR_KERNEL@l 1221 lis r4,MSR_KERNEL@h 1222 ori r4,r4,MSR_KERNEL@l
|
D | align.c | 868 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) { in fix_alignment() 908 if ((regs->msr & MSR_LE) != (MSR_KERNEL & MSR_LE)) in fix_alignment()
|
D | head_44x.S | 226 lis r3,MSR_KERNEL@h 227 ori r3,r3,MSR_KERNEL@l
|
D | setup_64.c | 212 get_paca()->kernel_msr = MSR_KERNEL; in cpu_ready_for_interrupts()
|
D | head_64.S | 700 LOAD_REG_IMMEDIATE(r4, MSR_KERNEL)
|
D | signal_64.c | 811 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal64()
|
D | exceptions-64e.S | 1395 lis r7,MSR_KERNEL@h 1396 ori r7,r7,MSR_KERNEL@l
|
D | signal_32.c | 1067 regs->msr |= (MSR_KERNEL & MSR_LE); in handle_rt_signal32()
|
/arch/microblaze/include/asm/ |
D | registers.h | 40 # define MSR_KERNEL (MSR_EE | MSR_VM) macro
|
/arch/powerpc/include/asm/ |
D | reg_booke.h | 45 #define MSR_KERNEL (MSR_ | MSR_64BIT) macro 49 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR|MSR_CE) macro 50 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE) 52 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_CE) macro 53 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
|
D | reg.h | 127 #define MSR_KERNEL (MSR_ | MSR_64BIT) macro 132 #define MSR_KERNEL (MSR_ME|MSR_RI|MSR_IR|MSR_DR) macro 133 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
|
D | kvm_book3s.h | 272 return (kvmppc_get_msr(vcpu) & MSR_LE) != (MSR_KERNEL & MSR_LE); in kvmppc_need_byteswap()
|
/arch/powerpc/platforms/85xx/ |
D | smp.c | 179 imsr = MSR_KERNEL; in wake_hw_thread()
|
/arch/microblaze/kernel/ |
D | head.S | 370 ori r4,r0,MSR_KERNEL
|
/arch/powerpc/lib/ |
D | sstep.c | 1987 regs->gpr[10] = MSR_KERNEL; in emulate_step() 1992 regs->msr = MSR_KERNEL; in emulate_step()
|