Home
last modified time | relevance | path

Searched refs:MSR_FP (Results 1 – 20 of 20) sorted by relevance

/arch/powerpc/kernel/
Dprocess.c122 if (cpu_has_feature(CPU_FTR_VSX) && (bits & MSR_FP)) in msr_check_and_set()
140 if (cpu_has_feature(CPU_FTR_VSX) && (bits & MSR_FP)) in __msr_check_and_clear()
156 msr &= ~MSR_FP; in __giveup_fpu()
168 msr_check_and_set(MSR_FP); in giveup_fpu()
170 msr_check_and_clear(MSR_FP); in giveup_fpu()
190 if (tsk->thread.regs->msr & MSR_FP) { in flush_fp_to_thread()
212 cpumsr = msr_check_and_set(MSR_FP); in enable_kernel_fp()
214 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) { in enable_kernel_fp()
330 if (tsk->thread.regs->msr & MSR_FP) in __giveup_vsx()
341 msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx()
[all …]
Dfpu.S87 ori r5,r5,MSR_FP
100 ori r9,r9,MSR_FP /* enable FP for current */
106 ori r12,r12,MSR_FP
Dtm.S117 ori r15, r15, MSR_FP
269 andi. r0, r4, MSR_FP
351 ori r5, r5, MSR_FP
383 andi. r0, r4, MSR_FP
Dsignal_64.c270 if (msr & MSR_FP) in setup_tm_sigcontexts()
368 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); in restore_sigcontext()
497 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); in restore_tm_sigcontexts()
563 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_sigcontexts()
564 if (msr & MSR_FP) { in restore_tm_sigcontexts()
566 regs->msr |= (MSR_FP | tsk->thread.fpexc_mode); in restore_tm_sigcontexts()
Dtraps.c1564 regs->msr |= (MSR_FP | current->thread.fpexc_mode); in fp_unavailable_tm()
1572 tm_recheckpoint(&current->thread, MSR_FP); in fp_unavailable_tm()
1597 if (regs->msr & MSR_FP) { in altivec_unavailable_tm()
1598 msr_check_and_set(MSR_FP); in altivec_unavailable_tm()
1622 if ((orig_msr & (MSR_FP | MSR_VEC)) == (MSR_FP | MSR_VEC)) { in vsx_unavailable_tm()
1630 regs->msr |= MSR_VEC | MSR_FP | current->thread.fpexc_mode | in vsx_unavailable_tm()
1638 msr_check_and_set(orig_msr & (MSR_FP | MSR_VEC)); in vsx_unavailable_tm()
1640 if (orig_msr & MSR_FP) in vsx_unavailable_tm()
Dvector.S119 andi. r5,r12,MSR_FP
179 ori r11,r10,MSR_FP
Dsignal_32.c590 if (msr & MSR_FP) { in save_tm_user_regs()
733 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1); in restore_user_regs()
819 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1); in restore_tm_user_regs()
882 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_user_regs()
883 if (msr & MSR_FP) { in restore_tm_user_regs()
885 regs->msr |= (MSR_FP | current->thread.fpexc_mode); in restore_tm_user_regs()
Dcpu_setup_6xx.S284 ori r11,r10,MSR_FP
Dentry_64.S220 andi. r0,r8,MSR_FP
1069 ori r9,r9,MSR_IR|MSR_DR|MSR_FE0|MSR_FE1|MSR_FP|MSR_RI|MSR_LE
Dentry_32.S624 li r0,MSR_FP /* Disable floating-point */
Dexceptions-64s.S1227 ori r10,r10,(MSR_FP|MSR_FE0|MSR_FE1)
/arch/powerpc/include/asm/
Dswitch_to.h34 msr_check_and_clear(MSR_FP); in disable_kernel_fp()
61 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in disable_kernel_vsx()
Dreg.h95 #define MSR_FP __MASK(MSR_FP_LG) /* Floating Point enable */ macro
/arch/powerpc/lib/
Dldstfp.S84 ori r7,r6,MSR_FP
111 ori r7,r6,MSR_FP
138 ori r7,r6,MSR_FP
165 ori r7,r6,MSR_FP
Dsstep.c1473 if (!(regs->msr & MSR_FP)) in analyse_instr()
1480 if (!(regs->msr & MSR_FP)) in analyse_instr()
1487 if (!(regs->msr & MSR_FP)) in analyse_instr()
1494 if (!(regs->msr & MSR_FP)) in analyse_instr()
1613 if (!(regs->msr & MSR_FP)) in analyse_instr()
1621 if (!(regs->msr & MSR_FP)) in analyse_instr()
1629 if (!(regs->msr & MSR_FP)) in analyse_instr()
1637 if (!(regs->msr & MSR_FP)) in analyse_instr()
/arch/powerpc/kvm/
Dbook3s_pr.c133 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_core_vcpu_put_pr()
398 if (kvmppc_get_msr(vcpu) & MSR_FP) in kvmppc_set_msr_pr()
399 kvmppc_handle_ext(vcpu, BOOK3S_INTERRUPT_FP_UNAVAIL, MSR_FP); in kvmppc_set_msr_pr()
663 msr |= MSR_FP | MSR_VEC; in kvmppc_giveup_ext()
673 if (msr & MSR_FP) { in kvmppc_giveup_ext()
679 if (t->regs->msr & MSR_FP) in kvmppc_giveup_ext()
744 msr = MSR_FP | MSR_VEC | MSR_VSX; in kvmppc_handle_ext()
756 if (msr & MSR_FP) { in kvmppc_handle_ext()
795 if (lost_ext & MSR_FP) { in kvmppc_handle_lost_ext()
1202 ext_msr = MSR_FP; in kvmppc_handle_exit_pr()
[all …]
Dbooke.c144 if (!(current->thread.regs->msr & MSR_FP)) { in kvmppc_load_guest_fp()
149 current->thread.regs->msr |= MSR_FP; in kvmppc_load_guest_fp()
161 if (current->thread.regs->msr & MSR_FP) in kvmppc_save_guest_fp()
172 vcpu->arch.shadow_msr &= ~MSR_FP; in kvmppc_vcpu_sync_fpu()
173 vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_FP; in kvmppc_vcpu_sync_fpu()
Dbook3s_emulate.c425 kvmppc_giveup_ext(vcpu, MSR_FP); in kvmppc_core_emulate_mtspr_pr()
Dbook3s_paired_singles.c667 if (!(kvmppc_get_msr(vcpu) & MSR_FP)) { in kvmppc_emulate_paired_single()
672 kvmppc_giveup_ext(vcpu, MSR_FP); in kvmppc_emulate_paired_single()
Dbook3s_hv_rmhandlers.S2488 ori r8,r5,MSR_FP
2523 ori r8,r9,MSR_FP
2680 ori r5, r5, MSR_FP