Home
last modified time | relevance | path

Searched refs:MSR_FP (Results 1 – 23 of 23) sorted by relevance

/arch/powerpc/kernel/
Dprocess.c129 if (cpu_has_feature(CPU_FTR_VSX) && (bits & MSR_FP)) in msr_check_and_set()
147 if (cpu_has_feature(CPU_FTR_VSX) && (bits & MSR_FP)) in __msr_check_and_clear()
162 msr &= ~(MSR_FP|MSR_FE0|MSR_FE1); in __giveup_fpu()
172 msr_check_and_set(MSR_FP); in giveup_fpu()
174 msr_check_and_clear(MSR_FP); in giveup_fpu()
194 if (tsk->thread.regs->msr & MSR_FP) { in flush_fp_to_thread()
216 cpumsr = msr_check_and_set(MSR_FP); in enable_kernel_fp()
218 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) { in enable_kernel_fp()
313 WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC))); in __giveup_vsx()
316 if (msr & MSR_FP) in __giveup_vsx()
[all …]
Dfpu.S96 ori r5,r5,MSR_FP
108 ori r9,r9,MSR_FP /* enable FP for current */
114 ori r12,r12,MSR_FP
Dsignal_64.c232 msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX); in setup_tm_sigcontexts()
280 if (msr & MSR_FP) in setup_tm_sigcontexts()
379 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX)); in __unsafe_restore_sigcontext()
498 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX)); in restore_tm_sigcontexts()
591 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_sigcontexts()
592 if (msr & MSR_FP) { in restore_tm_sigcontexts()
594 regs_set_return_msr(regs, regs->msr | (MSR_FP | tsk->thread.fpexc_mode)); in restore_tm_sigcontexts()
Dsignal_32.c412 if (msr & MSR_FP) in save_tm_user_regs_unsafe()
530 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1)); in restore_user_regs()
607 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1)); in restore_tm_user_regs()
690 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_user_regs()
691 if (msr & MSR_FP) { in restore_tm_user_regs()
693 regs_set_return_msr(regs, regs->msr | (MSR_FP | current->thread.fpexc_mode)); in restore_tm_user_regs()
Dvector.S126 andi. r5,r12,MSR_FP
188 ori r11,r10,MSR_FP
Dtm.S135 ori r15, r15, MSR_FP
387 ori r5, r5, MSR_FP
Dinterrupt.c370 unsigned long mathflags = MSR_FP; in interrupt_exit_user_prepare_main()
Dcpu_setup_6xx.S281 ori r11,r10,MSR_FP
Dexceptions-64s.S2560 ori r10,r10,(MSR_FP|MSR_FE0|MSR_FE1)
/arch/powerpc/include/asm/
Dswitch_to.h46 msr_check_and_clear(MSR_FP); in disable_kernel_fp()
72 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in disable_kernel_vsx()
Dreg.h97 #define MSR_FP __MASK(MSR_FP_LG) /* Floating Point enable */ macro
/arch/powerpc/lib/
Dldstfp.S23 ori r7, r6, MSR_FP
47 ori r7, r6, MSR_FP
213 ori r7, r6, MSR_FP
228 ori r7, r6, MSR_FP
Dsstep.c554 if (regs->msr & MSR_FP) in do_fp_load()
561 if (regs->msr & MSR_FP) in do_fp_load()
590 if (regs->msr & MSR_FP) in do_fp_store()
602 if (regs->msr & MSR_FP) in do_fp_store()
922 if (regs->msr & MSR_FP) { in do_vsx_load()
968 if (regs->msr & MSR_FP) { in do_vsx_store()
3375 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_FP)) in emulate_loadstore()
3446 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_FP)) in emulate_loadstore()
Dtest_emulate_step.c73 regs->msr |= MSR_FP; in init_pt_regs()
/arch/powerpc/kvm/
Dbook3s_pr.c173 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_core_vcpu_put_pr()
362 (MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_handle_lost_math_exts()
367 if (ext_diff == MSR_FP) in kvmppc_handle_lost_math_exts()
542 if (kvmppc_get_msr(vcpu) & MSR_FP) in kvmppc_set_msr_pr()
543 kvmppc_handle_ext(vcpu, BOOK3S_INTERRUPT_FP_UNAVAIL, MSR_FP); in kvmppc_set_msr_pr()
808 msr |= MSR_FP | MSR_VEC; in kvmppc_giveup_ext()
818 if (msr & MSR_FP) { in kvmppc_giveup_ext()
824 if (t->regs->msr & MSR_FP) in kvmppc_giveup_ext()
889 msr = MSR_FP | MSR_VEC | MSR_VSX; in kvmppc_handle_ext()
901 if (msr & MSR_FP) { in kvmppc_handle_ext()
[all …]
Demulate_loadstore.c30 if (!(kvmppc_get_msr(vcpu) & MSR_FP)) { in kvmppc_check_fp_disabled()
246 MSR_FP); in kvmppc_emulate_loadstore()
Dtm.S44 ori r8, r8, MSR_FP
238 ori r5, r5, MSR_FP
Dbooke.c160 if (!(current->thread.regs->msr & MSR_FP)) { in kvmppc_load_guest_fp()
165 current->thread.regs->msr |= MSR_FP; in kvmppc_load_guest_fp()
177 if (current->thread.regs->msr & MSR_FP) in kvmppc_save_guest_fp()
188 vcpu->arch.shadow_msr &= ~MSR_FP; in kvmppc_vcpu_sync_fpu()
189 vcpu->arch.shadow_msr |= vcpu->arch.shared->msr & MSR_FP; in kvmppc_vcpu_sync_fpu()
Dbook3s_paired_singles.c656 if (!(kvmppc_get_msr(vcpu) & MSR_FP)) { in kvmppc_emulate_paired_single()
661 kvmppc_giveup_ext(vcpu, MSR_FP); in kvmppc_emulate_paired_single()
Dbook3s_emulate.c737 kvmppc_giveup_ext(vcpu, MSR_FP); in kvmppc_core_emulate_mtspr_pr()
Dbook3s_hv_rmhandlers.S2461 ori r8,r5,MSR_FP
2496 ori r8,r9,MSR_FP
Dpowerpc.c1172 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_FP); in kvmppc_complete_mmio_load()
Dbook3s_hv.c3918 msr_check_and_set(MSR_FP | MSR_VEC | MSR_VSX); in kvmhv_p9_guest_entry()
4027 msr_check_and_set(MSR_FP | MSR_VEC | MSR_VSX); in kvmhv_p9_guest_entry()