Searched refs:MSR_TM_ACTIVE (Results 1 – 13 of 13) sorted by relevance
97 MSR_TM_ACTIVE(tsk->thread.regs->msr) && in check_if_tm_restore_required()232 if (!MSR_TM_ACTIVE(cpumsr) && in enable_kernel_fp()233 MSR_TM_ACTIVE(current->thread.regs->msr)) in enable_kernel_fp()297 if (!MSR_TM_ACTIVE(cpumsr) && in enable_kernel_altivec()298 MSR_TM_ACTIVE(current->thread.regs->msr)) in enable_kernel_altivec()383 if (!MSR_TM_ACTIVE(cpumsr) && in enable_kernel_vsx()384 MSR_TM_ACTIVE(current->thread.regs->msr)) in enable_kernel_vsx()528 if (!MSR_TM_ACTIVE(regs->msr) && in restore_math()896 if (!MSR_TM_ACTIVE(thr->regs->msr)) in tm_reclaim_task()963 if (!MSR_TM_ACTIVE(new->thread.regs->msr)){ in tm_recheckpoint_new_task()[all …]
807 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cgpr_active()843 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cgpr_get()908 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cgpr_set()972 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cfpr_active()1010 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cfpr_get()1056 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cfpr_set()1091 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cvmx_active()1131 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cvmx_get()1193 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cvmx_set()1236 if (!MSR_TM_ACTIVE(target->thread.regs->msr)) in tm_cvsx_active()[all …]
208 if (MSR_TM_ACTIVE(tsk->thread.regs->msr)) { in get_tm_stackpointer()
215 BUG_ON(!MSR_TM_ACTIVE(msr)); in setup_tm_sigcontexts()766 if (MSR_TM_ACTIVE(msr)) { in SYSCALL_DEFINE0()843 if (MSR_TM_ACTIVE(msr)) { in handle_rt_signal64()
943 if (MSR_TM_ACTIVE(msr)) { in handle_rt_signal32()1196 if (MSR_TM_ACTIVE(msr_hi<<32)) { in COMPAT_SYSCALL_DEFINE0()1405 if (MSR_TM_ACTIVE(msr)) { in handle_signal32()1492 if (MSR_TM_ACTIVE(msr_hi<<32)) { in COMPAT_SYSCALL_DEFINE0()
174 if (!MSR_TM_ACTIVE(msr)) { in kvmhv_p9_tm_emulation()212 if (MSR_TM_ACTIVE(msr) || !(vcpu->arch.texasr & TEXASR_FS)) { in kvmhv_p9_tm_emulation()
222 MSR_TM_ACTIVE(guest_msr)) { in kvmppc_emulate_tabort()282 !MSR_TM_ACTIVE(srr1)) in kvmppc_core_emulate_op_pr()573 if (!MSR_TM_ACTIVE(guest_msr)) { in kvmppc_core_emulate_op_pr()611 if (MSR_TM_ACTIVE(guest_msr) || in kvmppc_core_emulate_op_pr()795 if (MSR_TM_ACTIVE(kvmppc_get_msr(vcpu)) && in kvmppc_core_emulate_mtspr_pr()
793 if (MSR_TM_ACTIVE(old_msr)) in kvmppc_guest_entry_inject_int()
342 if (!(MSR_TM_ACTIVE(kvmppc_get_msr(vcpu)))) { in kvmppc_save_tm_pr()357 if (!MSR_TM_ACTIVE(kvmppc_get_msr(vcpu))) { in kvmppc_restore_tm_pr()
267 if (!MSR_TM_ACTIVE(l2_regs.msr)) in kvmhv_enter_nested_guest()
4264 if (MSR_TM_ACTIVE(current->thread.regs->msr)) { in kvmppc_vcpu_run_hv()
129 #define MSR_TM_ACTIVE(x) (((x) & MSR_TS_MASK) != 0) /* Transaction active? */ macro131 #define MSR_TM_ACTIVE(x) 0 macro
1689 MSR_TM_ACTIVE(current->thread.regs->msr)) { in tm_flush_hash_page()