Searched refs:is_smm (Results 1 – 10 of 10) sorted by relevance
149 static inline bool is_smm(struct kvm_vcpu *vcpu) in is_smm() function160 static inline bool is_smm(struct kvm_vcpu *vcpu) { return false; } in is_smm() function
238 return !is_smm(vcpu) && in kvm_apic_init_sipi_allowed()
222 bool (*is_smm)(struct x86_emulate_ctxt *ctxt); member
5222 events->smi.smm = is_smm(vcpu); in kvm_vcpu_ioctl_x86_get_vcpu_events()8304 return is_smm(emul_to_vcpu(ctxt)); in emulator_is_smm()8377 .is_smm = emulator_is_smm,9980 if (is_smm(vcpu)) in post_kvm_run_save()12055 WARN_ON_ONCE(is_guest_mode(vcpu) || is_smm(vcpu)); in kvm_vcpu_reset()
2319 if (!ctxt->ops->is_smm(ctxt)) in em_rsm()
849 if (is_smm(vcpu)) { in nested_svm_vmrun()1679 if (is_smm(vcpu) && (kvm_state->flags & KVM_STATE_NESTED_GUEST_MODE)) in svm_set_nested_state()
324 if (!is_smm(vcpu)) in svm_set_efer()4603 return is_smm(vcpu); in svm_smi_blocked()
5116 role.base.smm = is_smm(vcpu); in kvm_calc_cpu_role()5309 WARN_ON_ONCE(is_smm(vcpu)); in kvm_calc_shadow_ept_root_page_role()
4189 if (vcpu->arch.smi_pending && !is_smm(vcpu)) { in vmx_check_nested_events()6626 if (is_smm(vcpu) ? in vmx_set_nested_state()
3429 if ((cr4 & X86_CR4_VMXE) && is_smm(vcpu)) in vmx_is_valid_cr4()8157 return !is_smm(vcpu); in vmx_smi_allowed()