Lines Matching refs:get_cpl
716 if (kvm_x86_ops.get_cpl(vcpu) <= required_cpl) in kvm_require_cpl()
981 if (kvm_x86_ops.get_cpl(vcpu) != 0 || in kvm_set_xcr()
4093 vcpu->arch.preempted_in_kernel = !kvm_x86_ops.get_cpl(vcpu); in kvm_arch_vcpu_put()
6055 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_read()
6062 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_fetch()
6070 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_write()
6119 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_fetch_guest_virt()
6144 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_read_guest_virt()
6165 if (!system && kvm_x86_ops.get_cpl(vcpu) == 3) in emulator_read_std()
6218 if (!system && kvm_x86_ops.get_cpl(vcpu) == 3) in emulator_write_std()
6277 u32 access = ((kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0) in vcpu_mmio_gva_to_gpa()
6803 return kvm_x86_ops.get_cpl(emul_to_vcpu(ctxt)); in emulator_get_cpl()
7209 if (!is_guest_mode(vcpu) && kvm_x86_ops.get_cpl(vcpu) == 0) { in handle_emulation_failure()
7992 user_mode = kvm_x86_ops.get_cpl(__this_cpu_read(current_vcpu)); in kvm_is_user_mode()
8334 if (kvm_x86_ops.get_cpl(vcpu) != 0) { in kvm_emulate_hypercall()
11210 (vcpu->arch.apf.send_user_only && kvm_x86_ops.get_cpl(vcpu) == 0)) in kvm_can_deliver_async_pf()