Searched refs:X86_CR0_PE (Results 1 – 18 of 18) sorted by relevance
/arch/x86/platform/pvh/ |
D | head.S | 87 mov $(X86_CR0_PG | X86_CR0_PE), %eax 118 or $(X86_CR0_PG | X86_CR0_PE), %eax
|
/arch/x86/realmode/rm/ |
D | wakeup_asm.S | 53 orb $X86_CR0_PE, %al 63 andb $~X86_CR0_PE, %al
|
D | trampoline_64.S | 73 movl $X86_CR0_PE, %eax # protected mode (PE) bit 152 movl $(X86_CR0_PG | X86_CR0_WP | X86_CR0_PE), %eax
|
/arch/x86/include/uapi/asm/ |
D | processor-flags.h | 52 #define X86_CR0_PE _BITUL(X86_CR0_PE_BIT) macro 162 #define CR0_STATE (X86_CR0_PE | X86_CR0_MP | X86_CR0_ET | \
|
/arch/x86/boot/ |
D | pmjump.S | 38 orb $X86_CR0_PE, %dl # Protected mode
|
/arch/x86/boot/compressed/ |
D | head_64.S | 292 movl $(X86_CR0_PG | X86_CR0_PE), %eax /* Enable Paging and Protected mode */ 682 movl $(X86_CR0_PG | X86_CR0_PE), %eax 951 movl $(X86_CR0_PG | X86_CR0_PE), %ecx /* Enable Paging and Protected mode */
|
/arch/x86/kvm/vmx/ |
D | nested.h | 265 fixed0 &= ~(X86_CR0_PE | X86_CR0_PG); in nested_guest_cr0_valid()
|
D | nested.c | 2803 bool prot_mode = !urg || vmcs12->guest_cr0 & X86_CR0_PE; in nested_check_vm_entry_controls() 3020 if (CC((vmcs12->guest_cr0 & (X86_CR0_PG | X86_CR0_PE)) == X86_CR0_PG)) in nested_vmx_check_guest_state() 6786 #define VMXON_CR0_ALWAYSON (X86_CR0_PE | X86_CR0_PG | X86_CR0_NE) in nested_vmx_setup_ctls_msrs()
|
D | vmx.c | 139 (KVM_VM_CR0_ALWAYS_ON_UNRESTRICTED_GUEST | X86_CR0_PG | X86_CR0_PE) 3123 if (vmx->rmode.vm86_active && (cr0 & X86_CR0_PE)) in vmx_set_cr0() 3126 if (!vmx->rmode.vm86_active && !(cr0 & X86_CR0_PE)) in vmx_set_cr0()
|
/arch/x86/kernel/ |
D | relocate_kernel_64.S | 131 orl $(X86_CR0_PG | X86_CR0_PE), %eax
|
D | relocate_kernel_32.S | 118 orl $(X86_CR0_PE), %eax
|
/arch/x86/kvm/ |
D | x86.h | 141 return kvm_read_cr0_bits(vcpu, X86_CR0_PE); in is_protmode()
|
D | pmu.c | 371 (kvm_read_cr0(vcpu) & X86_CR0_PE)) in kvm_pmu_rdpmc()
|
D | emulate.c | 823 if (!(ctxt->ops->get_cr(ctxt, 0) & X86_CR0_PE)) { in emulator_recalc_and_set_mode() 2609 if (cr0 & X86_CR0_PE) in em_rsm() 2610 ctxt->ops->set_cr(ctxt, 0, cr0 & ~(X86_CR0_PG | X86_CR0_PE)); in em_rsm()
|
D | x86.c | 889 if ((cr0 & X86_CR0_PG) && !(cr0 & X86_CR0_PE)) in kvm_is_valid_cr0() 9448 cr0 = vcpu->arch.cr0 & ~(X86_CR0_PE | X86_CR0_EM | X86_CR0_TS | X86_CR0_PG); in enter_smm()
|
/arch/x86/kvm/svm/ |
D | nested.c | 273 CC(!(save->cr0 & X86_CR0_PE)) || in nested_vmcb_check_cr3_cr4() 840 svm_set_cr0(vcpu, svm->vmcb->save.cr0 | X86_CR0_PE); in nested_svm_vmexit() 1353 !(save->cr0 & X86_CR0_PE) || in svm_set_nested_state()
|
D | svm.c | 4212 if (cr0 & X86_CR0_PE) in svm_check_intercept() 4213 val |= X86_CR0_PE; in svm_check_intercept()
|
/arch/x86/include/asm/ |
D | kvm_host.h | 108 (~(unsigned long)(X86_CR0_PE | X86_CR0_MP | X86_CR0_EM | X86_CR0_TS \
|