Searched refs:walk_mmu (Results 1 – 10 of 10) sorted by relevance
/arch/x86/kvm/ |
D | kvm_cache_regs.h | 113 return vcpu->arch.walk_mmu->pdptrs[index]; in kvm_pdptr_read()
|
D | x86.h | 125 return vcpu->arch.walk_mmu == &vcpu->arch.nested_mmu; in mmu_is_nested()
|
D | x86.c | 675 vcpu->arch.walk_mmu; in kvm_inject_emulated_page_fault() 760 return kvm_read_guest_page_mmu(vcpu, vcpu->arch.walk_mmu, gfn, in kvm_read_nested_guest_page() 808 u64 pdpte[ARRAY_SIZE(vcpu->arch.walk_mmu->pdptrs)]; in pdptrs_changed() 826 return memcmp(pdpte, vcpu->arch.walk_mmu->pdptrs, sizeof(pdpte)) != 0; in pdptrs_changed() 865 !load_pdptrs(vcpu, vcpu->arch.walk_mmu, kvm_read_cr3(vcpu))) in kvm_set_cr0() 1022 && !load_pdptrs(vcpu, vcpu->arch.walk_mmu, in kvm_set_cr4() 1072 !load_pdptrs(vcpu, vcpu->arch.walk_mmu, cr3)) in kvm_set_cr3() 6111 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_read() 6119 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_fetch() 6127 return vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in kvm_mmu_gva_to_gpa_write() [all …]
|
/arch/x86/kvm/svm/ |
D | nested.c | 106 vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu; in nested_svm_init_mmu_context() 112 vcpu->arch.walk_mmu = &vcpu->arch.root_mmu; in nested_svm_uninit_mmu_context() 380 if (!load_pdptrs(vcpu, vcpu->arch.walk_mmu, cr3)) in nested_svm_load_cr3()
|
D | svm.c | 1457 load_pdptrs(vcpu, vcpu->arch.walk_mmu, kvm_read_cr3(vcpu)); in svm_cache_reg()
|
/arch/x86/kvm/vmx/ |
D | nested.c | 368 vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu; in nested_ept_init_mmu_context() 374 vcpu->arch.walk_mmu = &vcpu->arch.root_mmu; in nested_ept_uninit_mmu_context() 1138 if (CC(!load_pdptrs(vcpu, vcpu->arch.walk_mmu, cr3))) { in nested_vmx_load_cr3() 2635 vcpu->arch.walk_mmu->inject_page_fault = vmx_inject_page_fault_nested; in prepare_vmcs02() 4267 vcpu->arch.walk_mmu->inject_page_fault = kvm_inject_page_fault; in load_vmcs12_host_state()
|
D | vmx.c | 3043 struct kvm_mmu *mmu = vcpu->arch.walk_mmu; in vmx_ept_load_pdptrs() 3058 struct kvm_mmu *mmu = vcpu->arch.walk_mmu; in ept_save_pdptrs()
|
/arch/x86/kvm/mmu/ |
D | paging_tmpl.h | 504 walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu; in FNAME()
|
D | mmu.c | 5168 kvm_mmu_invalidate_gva(vcpu, vcpu->arch.walk_mmu, gva, INVALID_PAGE); in kvm_mmu_invlpg() 5357 vcpu->arch.walk_mmu = &vcpu->arch.root_mmu; in kvm_mmu_create()
|
/arch/x86/include/asm/ |
D | kvm_host.h | 593 struct kvm_mmu *walk_mmu; member
|