Lines Matching refs:mmu
34 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument
53 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument
66 #define PT_HAVE_ACCESSED_DIRTY(mmu) ((mmu)->ept_ad) argument
105 static inline void FNAME(protect_clean_gpte)(struct kvm_mmu *mmu, unsigned *access, in FNAME()
111 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME()
132 static int FNAME(cmpxchg_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in FNAME()
175 if (is_rsvd_bits_set(vcpu->arch.mmu, gpte, PT_PAGE_TABLE_LEVEL)) in FNAME()
182 if (PT_HAVE_ACCESSED_DIRTY(vcpu->arch.mmu) && in FNAME()
218 struct kvm_mmu *mmu, in FNAME()
229 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME()
269 ret = FNAME(cmpxchg_gpte)(vcpu, mmu, ptep_user, index, orig_pte, pte); in FNAME()
294 struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, in FNAME()
317 walker->level = mmu->root_level; in FNAME()
318 pte = mmu->get_cr3(vcpu); in FNAME()
319 have_ad = PT_HAVE_ACCESSED_DIRTY(mmu); in FNAME()
324 pte = mmu->get_pdptr(vcpu, (addr >> 30) & 3); in FNAME()
360 real_gfn = mmu->translate_gpa(vcpu, gfn_to_gpa(table_gfn), in FNAME()
400 if (unlikely(is_rsvd_bits_set(mmu, pte, walker->level))) { in FNAME()
409 } while (!is_last_gpte(mmu, walker->level, pte)); in FNAME()
416 errcode = permission_fault(vcpu, mmu, walker->pte_access, pte_pkey, access); in FNAME()
426 real_gpa = mmu->translate_gpa(vcpu, gfn_to_gpa(gfn), access, &walker->fault); in FNAME()
433 FNAME(protect_clean_gpte)(mmu, &walker->pte_access, pte); in FNAME()
444 ret = FNAME(update_accessed_dirty_bits)(vcpu, mmu, walker, in FNAME()
459 if (fetch_fault && (mmu->nx || in FNAME()
493 walker->fault.nested_page_fault = mmu != vcpu->arch.walk_mmu; in FNAME()
502 return FNAME(walk_addr_generic)(walker, vcpu, vcpu->arch.mmu, addr, in FNAME()
531 FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte); in FNAME()
629 top_level = vcpu->arch.mmu->root_level; in FNAME()
641 if (!VALID_PAGE(vcpu->arch.mmu->root_hpa)) in FNAME()
1050 FNAME(protect_clean_gpte)(vcpu->arch.mmu, &pte_access, gpte); in FNAME()