• Home
  • Raw
  • Download

Lines Matching refs:gpa

2958 	gpa_t gpa = data & ~0x3f;  in kvm_pv_enable_async_pf()  local
2983 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, in kvm_pv_enable_async_pf()
3034 gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS; in record_steal_time() local
3047 gpa != ghc->gpa || in record_steal_time()
3052 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, gpa, sizeof(*st)) || in record_steal_time()
3119 mark_page_dirty_in_slot(ghc->memslot, gpa_to_gfn(ghc->gpa)); in record_steal_time()
4110 gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS; in kvm_steal_time_set_preempted() local
4138 gpa != ghc->gpa || in kvm_steal_time_set_preempted()
4148 mark_page_dirty_in_slot(ghc->memslot, gpa_to_gfn(ghc->gpa)); in kvm_steal_time_set_preempted()
6093 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, in translate_nested_gpa() argument
6102 t_gpa = vcpu->arch.mmu->gva_to_gpa(vcpu, gpa, access, exception); in translate_nested_gpa()
6145 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access, in kvm_read_guest_virt_helper() local
6151 if (gpa == UNMAPPED_GVA) in kvm_read_guest_virt_helper()
6153 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, data, in kvm_read_guest_virt_helper()
6179 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, access|PFERR_FETCH_MASK, in kvm_fetch_guest_virt() local
6181 if (unlikely(gpa == UNMAPPED_GVA)) in kvm_fetch_guest_virt()
6187 ret = kvm_vcpu_read_guest_page(vcpu, gpa >> PAGE_SHIFT, val, in kvm_fetch_guest_virt()
6243 gpa_t gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, addr, in kvm_write_guest_virt_helper() local
6250 if (gpa == UNMAPPED_GVA) in kvm_write_guest_virt_helper()
6252 ret = kvm_vcpu_write_guest(vcpu, gpa, data, towrite); in kvm_write_guest_virt_helper()
6314 gpa_t gpa, bool write) in vcpu_is_mmio_gpa() argument
6317 if ((gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) in vcpu_is_mmio_gpa()
6320 if (vcpu_match_mmio_gpa(vcpu, gpa)) { in vcpu_is_mmio_gpa()
6321 trace_vcpu_match_mmio(gva, gpa, write, true); in vcpu_is_mmio_gpa()
6329 gpa_t *gpa, struct x86_exception *exception, in vcpu_mmio_gva_to_gpa() argument
6343 *gpa = vcpu->arch.mmio_gfn << PAGE_SHIFT | in vcpu_mmio_gva_to_gpa()
6345 trace_vcpu_match_mmio(gva, *gpa, write, false); in vcpu_mmio_gva_to_gpa()
6349 *gpa = vcpu->arch.walk_mmu->gva_to_gpa(vcpu, gva, access, exception); in vcpu_mmio_gva_to_gpa()
6351 if (*gpa == UNMAPPED_GVA) in vcpu_mmio_gva_to_gpa()
6354 return vcpu_is_mmio_gpa(vcpu, gva, *gpa, write); in vcpu_mmio_gva_to_gpa()
6357 int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa, in emulator_write_phys() argument
6362 ret = kvm_vcpu_write_guest(vcpu, gpa, val, bytes); in emulator_write_phys()
6365 kvm_page_track_write(vcpu, gpa, val, bytes); in emulator_write_phys()
6372 int (*read_write_emulate)(struct kvm_vcpu *vcpu, gpa_t gpa,
6374 int (*read_write_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
6376 int (*read_write_exit_mmio)(struct kvm_vcpu *vcpu, gpa_t gpa,
6385 vcpu->mmio_fragments[0].gpa, val); in read_prepare()
6393 static int read_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, in read_emulate() argument
6396 return !kvm_vcpu_read_guest(vcpu, gpa, val, bytes); in read_emulate()
6399 static int write_emulate(struct kvm_vcpu *vcpu, gpa_t gpa, in write_emulate() argument
6402 return emulator_write_phys(vcpu, gpa, val, bytes); in write_emulate()
6405 static int write_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, int bytes, void *val) in write_mmio() argument
6407 trace_kvm_mmio(KVM_TRACE_MMIO_WRITE, bytes, gpa, val); in write_mmio()
6408 return vcpu_mmio_write(vcpu, gpa, bytes, val); in write_mmio()
6411 static int read_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, in read_exit_mmio() argument
6414 trace_kvm_mmio(KVM_TRACE_MMIO_READ_UNSATISFIED, bytes, gpa, NULL); in read_exit_mmio()
6418 static int write_exit_mmio(struct kvm_vcpu *vcpu, gpa_t gpa, in write_exit_mmio() argument
6447 gpa_t gpa; in emulator_read_write_onepage() local
6462 gpa = ctxt->gpa_val; in emulator_read_write_onepage()
6463 ret = vcpu_is_mmio_gpa(vcpu, addr, gpa, write); in emulator_read_write_onepage()
6465 ret = vcpu_mmio_gva_to_gpa(vcpu, addr, &gpa, exception, write); in emulator_read_write_onepage()
6470 if (!ret && ops->read_write_emulate(vcpu, gpa, val, bytes)) in emulator_read_write_onepage()
6476 handled = ops->read_write_mmio(vcpu, gpa, bytes, val); in emulator_read_write_onepage()
6480 gpa += handled; in emulator_read_write_onepage()
6486 frag->gpa = gpa; in emulator_read_write_onepage()
6499 gpa_t gpa; in emulator_read_write() local
6533 gpa = vcpu->mmio_fragments[0].gpa; in emulator_read_write()
6541 vcpu->run->mmio.phys_addr = gpa; in emulator_read_write()
6543 return ops->read_write_exit_mmio(vcpu, gpa, val, bytes); in emulator_read_write()
6586 gpa_t gpa; in emulator_cmpxchg_emulated() local
6594 gpa = kvm_mmu_gva_to_gpa_write(vcpu, addr, NULL); in emulator_cmpxchg_emulated()
6596 if (gpa == UNMAPPED_GVA || in emulator_cmpxchg_emulated()
6597 (gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) in emulator_cmpxchg_emulated()
6609 if (((gpa + bytes - 1) & page_line_mask) != (gpa & page_line_mask)) in emulator_cmpxchg_emulated()
6612 if (kvm_vcpu_map(vcpu, gpa_to_gfn(gpa), &map)) in emulator_cmpxchg_emulated()
6615 kaddr = map.hva + offset_in_page(gpa); in emulator_cmpxchg_emulated()
6639 kvm_page_track_write(vcpu, gpa, new, bytes); in emulator_cmpxchg_emulated()
7278 gpa_t gpa = cr2_or_gpa; in reexecute_instruction() local
7293 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2_or_gpa, NULL); in reexecute_instruction()
7299 if (gpa == UNMAPPED_GVA) in reexecute_instruction()
7309 pfn = gfn_to_pfn(vcpu->kvm, gpa_to_gfn(gpa)); in reexecute_instruction()
7329 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); in reexecute_instruction()
7339 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); in reexecute_instruction()
7353 unsigned long last_retry_eip, last_retry_addr, gpa = cr2_or_gpa; in retry_instruction() local
7390 gpa = kvm_mmu_gva_to_gpa_write(vcpu, cr2_or_gpa, NULL); in retry_instruction()
7392 kvm_mmu_unprotect_page(vcpu->kvm, gpa_to_gfn(gpa)); in retry_instruction()
9528 frag->gpa += len; in complete_emulated_mmio()
9543 run->mmio.phys_addr = frag->gpa; in complete_emulated_mmio()
10072 gpa_t gpa; in kvm_arch_vcpu_ioctl_translate() local
10078 gpa = kvm_mmu_gva_to_gpa_system(vcpu, vaddr, NULL); in kvm_arch_vcpu_ioctl_translate()
10080 tr->physical_address = gpa; in kvm_arch_vcpu_ioctl_translate()
10081 tr->valid = gpa != UNMAPPED_GVA; in kvm_arch_vcpu_ioctl_translate()
10703 int __x86_set_memory_region(struct kvm *kvm, int id, gpa_t gpa, u32 size) in __x86_set_memory_region() argument
10740 m.guest_phys_addr = gpa; in __x86_set_memory_region()