/arch/powerpc/kvm/ |
D | book3s_64_mmu_radix.c | 137 u64 pte, base, gpa; in kvmppc_mmu_walk_radix_tree() local 191 gpa = pte & 0x01fffffffffff000ul; in kvmppc_mmu_walk_radix_tree() 192 if (gpa & ((1ul << offset) - 1)) in kvmppc_mmu_walk_radix_tree() 194 gpa |= eaddr & ((1ul << offset) - 1); in kvmppc_mmu_walk_radix_tree() 202 gpte->raddr = gpa; in kvmppc_mmu_walk_radix_tree() 394 void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa, in kvmppc_unmap_pte() argument 401 unsigned long gfn = gpa >> PAGE_SHIFT; in kvmppc_unmap_pte() 405 old = kvmppc_radix_update_pte(kvm, pte, ~0UL, 0, gpa, shift); in kvmppc_unmap_pte() 406 kvmppc_radix_tlbie_page(kvm, gpa, shift, lpid); in kvmppc_unmap_pte() 425 gpa &= ~(page_size - 1); in kvmppc_unmap_pte() [all …]
|
D | book3s_hv_uvmem.c | 232 unsigned long gpa; member 509 struct kvm *kvm, unsigned long gpa) in __kvmppc_svm_page_out() argument 528 if (!kvmppc_gfn_is_uvmem_pfn(gpa >> page_shift, kvm, NULL)) in __kvmppc_svm_page_out() 561 gpa, 0, page_shift); in __kvmppc_svm_page_out() 581 struct kvm *kvm, unsigned long gpa) in kvmppc_svm_page_out() argument 586 ret = __kvmppc_svm_page_out(vma, start, end, page_shift, kvm, gpa); in kvmppc_svm_page_out() 635 PAGE_SHIFT, kvm, pvt->gpa)) in kvmppc_uvmem_drop_pages() 637 pvt->gpa, addr); in kvmppc_uvmem_drop_pages() 685 static struct page *kvmppc_uvmem_get_page(unsigned long gpa, struct kvm *kvm) in kvmppc_uvmem_get_page() argument 709 kvmppc_gfn_secure_uvmem_pfn(gpa >> PAGE_SHIFT, uvmem_pfn, kvm); in kvmppc_uvmem_get_page() [all …]
|
D | book3s_hv_nested.c | 840 unsigned long gpa; in kvmhv_update_nest_rmap_rc() local 844 gpa = n_rmap & RMAP_NESTED_GPA_MASK; in kvmhv_update_nest_rmap_rc() 848 ptep = find_kvm_nested_guest_pte(kvm, lpid, gpa, &shift); in kvmhv_update_nest_rmap_rc() 857 kvmppc_radix_tlbie_page(kvm, gpa, shift, lpid); in kvmhv_update_nest_rmap_rc() 887 unsigned long gpa; in kvmhv_remove_nest_rmap() local 891 gpa = n_rmap & RMAP_NESTED_GPA_MASK; in kvmhv_remove_nest_rmap() 898 ptep = find_kvm_nested_guest_pte(kvm, lpid, gpa, &shift); in kvmhv_remove_nest_rmap() 901 kvmppc_unmap_pte(kvm, ptep, gpa, shift, NULL, gp->shadow_lpid); in kvmhv_remove_nest_rmap() 920 unsigned long gpa, unsigned long hpa, in kvmhv_remove_nest_rmap_range() argument 928 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; in kvmhv_remove_nest_rmap_range() [all …]
|
D | e500.h | 255 gpa_t gpa; in tlbe_is_host_safe() local 267 gpa = get_tlb_raddr(tlbe); in tlbe_is_host_safe() 268 if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT)) in tlbe_is_host_safe()
|
/arch/s390/kvm/ |
D | gaccess.h | 152 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in write_guest_lc() local 154 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_lc() 178 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in read_guest_lc() local 180 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_lc() 190 u8 ar, unsigned long *gpa, enum gacc_mode mode); 310 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in write_guest_abs() argument 313 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_abs() 333 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in read_guest_abs() argument 336 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_abs()
|
D | vsie.c | 643 static int pin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t *hpa) in pin_guest_page() argument 647 page = gfn_to_page(kvm, gpa_to_gfn(gpa)); in pin_guest_page() 650 *hpa = (hpa_t) page_to_virt(page) + (gpa & ~PAGE_MASK); in pin_guest_page() 655 static void unpin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t hpa) in unpin_guest_page() argument 659 mark_page_dirty(kvm, gpa_to_gfn(gpa)); in unpin_guest_page() 724 gpa_t gpa; in pin_blocks() local 727 gpa = READ_ONCE(scb_o->scaol) & ~0xfUL; in pin_blocks() 729 gpa |= (u64) READ_ONCE(scb_o->scaoh) << 32; in pin_blocks() 730 if (gpa) { in pin_blocks() 731 if (gpa < 2 * PAGE_SIZE) in pin_blocks() [all …]
|
D | gaccess.c | 588 static int deref_table(struct kvm *kvm, unsigned long gpa, unsigned long *val) in deref_table() argument 590 return kvm_read_guest(kvm, gpa, val, sizeof(*val)); in deref_table() 615 unsigned long *gpa, const union asce asce, in guest_translate() argument 774 *gpa = raddr.addr; in guest_translate() 834 unsigned long _len, nr_pages, gpa, idx; in access_guest() local 858 gpa = *(pages + idx) + (ga & ~PAGE_MASK); in access_guest() 859 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); in access_guest() 861 rc = kvm_write_guest(vcpu->kvm, gpa, data, _len); in access_guest() 863 rc = kvm_read_guest(vcpu->kvm, gpa, data, _len); in access_guest() 878 unsigned long _len, gpa; in access_guest_real() local [all …]
|
/arch/x86/include/asm/uv/ |
D | uv_hub.h | 460 uv_gpa_in_mmr_space(unsigned long gpa) in uv_gpa_in_mmr_space() argument 462 return (gpa >> 62) == 0x3UL; in uv_gpa_in_mmr_space() 466 static inline unsigned long uv_gpa_to_soc_phys_ram(unsigned long gpa) in uv_gpa_to_soc_phys_ram() argument 474 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | in uv_gpa_to_soc_phys_ram() 475 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); in uv_gpa_to_soc_phys_ram() 477 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram() 484 static inline unsigned long uv_gpa_to_gnode(unsigned long gpa) in uv_gpa_to_gnode() argument 489 return gpa >> n_lshift; in uv_gpa_to_gnode() 491 return uv_gam_range(gpa)->nasid >> 1; in uv_gpa_to_gnode() 495 static inline int uv_gpa_to_pnode(unsigned long gpa) in uv_gpa_to_pnode() argument [all …]
|
/arch/arm64/kvm/ |
D | hypercalls.c | 17 gpa_t gpa; in kvm_hvc_call_handler() local 82 gpa = kvm_init_stolen_time(vcpu); in kvm_hvc_call_handler() 83 if (gpa != GPA_INVALID) in kvm_hvc_call_handler() 84 val = gpa; in kvm_hvc_call_handler()
|
D | mmu.c | 520 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot() local 521 unmap_stage2_range(&kvm->arch.mmu, gpa, vm_end - vm_start); in stage2_unmap_memslot() 1154 gpa_t gpa, u64 size, in handle_hva_to_gpa() argument 1167 gfn_t gpa; in handle_hva_to_gpa() local 1175 gpa = hva_to_gfn_memslot(hva_start, memslot) << PAGE_SHIFT; in handle_hva_to_gpa() 1176 ret |= handler(kvm, gpa, (u64)(hva_end - hva_start), data); in handle_hva_to_gpa() 1182 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, u64 size, void *data) in kvm_unmap_hva_handler() argument 1187 __unmap_stage2_range(&kvm->arch.mmu, gpa, size, may_block); in kvm_unmap_hva_handler() 1202 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, u64 size, void *data) in kvm_set_spte_handler() argument 1214 kvm_pgtable_stage2_map(kvm->arch.mmu.pgt, gpa, PAGE_SIZE, in kvm_set_spte_handler() [all …]
|
/arch/x86/xen/ |
D | mmu_hvm.c | 53 a.gpa = __pa(mm->pgd); in xen_hvm_exit_mmap() 64 a.gpa = 0x00; in is_pagetable_dying_supported()
|
/arch/x86/include/asm/ |
D | kvm_page_track.h | 35 void (*track_write)(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new, 71 void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new,
|
/arch/arm64/include/asm/ |
D | kvm_mmu.h | 235 gpa_t gpa, void *data, unsigned long len) in kvm_read_guest_lock() argument 238 int ret = kvm_read_guest(kvm, gpa, data, len); in kvm_read_guest_lock() 245 static inline int kvm_write_guest_lock(struct kvm *kvm, gpa_t gpa, in kvm_write_guest_lock() argument 249 int ret = kvm_write_guest(kvm, gpa, data, len); in kvm_write_guest_lock()
|
/arch/x86/kvm/vmx/ |
D | vmx_ops.h | 22 void invept_error(unsigned long ext, u64 eptp, gpa_t gpa); 266 static inline void __invept(unsigned long ext, u64 eptp, gpa_t gpa) in __invept() argument 269 u64 eptp, gpa; in __invept() member 270 } operand = {eptp, gpa}; in __invept() 272 vmx_asm2(invept, "r"(ext), "m"(operand), ext, eptp, gpa); in __invept()
|
/arch/mips/kvm/ |
D | mmu.c | 501 gpa_t gpa = gfn << PAGE_SHIFT; in kvm_set_spte_handler() local 503 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_set_spte_handler() 549 gpa_t gpa = gfn << PAGE_SHIFT; in kvm_test_age_hva_handler() local 550 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_test_age_hva_handler() 585 static int _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, in _kvm_mips_map_page_fast() argument 590 gfn_t gfn = gpa >> PAGE_SHIFT; in _kvm_mips_map_page_fast() 599 ptep = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in _kvm_mips_map_page_fast() 661 static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, in kvm_mips_map_page() argument 667 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mips_map_page() 677 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page() [all …]
|
/arch/x86/kvm/ |
D | cpuid.h | 39 static inline bool kvm_vcpu_is_illegal_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in kvm_vcpu_is_illegal_gpa() argument 41 return (gpa >= BIT_ULL(cpuid_maxphyaddr(vcpu))); in kvm_vcpu_is_illegal_gpa() 328 static inline bool page_address_valid(struct kvm_vcpu *vcpu, gpa_t gpa) in page_address_valid() argument 330 return PAGE_ALIGNED(gpa) && !(gpa >> cpuid_maxphyaddr(vcpu)); in page_address_valid()
|
D | x86.c | 2958 gpa_t gpa = data & ~0x3f; in kvm_pv_enable_async_pf() local 2983 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, in kvm_pv_enable_async_pf() 3034 gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS; in record_steal_time() local 3047 gpa != ghc->gpa || in record_steal_time() 3052 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, gpa, sizeof(*st)) || in record_steal_time() 3119 mark_page_dirty_in_slot(ghc->memslot, gpa_to_gfn(ghc->gpa)); in record_steal_time() 4110 gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS; in kvm_steal_time_set_preempted() local 4138 gpa != ghc->gpa || in kvm_steal_time_set_preempted() 4148 mark_page_dirty_in_slot(ghc->memslot, gpa_to_gfn(ghc->gpa)); in kvm_steal_time_set_preempted() 6093 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, in translate_nested_gpa() argument [all …]
|
/arch/powerpc/include/asm/ |
D | ultravisor.h | 75 static inline int uv_page_inval(u64 lpid, u64 gpa, u64 page_shift) in uv_page_inval() argument 77 return ucall_norets(UV_PAGE_INVAL, lpid, gpa, page_shift); in uv_page_inval()
|
D | kvm_book3s.h | 163 unsigned long gpa, gva_t ea, int is_store); 195 extern void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa, 200 bool writing, unsigned long gpa, 203 unsigned long gpa, 236 extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa, 249 unsigned long gpa, bool dirty);
|
D | iommu.h | 300 unsigned long gpa); 306 #define iommu_tce_put_param_check(tbl, ioba, gpa) \ argument 310 iommu_tce_check_gpa((tbl)->it_page_shift, (gpa)))
|
/arch/ia64/include/asm/uv/ |
D | uv_hub.h | 166 static inline void *uv_va(unsigned long gpa) in uv_va() argument 168 return __va(gpa & uv_hub_info->gpa_mask); in uv_va()
|
/arch/x86/kvm/mmu/ |
D | spte.c | 52 u64 gpa = gfn << PAGE_SHIFT; in make_mmio_spte() local 56 mask |= gpa | shadow_nonpresent_or_rsvd_mask; in make_mmio_spte() 57 mask |= (gpa & shadow_nonpresent_or_rsvd_mask) in make_mmio_spte()
|
/arch/arm/boot/dts/ |
D | s3c64xx-pinctrl.dtsi | 19 gpa: gpa-gpio-bank { label 135 samsung,pins = "gpa-0", "gpa-1"; 141 samsung,pins = "gpa-2", "gpa-3"; 147 samsung,pins = "gpa-4", "gpa-5"; 153 samsung,pins = "gpa-6", "gpa-7";
|
D | s3c2416-pinctrl.dtsi | 15 gpa: gpa { label
|
/arch/arm64/kvm/vgic/ |
D | vgic-its.c | 2103 gpa_t gpa = base; in scan_its_table() local 2113 ret = kvm_read_guest_lock(kvm, gpa, entry, esz); in scan_its_table() 2126 gpa += byte_offset; in scan_its_table() 2136 struct its_ite *ite, gpa_t gpa, int ite_esz) in vgic_its_save_ite() argument 2147 return kvm_write_guest_lock(kvm, gpa, &val, ite_esz); in vgic_its_save_ite() 2229 gpa_t gpa = base + ite->event_id * ite_esz; in vgic_its_save_itt() local 2240 ret = vgic_its_save_ite(its, device, ite, gpa, ite_esz); in vgic_its_save_itt() 2414 gpa_t gpa; in handle_l1_dte() local 2422 gpa = entry & KVM_ITS_L1E_ADDR_MASK; in handle_l1_dte() 2424 ret = scan_its_table(its, gpa, SZ_64K, dte_esz, in handle_l1_dte() [all …]
|