Home
last modified time | relevance | path

Searched refs:gpa (Results 1 – 25 of 64) sorted by relevance

123

/arch/powerpc/kvm/
Dbook3s_64_mmu_radix.c137 u64 pte, base, gpa; in kvmppc_mmu_walk_radix_tree() local
191 gpa = pte & 0x01fffffffffff000ul; in kvmppc_mmu_walk_radix_tree()
192 if (gpa & ((1ul << offset) - 1)) in kvmppc_mmu_walk_radix_tree()
194 gpa |= eaddr & ((1ul << offset) - 1); in kvmppc_mmu_walk_radix_tree()
202 gpte->raddr = gpa; in kvmppc_mmu_walk_radix_tree()
394 void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa, in kvmppc_unmap_pte() argument
401 unsigned long gfn = gpa >> PAGE_SHIFT; in kvmppc_unmap_pte()
405 old = kvmppc_radix_update_pte(kvm, pte, ~0UL, 0, gpa, shift); in kvmppc_unmap_pte()
406 kvmppc_radix_tlbie_page(kvm, gpa, shift, lpid); in kvmppc_unmap_pte()
425 gpa &= ~(page_size - 1); in kvmppc_unmap_pte()
[all …]
Dbook3s_hv_uvmem.c232 unsigned long gpa; member
509 struct kvm *kvm, unsigned long gpa) in __kvmppc_svm_page_out() argument
528 if (!kvmppc_gfn_is_uvmem_pfn(gpa >> page_shift, kvm, NULL)) in __kvmppc_svm_page_out()
561 gpa, 0, page_shift); in __kvmppc_svm_page_out()
581 struct kvm *kvm, unsigned long gpa) in kvmppc_svm_page_out() argument
586 ret = __kvmppc_svm_page_out(vma, start, end, page_shift, kvm, gpa); in kvmppc_svm_page_out()
635 PAGE_SHIFT, kvm, pvt->gpa)) in kvmppc_uvmem_drop_pages()
637 pvt->gpa, addr); in kvmppc_uvmem_drop_pages()
685 static struct page *kvmppc_uvmem_get_page(unsigned long gpa, struct kvm *kvm) in kvmppc_uvmem_get_page() argument
709 kvmppc_gfn_secure_uvmem_pfn(gpa >> PAGE_SHIFT, uvmem_pfn, kvm); in kvmppc_uvmem_get_page()
[all …]
Dbook3s_hv_nested.c840 unsigned long gpa; in kvmhv_update_nest_rmap_rc() local
844 gpa = n_rmap & RMAP_NESTED_GPA_MASK; in kvmhv_update_nest_rmap_rc()
848 ptep = find_kvm_nested_guest_pte(kvm, lpid, gpa, &shift); in kvmhv_update_nest_rmap_rc()
857 kvmppc_radix_tlbie_page(kvm, gpa, shift, lpid); in kvmhv_update_nest_rmap_rc()
887 unsigned long gpa; in kvmhv_remove_nest_rmap() local
891 gpa = n_rmap & RMAP_NESTED_GPA_MASK; in kvmhv_remove_nest_rmap()
898 ptep = find_kvm_nested_guest_pte(kvm, lpid, gpa, &shift); in kvmhv_remove_nest_rmap()
901 kvmppc_unmap_pte(kvm, ptep, gpa, shift, NULL, gp->shadow_lpid); in kvmhv_remove_nest_rmap()
920 unsigned long gpa, unsigned long hpa, in kvmhv_remove_nest_rmap_range() argument
928 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; in kvmhv_remove_nest_rmap_range()
[all …]
De500.h255 gpa_t gpa; in tlbe_is_host_safe() local
267 gpa = get_tlb_raddr(tlbe); in tlbe_is_host_safe()
268 if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT)) in tlbe_is_host_safe()
/arch/s390/kvm/
Dgaccess.h152 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in write_guest_lc() local
154 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_lc()
178 unsigned long gpa = gra + kvm_s390_get_prefix(vcpu); in read_guest_lc() local
180 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_lc()
190 u8 ar, unsigned long *gpa, enum gacc_mode mode);
310 int write_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in write_guest_abs() argument
313 return kvm_write_guest(vcpu->kvm, gpa, data, len); in write_guest_abs()
333 int read_guest_abs(struct kvm_vcpu *vcpu, unsigned long gpa, void *data, in read_guest_abs() argument
336 return kvm_read_guest(vcpu->kvm, gpa, data, len); in read_guest_abs()
Dvsie.c643 static int pin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t *hpa) in pin_guest_page() argument
647 page = gfn_to_page(kvm, gpa_to_gfn(gpa)); in pin_guest_page()
650 *hpa = (hpa_t) page_to_virt(page) + (gpa & ~PAGE_MASK); in pin_guest_page()
655 static void unpin_guest_page(struct kvm *kvm, gpa_t gpa, hpa_t hpa) in unpin_guest_page() argument
659 mark_page_dirty(kvm, gpa_to_gfn(gpa)); in unpin_guest_page()
724 gpa_t gpa; in pin_blocks() local
727 gpa = READ_ONCE(scb_o->scaol) & ~0xfUL; in pin_blocks()
729 gpa |= (u64) READ_ONCE(scb_o->scaoh) << 32; in pin_blocks()
730 if (gpa) { in pin_blocks()
731 if (gpa < 2 * PAGE_SIZE) in pin_blocks()
[all …]
Dgaccess.c588 static int deref_table(struct kvm *kvm, unsigned long gpa, unsigned long *val) in deref_table() argument
590 return kvm_read_guest(kvm, gpa, val, sizeof(*val)); in deref_table()
615 unsigned long *gpa, const union asce asce, in guest_translate() argument
774 *gpa = raddr.addr; in guest_translate()
834 unsigned long _len, nr_pages, gpa, idx; in access_guest() local
858 gpa = *(pages + idx) + (ga & ~PAGE_MASK); in access_guest()
859 _len = min(PAGE_SIZE - (gpa & ~PAGE_MASK), len); in access_guest()
861 rc = kvm_write_guest(vcpu->kvm, gpa, data, _len); in access_guest()
863 rc = kvm_read_guest(vcpu->kvm, gpa, data, _len); in access_guest()
878 unsigned long _len, gpa; in access_guest_real() local
[all …]
/arch/x86/include/asm/uv/
Duv_hub.h460 uv_gpa_in_mmr_space(unsigned long gpa) in uv_gpa_in_mmr_space() argument
462 return (gpa >> 62) == 0x3UL; in uv_gpa_in_mmr_space()
466 static inline unsigned long uv_gpa_to_soc_phys_ram(unsigned long gpa) in uv_gpa_to_soc_phys_ram() argument
474 gpa = ((gpa << uv_hub_info->m_shift) >> uv_hub_info->m_shift) | in uv_gpa_to_soc_phys_ram()
475 ((gpa >> uv_hub_info->n_lshift) << uv_hub_info->m_val); in uv_gpa_to_soc_phys_ram()
477 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram()
484 static inline unsigned long uv_gpa_to_gnode(unsigned long gpa) in uv_gpa_to_gnode() argument
489 return gpa >> n_lshift; in uv_gpa_to_gnode()
491 return uv_gam_range(gpa)->nasid >> 1; in uv_gpa_to_gnode()
495 static inline int uv_gpa_to_pnode(unsigned long gpa) in uv_gpa_to_pnode() argument
[all …]
/arch/arm64/kvm/
Dhypercalls.c17 gpa_t gpa; in kvm_hvc_call_handler() local
82 gpa = kvm_init_stolen_time(vcpu); in kvm_hvc_call_handler()
83 if (gpa != GPA_INVALID) in kvm_hvc_call_handler()
84 val = gpa; in kvm_hvc_call_handler()
Dmmu.c520 gpa_t gpa = addr + (vm_start - memslot->userspace_addr); in stage2_unmap_memslot() local
521 unmap_stage2_range(&kvm->arch.mmu, gpa, vm_end - vm_start); in stage2_unmap_memslot()
1154 gpa_t gpa, u64 size, in handle_hva_to_gpa() argument
1167 gfn_t gpa; in handle_hva_to_gpa() local
1175 gpa = hva_to_gfn_memslot(hva_start, memslot) << PAGE_SHIFT; in handle_hva_to_gpa()
1176 ret |= handler(kvm, gpa, (u64)(hva_end - hva_start), data); in handle_hva_to_gpa()
1182 static int kvm_unmap_hva_handler(struct kvm *kvm, gpa_t gpa, u64 size, void *data) in kvm_unmap_hva_handler() argument
1187 __unmap_stage2_range(&kvm->arch.mmu, gpa, size, may_block); in kvm_unmap_hva_handler()
1202 static int kvm_set_spte_handler(struct kvm *kvm, gpa_t gpa, u64 size, void *data) in kvm_set_spte_handler() argument
1214 kvm_pgtable_stage2_map(kvm->arch.mmu.pgt, gpa, PAGE_SIZE, in kvm_set_spte_handler()
[all …]
/arch/x86/xen/
Dmmu_hvm.c53 a.gpa = __pa(mm->pgd); in xen_hvm_exit_mmap()
64 a.gpa = 0x00; in is_pagetable_dying_supported()
/arch/x86/include/asm/
Dkvm_page_track.h35 void (*track_write)(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new,
71 void kvm_page_track_write(struct kvm_vcpu *vcpu, gpa_t gpa, const u8 *new,
/arch/arm64/include/asm/
Dkvm_mmu.h235 gpa_t gpa, void *data, unsigned long len) in kvm_read_guest_lock() argument
238 int ret = kvm_read_guest(kvm, gpa, data, len); in kvm_read_guest_lock()
245 static inline int kvm_write_guest_lock(struct kvm *kvm, gpa_t gpa, in kvm_write_guest_lock() argument
249 int ret = kvm_write_guest(kvm, gpa, data, len); in kvm_write_guest_lock()
/arch/x86/kvm/vmx/
Dvmx_ops.h22 void invept_error(unsigned long ext, u64 eptp, gpa_t gpa);
266 static inline void __invept(unsigned long ext, u64 eptp, gpa_t gpa) in __invept() argument
269 u64 eptp, gpa; in __invept() member
270 } operand = {eptp, gpa}; in __invept()
272 vmx_asm2(invept, "r"(ext), "m"(operand), ext, eptp, gpa); in __invept()
/arch/mips/kvm/
Dmmu.c501 gpa_t gpa = gfn << PAGE_SHIFT; in kvm_set_spte_handler() local
503 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_set_spte_handler()
549 gpa_t gpa = gfn << PAGE_SHIFT; in kvm_test_age_hva_handler() local
550 pte_t *gpa_pte = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in kvm_test_age_hva_handler()
585 static int _kvm_mips_map_page_fast(struct kvm_vcpu *vcpu, unsigned long gpa, in _kvm_mips_map_page_fast() argument
590 gfn_t gfn = gpa >> PAGE_SHIFT; in _kvm_mips_map_page_fast()
599 ptep = kvm_mips_pte_for_gpa(kvm, NULL, gpa); in _kvm_mips_map_page_fast()
661 static int kvm_mips_map_page(struct kvm_vcpu *vcpu, unsigned long gpa, in kvm_mips_map_page() argument
667 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mips_map_page()
677 err = _kvm_mips_map_page_fast(vcpu, gpa, write_fault, out_entry, in kvm_mips_map_page()
[all …]
/arch/x86/kvm/
Dcpuid.h39 static inline bool kvm_vcpu_is_illegal_gpa(struct kvm_vcpu *vcpu, gpa_t gpa) in kvm_vcpu_is_illegal_gpa() argument
41 return (gpa >= BIT_ULL(cpuid_maxphyaddr(vcpu))); in kvm_vcpu_is_illegal_gpa()
328 static inline bool page_address_valid(struct kvm_vcpu *vcpu, gpa_t gpa) in page_address_valid() argument
330 return PAGE_ALIGNED(gpa) && !(gpa >> cpuid_maxphyaddr(vcpu)); in page_address_valid()
Dx86.c2958 gpa_t gpa = data & ~0x3f; in kvm_pv_enable_async_pf() local
2983 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, &vcpu->arch.apf.data, gpa, in kvm_pv_enable_async_pf()
3034 gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS; in record_steal_time() local
3047 gpa != ghc->gpa || in record_steal_time()
3052 if (kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, gpa, sizeof(*st)) || in record_steal_time()
3119 mark_page_dirty_in_slot(ghc->memslot, gpa_to_gfn(ghc->gpa)); in record_steal_time()
4110 gpa_t gpa = vcpu->arch.st.msr_val & KVM_STEAL_VALID_BITS; in kvm_steal_time_set_preempted() local
4138 gpa != ghc->gpa || in kvm_steal_time_set_preempted()
4148 mark_page_dirty_in_slot(ghc->memslot, gpa_to_gfn(ghc->gpa)); in kvm_steal_time_set_preempted()
6093 gpa_t translate_nested_gpa(struct kvm_vcpu *vcpu, gpa_t gpa, u32 access, in translate_nested_gpa() argument
[all …]
/arch/powerpc/include/asm/
Dultravisor.h75 static inline int uv_page_inval(u64 lpid, u64 gpa, u64 page_shift) in uv_page_inval() argument
77 return ucall_norets(UV_PAGE_INVAL, lpid, gpa, page_shift); in uv_page_inval()
Dkvm_book3s.h163 unsigned long gpa, gva_t ea, int is_store);
195 extern void kvmppc_unmap_pte(struct kvm *kvm, pte_t *pte, unsigned long gpa,
200 bool writing, unsigned long gpa,
203 unsigned long gpa,
236 extern kvm_pfn_t kvmppc_gpa_to_pfn(struct kvm_vcpu *vcpu, gpa_t gpa,
249 unsigned long gpa, bool dirty);
Diommu.h300 unsigned long gpa);
306 #define iommu_tce_put_param_check(tbl, ioba, gpa) \ argument
310 iommu_tce_check_gpa((tbl)->it_page_shift, (gpa)))
/arch/ia64/include/asm/uv/
Duv_hub.h166 static inline void *uv_va(unsigned long gpa) in uv_va() argument
168 return __va(gpa & uv_hub_info->gpa_mask); in uv_va()
/arch/x86/kvm/mmu/
Dspte.c52 u64 gpa = gfn << PAGE_SHIFT; in make_mmio_spte() local
56 mask |= gpa | shadow_nonpresent_or_rsvd_mask; in make_mmio_spte()
57 mask |= (gpa & shadow_nonpresent_or_rsvd_mask) in make_mmio_spte()
/arch/arm/boot/dts/
Ds3c64xx-pinctrl.dtsi19 gpa: gpa-gpio-bank { label
135 samsung,pins = "gpa-0", "gpa-1";
141 samsung,pins = "gpa-2", "gpa-3";
147 samsung,pins = "gpa-4", "gpa-5";
153 samsung,pins = "gpa-6", "gpa-7";
Ds3c2416-pinctrl.dtsi15 gpa: gpa { label
/arch/arm64/kvm/vgic/
Dvgic-its.c2103 gpa_t gpa = base; in scan_its_table() local
2113 ret = kvm_read_guest_lock(kvm, gpa, entry, esz); in scan_its_table()
2126 gpa += byte_offset; in scan_its_table()
2136 struct its_ite *ite, gpa_t gpa, int ite_esz) in vgic_its_save_ite() argument
2147 return kvm_write_guest_lock(kvm, gpa, &val, ite_esz); in vgic_its_save_ite()
2229 gpa_t gpa = base + ite->event_id * ite_esz; in vgic_its_save_itt() local
2240 ret = vgic_its_save_ite(its, device, ite, gpa, ite_esz); in vgic_its_save_itt()
2414 gpa_t gpa; in handle_l1_dte() local
2422 gpa = entry & KVM_ITS_L1E_ADDR_MASK; in handle_l1_dte()
2424 ret = scan_its_table(its, gpa, SZ_64K, dte_esz, in handle_l1_dte()
[all …]

123