Home
last modified time | relevance | path

Searched refs:base_gfn (Results 1 – 19 of 19) sorted by relevance

/arch/powerpc/kvm/
Dtrace_hv.h285 __field(u64, base_gfn)
297 __entry->base_gfn = memslot ? memslot->base_gfn : -1UL;
305 __entry->base_gfn, __entry->slot_flags)
Dbook3s_64_mmu_hv.c576 if (gfn_base < memslot->base_gfn) in kvmppc_book3s_hv_page_fault()
687 rmap = &memslot->arch.rmap[gfn_base - memslot->base_gfn]; in kvmppc_book3s_hv_page_fault()
861 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_unmap_rmapp()
907 gfn = memslot->base_gfn; in kvmppc_core_flush_memslot_hv()
936 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_rmapp()
999 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_test_age_rmapp()
1138 if (gfn < memslot->base_gfn || in kvmppc_harvest_vpa_dirty()
1139 gfn >= memslot->base_gfn + memslot->npages) in kvmppc_harvest_vpa_dirty()
1144 __set_bit_le(gfn - memslot->base_gfn, map); in kvmppc_harvest_vpa_dirty()
1219 set_bit_le(gfn - memslot->base_gfn, memslot->dirty_bitmap); in kvmppc_unpin_guest_page()
[all …]
Dbook3s_64_mmu_radix.c1011 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_age_radix()
1039 unsigned long gfn = memslot->base_gfn + pagenum; in kvm_radix_test_clear_dirty()
1056 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvm_radix_test_clear_dirty()
1098 gpa = memslot->base_gfn << PAGE_SHIFT; in kvmppc_radix_flush_memslot()
Dbook3s_hv_rm_mmu.c117 gfn -= memslot->base_gfn; in kvmppc_update_dirty_map()
155 rmap = real_vmalloc_addr(&memslot->arch.rmap[gfn - memslot->base_gfn]); in revmap_for_hpte()
246 slot_fn = gfn - memslot->base_gfn; in kvmppc_do_h_enter()
Dbook3s_64_vio_hv.c92 *prmap = &memslot->arch.rmap[gfn - memslot->base_gfn]; in kvmppc_rm_tce_to_ua()
De500_mmu_host.c381 slot_start = pfn - (gfn - slot->base_gfn); in kvmppc_e500_shadow_map()
Dbook3s_hv_nested.c908 gfn = (gpa >> PAGE_SHIFT) - memslot->base_gfn; in kvmhv_remove_nest_rmap_range()
1437 rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn]; in __kvmhv_nested_page_fault()
Dbook3s_pr.c1884 ga = memslot->base_gfn << PAGE_SHIFT; in kvm_vm_ioctl_get_dirty_log_pr()
Dbook3s_hv.c814 if ((from + len) >= ((from_memslot->base_gfn + from_memslot->npages) in kvmppc_copy_guest()
826 if ((to + len) >= ((to_memslot->base_gfn + to_memslot->npages) in kvmppc_copy_guest()
/arch/x86/kvm/
Dpage_track.c68 index = gfn_to_index(gfn, slot->base_gfn, PT_PAGE_TABLE_LEVEL); in update_gfn_track()
158 index = gfn_to_index(gfn, slot->base_gfn, PT_PAGE_TABLE_LEVEL); in kvm_page_track_is_active()
Dpaging_tmpl.h625 gfn_t gfn, base_gfn; in FNAME() local
676 base_gfn = gfn; in FNAME()
689 base_gfn = gfn & ~(KVM_PAGES_PER_HPAGE(it.level) - 1); in FNAME()
698 sp = kvm_mmu_get_page(vcpu, base_gfn, addr, in FNAME()
707 it.level, base_gfn, pfn, prefault, map_writable); in FNAME()
Dmmu.c1198 idx = gfn_to_index(gfn, slot->base_gfn, level); in lpage_info_slot()
1485 idx = gfn_to_index(gfn, slot->base_gfn, level); in __gfn_to_rmap()
1769 rmap_head = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), in kvm_mmu_write_protect_pt_masked()
1795 rmap_head = __gfn_to_rmap(slot->base_gfn + gfn_offset + __ffs(mask), in kvm_mmu_clear_dirty_pt_masked()
3338 gfn_t base_gfn = gfn; in __direct_map() local
3351 base_gfn = gfn & ~(KVM_PAGES_PER_HPAGE(it.level) - 1); in __direct_map()
3357 sp = kvm_mmu_get_page(vcpu, base_gfn, it.addr, in __direct_map()
3367 write, level, base_gfn, pfn, prefault, in __direct_map()
5720 end_level, memslot->base_gfn, in slot_handle_level()
5721 memslot->base_gfn + memslot->npages - 1, in slot_handle_level()
[all …]
Dx86.c9840 lpages = gfn_to_index(slot->base_gfn + npages - 1, in kvm_arch_create_memslot()
9841 slot->base_gfn, level) + 1; in kvm_arch_create_memslot()
9857 if (slot->base_gfn & (KVM_PAGES_PER_HPAGE(level) - 1)) in kvm_arch_create_memslot()
9859 if ((slot->base_gfn + npages) & (KVM_PAGES_PER_HPAGE(level) - 1)) in kvm_arch_create_memslot()
9867 if ((slot->base_gfn ^ ugfn) & (KVM_PAGES_PER_HPAGE(level) - 1) || in kvm_arch_create_memslot()
/arch/mips/kvm/
Dmmu.c445 gfn_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() local
446 gfn_t start = base_gfn + __ffs(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked()
447 gfn_t end = base_gfn + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked()
Dmips.c218 kvm_mips_flush_gpa_pt(kvm, slot->base_gfn, in kvm_arch_flush_shadow_memslot()
219 slot->base_gfn + slot->npages - 1); in kvm_arch_flush_shadow_memslot()
259 needs_flush = kvm_mips_mkclean_gpa_pt(kvm, new->base_gfn, in kvm_arch_commit_memory_region()
260 new->base_gfn + new->npages - 1); in kvm_arch_commit_memory_region()
/arch/s390/kvm/
Dkvm-s390.c585 cur_gfn = memslot->base_gfn; in kvm_s390_sync_dirty_log()
586 last_gfn = memslot->base_gfn + memslot->npages; in kvm_s390_sync_dirty_log()
1922 if (gfn >= memslots[slot].base_gfn && in gfn_to_memslot_approx()
1923 gfn < memslots[slot].base_gfn + memslots[slot].npages) in gfn_to_memslot_approx()
1929 if (gfn >= memslots[slot].base_gfn) in gfn_to_memslot_approx()
1938 if (gfn >= memslots[start].base_gfn && in gfn_to_memslot_approx()
1939 gfn < memslots[start].base_gfn + memslots[start].npages) { in gfn_to_memslot_approx()
1974 unsigned long ofs = cur_gfn - ms->base_gfn; in kvm_s390_next_dirty_cmma()
1976 if (ms->base_gfn + ms->npages <= cur_gfn) { in kvm_s390_next_dirty_cmma()
1986 if (cur_gfn < ms->base_gfn) in kvm_s390_next_dirty_cmma()
[all …]
Dpriv.c1179 if (ms && !test_and_set_bit(gfn - ms->base_gfn, kvm_second_dirty_bitmap(ms))) in __do_essa()
/arch/powerpc/include/asm/
Dkvm_book3s_64.h493 return !(memslot->base_gfn & mask) && !(memslot->npages & mask); in slot_is_aligned()
/arch/x86/include/asm/
Dkvm_host.h120 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() argument
124 (base_gfn >> KVM_HPAGE_GFN_SHIFT(level)); in gfn_to_index()