Lines Matching refs:pfn
164 kvm_pfn_t pfn; in kvmppc_map_magic() local
166 pfn = (kvm_pfn_t)virt_to_phys((void *)shared_page) >> PAGE_SHIFT; in kvmppc_map_magic()
167 get_page(pfn_to_page(pfn)); in kvmppc_map_magic()
175 magic.mas7_3 = ((u64)pfn << PAGE_SHIFT) | in kvmppc_map_magic()
247 kvm_pfn_t pfn, unsigned int wimg) in kvmppc_e500_ref_setup() argument
249 ref->pfn = pfn; in kvmppc_e500_ref_setup()
256 kvm_set_pfn_accessed(pfn); in kvmppc_e500_ref_setup()
259 kvm_set_pfn_dirty(pfn); in kvmppc_e500_ref_setup()
266 trace_kvm_booke206_ref_release(ref->pfn, ref->flags); in kvmppc_e500_ref_release()
310 kvm_pfn_t pfn = ref->pfn; in kvmppc_e500_setup_stlbe() local
318 stlbe->mas7_3 = ((u64)pfn << PAGE_SHIFT) | in kvmppc_e500_setup_stlbe()
328 unsigned long pfn = 0; /* silence GCC warning */ in kvmppc_e500_shadow_map() local
379 pfn = start + ((hva - vma->vm_start) >> PAGE_SHIFT); in kvmppc_e500_shadow_map()
381 slot_start = pfn - (gfn - slot->base_gfn); in kvmppc_e500_shadow_map()
412 if (gfn_start + pfn - gfn < start) in kvmppc_e500_shadow_map()
414 if (gfn_end + pfn - gfn > end) in kvmppc_e500_shadow_map()
417 (pfn & (tsize_pages - 1))) in kvmppc_e500_shadow_map()
421 pfn &= ~(tsize_pages - 1); in kvmppc_e500_shadow_map()
449 pfn = gfn_to_pfn_memslot(slot, gfn); in kvmppc_e500_shadow_map()
450 if (is_error_noslot_pfn(pfn)) { in kvmppc_e500_shadow_map()
458 pfn &= ~(tsize_pages - 1); in kvmppc_e500_shadow_map()
488 __func__, (long)gfn, pfn); in kvmppc_e500_shadow_map()
493 kvmppc_e500_ref_setup(ref, gtlbe, pfn, wimg); in kvmppc_e500_shadow_map()
499 kvmppc_mmu_flush_icache(pfn); in kvmppc_e500_shadow_map()
505 kvm_release_pfn_clean(pfn); in kvmppc_e500_shadow_map()
630 hfn_t pfn; in kvmppc_load_last_inst() local
697 pfn = addr >> PAGE_SHIFT; in kvmppc_load_last_inst()
700 if (unlikely(!page_is_ram(pfn))) { in kvmppc_load_last_inst()
707 page = pfn_to_page(pfn); in kvmppc_load_last_inst()