Lines Matching refs:vma
950 struct vm_area_struct *vma = find_vma(current->mm, hva); in stage2_unmap_memslot() local
953 if (!vma || vma->vm_start >= reg_end) in stage2_unmap_memslot()
959 vm_start = max(hva, vma->vm_start); in stage2_unmap_memslot()
960 vm_end = min(reg_end, vma->vm_end); in stage2_unmap_memslot()
962 if (!(vma->vm_flags & VM_PFNMAP)) { in stage2_unmap_memslot()
1606 struct vm_area_struct *vma) in kvm_send_hwpoison_signal() argument
1610 if (is_vm_hugetlb_page(vma)) in kvm_send_hwpoison_signal()
1611 lsb = huge_page_shift(hstate_vma(vma)); in kvm_send_hwpoison_signal()
1686 struct vm_area_struct *vma; in user_mem_abort() local
1703 vma = find_vma_intersection(current->mm, hva, hva + 1); in user_mem_abort()
1704 if (unlikely(!vma)) { in user_mem_abort()
1710 vma_pagesize = vma_kernel_pagesize(vma); in user_mem_abort()
1712 (vma->vm_flags & VM_PFNMAP) || in user_mem_abort()
1727 gfn = (fault_ipa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT; in user_mem_abort()
1750 kvm_send_hwpoison_signal(hva, vma); in user_mem_abort()
2327 struct vm_area_struct *vma = find_vma(current->mm, hva); in kvm_arch_prepare_memory_region() local
2330 if (!vma || vma->vm_start >= reg_end) in kvm_arch_prepare_memory_region()
2337 if (writable && !(vma->vm_flags & VM_WRITE)) { in kvm_arch_prepare_memory_region()
2345 vm_start = max(hva, vma->vm_start); in kvm_arch_prepare_memory_region()
2346 vm_end = min(reg_end, vma->vm_end); in kvm_arch_prepare_memory_region()
2348 if (vma->vm_flags & VM_PFNMAP) { in kvm_arch_prepare_memory_region()
2353 pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT; in kvm_arch_prepare_memory_region()
2354 pa += vm_start - vma->vm_start; in kvm_arch_prepare_memory_region()