• Home
  • Raw
  • Download

Lines Matching refs:hmm_vma_walk

29 struct hmm_vma_walk {  struct
64 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_fault() local
69 hmm_vma_walk->last = addr; in hmm_vma_fault()
84 static unsigned int hmm_pte_need_fault(const struct hmm_vma_walk *hmm_vma_walk, in hmm_pte_need_fault() argument
88 struct hmm_range *range = hmm_vma_walk->range; in hmm_pte_need_fault()
119 hmm_range_need_fault(const struct hmm_vma_walk *hmm_vma_walk, in hmm_range_need_fault() argument
123 struct hmm_range *range = hmm_vma_walk->range; in hmm_range_need_fault()
137 required_fault |= hmm_pte_need_fault(hmm_vma_walk, hmm_pfns[i], in hmm_range_need_fault()
148 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_walk_hole() local
149 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_hole()
158 hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, 0); in hmm_vma_walk_hole()
189 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_handle_pmd() local
190 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_pmd()
198 hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, cpu_flags); in hmm_vma_handle_pmd()
233 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_handle_pte() local
234 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_handle_pte()
242 hmm_pte_need_fault(hmm_vma_walk, pfn_req_flags, 0); in hmm_vma_handle_pte()
266 hmm_pte_need_fault(hmm_vma_walk, pfn_req_flags, 0); in hmm_vma_handle_pte()
277 hmm_vma_walk->last = addr; in hmm_vma_handle_pte()
289 hmm_pte_need_fault(hmm_vma_walk, pfn_req_flags, cpu_flags); in hmm_vma_handle_pte()
302 if (hmm_pte_need_fault(hmm_vma_walk, pfn_req_flags, 0)) { in hmm_vma_handle_pte()
324 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_walk_pmd() local
325 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_pmd()
339 if (hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, 0)) { in hmm_vma_walk_pmd()
340 hmm_vma_walk->last = addr; in hmm_vma_walk_pmd()
348 if (hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, 0)) in hmm_vma_walk_pmd()
378 if (hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, 0)) in hmm_vma_walk_pmd()
412 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_walk_pud() local
413 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_pud()
447 required_fault = hmm_range_need_fault(hmm_vma_walk, hmm_pfns, in hmm_vma_walk_pud()
477 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_walk_hugetlb_entry() local
478 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_hugetlb_entry()
494 hmm_pte_need_fault(hmm_vma_walk, pfn_req_flags, cpu_flags); in hmm_vma_walk_hugetlb_entry()
514 struct hmm_vma_walk *hmm_vma_walk = walk->private; in hmm_vma_walk_test() local
515 struct hmm_range *range = hmm_vma_walk->range; in hmm_vma_walk_test()
533 if (hmm_range_need_fault(hmm_vma_walk, in hmm_vma_walk_test()
574 struct hmm_vma_walk hmm_vma_walk = { in hmm_range_fault() local
588 ret = walk_page_range(mm, hmm_vma_walk.last, range->end, in hmm_range_fault()
589 &hmm_walk_ops, &hmm_vma_walk); in hmm_range_fault()