• Home
  • Raw
  • Download

Lines Matching refs:end

109 		if (nrange->end < range->start || nrange->start >= range->end)  in hmm_invalidate_range_start()
256 unsigned long end, in hmm_pfns_bad() argument
265 for (; addr < end; addr += PAGE_SIZE, i++) in hmm_pfns_bad()
283 static int hmm_vma_walk_hole_(unsigned long addr, unsigned long end, in hmm_vma_walk_hole_() argument
298 for (; addr < end; addr += PAGE_SIZE, i++) { in hmm_vma_walk_hole_()
378 static int hmm_vma_walk_hole(unsigned long addr, unsigned long end, in hmm_vma_walk_hole() argument
388 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_hole()
392 return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); in hmm_vma_walk_hole()
406 unsigned long end, uint64_t *pfns, pmd_t pmd) in hmm_vma_handle_pmd() argument
414 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_handle_pmd()
420 return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); in hmm_vma_handle_pmd()
423 for (i = 0; addr < end; addr += PAGE_SIZE, i++, pfn++) { in hmm_vma_handle_pmd()
436 hmm_vma_walk->last = end; in hmm_vma_handle_pmd()
442 unsigned long end, uint64_t *pfns, pmd_t pmd);
455 unsigned long end, pmd_t *pmdp, pte_t *ptep, in hmm_vma_handle_pte() argument
549 return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); in hmm_vma_handle_pte()
554 unsigned long end, in hmm_vma_walk_pmd() argument
567 return hmm_vma_walk_hole(start, end, walk); in hmm_vma_walk_pmd()
575 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_pmd()
587 return hmm_pfns_bad(start, end, walk); in hmm_vma_walk_pmd()
605 return hmm_vma_handle_pmd(walk, addr, end, &pfns[i], pmd); in hmm_vma_walk_pmd()
615 return hmm_pfns_bad(start, end, walk); in hmm_vma_walk_pmd()
619 for (; addr < end; addr += PAGE_SIZE, ptep++, i++) { in hmm_vma_walk_pmd()
622 r = hmm_vma_handle_pte(walk, addr, end, pmdp, ptep, &pfns[i]); in hmm_vma_walk_pmd()
656 static int hmm_vma_walk_pud(pud_t *pudp, unsigned long start, unsigned long end, in hmm_vma_walk_pud() argument
669 return hmm_vma_walk_hole(start, end, walk); in hmm_vma_walk_pud()
677 return hmm_vma_walk_hole(start, end, walk); in hmm_vma_walk_pud()
680 npages = (end - addr) >> PAGE_SHIFT; in hmm_vma_walk_pud()
687 return hmm_vma_walk_hole_(addr, end, fault, in hmm_vma_walk_pud()
703 hmm_vma_walk->last = end; in hmm_vma_walk_pud()
713 next = pmd_addr_end(addr, end); in hmm_vma_walk_pud()
717 } while (pmdp++, addr = next, addr != end); in hmm_vma_walk_pud()
727 unsigned long start, unsigned long end, in hmm_vma_walk_hugetlb_entry() argument
756 for (; addr < end; addr += PAGE_SIZE, i++, pfn++) in hmm_vma_walk_hugetlb_entry()
759 hmm_vma_walk->last = end; in hmm_vma_walk_hugetlb_entry()
765 return hmm_vma_walk_hole_(addr, end, fault, write_fault, walk); in hmm_vma_walk_hugetlb_entry()
776 unsigned long end) in hmm_pfns_clear() argument
778 for (; addr < end; addr += PAGE_SIZE, pfns++) in hmm_pfns_clear()
799 if ((range->start & (PAGE_SIZE - 1)) || (range->end & (PAGE_SIZE - 1))) in hmm_range_register()
801 if (range->start >= range->end) in hmm_range_register()
893 unsigned long start = range->start, end; in hmm_range_fault() local
917 range->start, range->end); in hmm_range_fault()
925 end = min(range->end, vma->vm_end); in hmm_range_fault()
927 walk_page_range(vma->vm_mm, start, end, &hmm_walk_ops, in hmm_range_fault()
931 ret = walk_page_range(vma->vm_mm, start, end, in hmm_range_fault()
943 hmm_vma_walk.last, range->end); in hmm_range_fault()
946 start = end; in hmm_range_fault()
948 } while (start < range->end); in hmm_range_fault()
974 npages = (range->end - range->start) >> PAGE_SHIFT; in hmm_range_dma_map()
1060 if (range->end <= range->start) in hmm_range_dma_unmap()
1067 npages = (range->end - range->start) >> PAGE_SHIFT; in hmm_range_dma_unmap()