Home
last modified time | relevance | path

Searched refs:hmm_pfns (Results 1 – 7 of 7) sorted by relevance

/kernel/linux/linux-5.10/mm/
Dhmm.c46 range->hmm_pfns[i] = cpu_flags; in hmm_pfns_fill()
120 const unsigned long hmm_pfns[], unsigned long npages, in hmm_range_need_fault() argument
137 required_fault |= hmm_pte_need_fault(hmm_vma_walk, hmm_pfns[i], in hmm_range_need_fault()
152 unsigned long *hmm_pfns; in hmm_vma_walk_hole() local
156 hmm_pfns = &range->hmm_pfns[i]; in hmm_vma_walk_hole()
158 hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, 0); in hmm_vma_walk_hole()
186 unsigned long end, unsigned long hmm_pfns[], in hmm_vma_handle_pmd() argument
198 hmm_range_need_fault(hmm_vma_walk, hmm_pfns, npages, cpu_flags); in hmm_vma_handle_pmd()
204 hmm_pfns[i] = pfn | cpu_flags; in hmm_vma_handle_pmd()
210 unsigned long end, unsigned long hmm_pfns[], pmd_t pmd);
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/core/
Dumem_odp.c396 range.hmm_pfns = &(umem_odp->pfn_list[pfn_start_idx]); in ib_umem_odp_map_dma_and_lock()
429 WARN_ON(range.hmm_pfns[pfn_index] & HMM_PFN_ERROR); in ib_umem_odp_map_dma_and_lock()
430 WARN_ON(!(range.hmm_pfns[pfn_index] & HMM_PFN_VALID)); in ib_umem_odp_map_dma_and_lock()
432 if (!(range.hmm_pfns[pfn_index] & HMM_PFN_VALID)) { in ib_umem_odp_map_dma_and_lock()
437 if (range.hmm_pfns[pfn_index] & HMM_PFN_WRITE) in ib_umem_odp_map_dma_and_lock()
441 hmm_order = hmm_pfn_to_map_order(range.hmm_pfns[pfn_index]); in ib_umem_odp_map_dma_and_lock()
454 umem_odp, dma_index, hmm_pfn_to_page(range.hmm_pfns[pfn_index]), in ib_umem_odp_map_dma_and_lock()
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/
Dnouveau_svm.c527 if (!(range->hmm_pfns[0] & HMM_PFN_VALID)) { in nouveau_hmm_convert_pfn()
532 page = hmm_pfn_to_page(range->hmm_pfns[0]); in nouveau_hmm_convert_pfn()
541 if (hmm_pfn_to_map_order(range->hmm_pfns[0])) { in nouveau_hmm_convert_pfn()
544 args->p.page = hmm_pfn_to_map_order(range->hmm_pfns[0]) + in nouveau_hmm_convert_pfn()
558 if (range->hmm_pfns[0] & HMM_PFN_WRITE) in nouveau_hmm_convert_pfn()
571 unsigned long hmm_pfns[1]; in nouveau_range_fault() local
577 .hmm_pfns = hmm_pfns, in nouveau_range_fault()
/kernel/linux/linux-5.10/include/linux/
Dhmm.h101 unsigned long *hmm_pfns; member
/kernel/linux/linux-5.10/lib/
Dtest_hmm.c162 unsigned long *pfns = range->hmm_pfns; in dmirror_do_fault()
288 .hmm_pfns = pfns, in dmirror_fault()
851 dmirror_mkentry(dmirror, range, perm + i, range->hmm_pfns[i]); in dmirror_range_snapshot()
871 .hmm_pfns = pfns, in dmirror_snapshot()
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/
Damdgpu_ttm.c871 range->hmm_pfns = kvmalloc_array(ttm->num_pages, in amdgpu_ttm_tt_get_user_pages()
872 sizeof(*range->hmm_pfns), GFP_KERNEL); in amdgpu_ttm_tt_get_user_pages()
873 if (unlikely(!range->hmm_pfns)) { in amdgpu_ttm_tt_get_user_pages()
914 pages[i] = hmm_pfn_to_page(range->hmm_pfns[i]); in amdgpu_ttm_tt_get_user_pages()
924 kvfree(range->hmm_pfns); in amdgpu_ttm_tt_get_user_pages()
949 WARN_ONCE(!gtt->range || !gtt->range->hmm_pfns, in amdgpu_ttm_tt_get_user_pages_done()
959 kvfree(gtt->range->hmm_pfns); in amdgpu_ttm_tt_get_user_pages_done()
1050 hmm_pfn_to_page(gtt->range->hmm_pfns[i])) in amdgpu_ttm_tt_unpin_userptr()
/kernel/linux/linux-5.10/Documentation/vm/
Dhmm.rst187 range.hmm_pfns = ...;