Lines Matching refs:range
355 struct hmm_range range = {}; in ib_umem_odp_map_dma_and_lock() local
378 range.notifier = &umem_odp->notifier; in ib_umem_odp_map_dma_and_lock()
379 range.start = ALIGN_DOWN(user_virt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock()
380 range.end = ALIGN(user_virt + bcnt, 1UL << page_shift); in ib_umem_odp_map_dma_and_lock()
381 pfn_start_idx = (range.start - ib_umem_start(umem_odp)) >> PAGE_SHIFT; in ib_umem_odp_map_dma_and_lock()
382 num_pfns = (range.end - range.start) >> PAGE_SHIFT; in ib_umem_odp_map_dma_and_lock()
384 range.default_flags = HMM_PFN_REQ_FAULT; in ib_umem_odp_map_dma_and_lock()
387 range.default_flags |= HMM_PFN_REQ_WRITE; in ib_umem_odp_map_dma_and_lock()
390 range.hmm_pfns = &(umem_odp->pfn_list[pfn_start_idx]); in ib_umem_odp_map_dma_and_lock()
394 current_seq = range.notifier_seq = in ib_umem_odp_map_dma_and_lock()
398 ret = hmm_range_fault(&range); in ib_umem_odp_map_dma_and_lock()
406 start_idx = (range.start - ib_umem_start(umem_odp)) >> page_shift; in ib_umem_odp_map_dma_and_lock()
423 WARN_ON(range.hmm_pfns[pfn_index] & HMM_PFN_ERROR); in ib_umem_odp_map_dma_and_lock()
424 WARN_ON(!(range.hmm_pfns[pfn_index] & HMM_PFN_VALID)); in ib_umem_odp_map_dma_and_lock()
426 if (!(range.hmm_pfns[pfn_index] & HMM_PFN_VALID)) { in ib_umem_odp_map_dma_and_lock()
431 if (range.hmm_pfns[pfn_index] & HMM_PFN_WRITE) in ib_umem_odp_map_dma_and_lock()
435 hmm_order = hmm_pfn_to_map_order(range.hmm_pfns[pfn_index]); in ib_umem_odp_map_dma_and_lock()
448 umem_odp, dma_index, hmm_pfn_to_page(range.hmm_pfns[pfn_index]), in ib_umem_odp_map_dma_and_lock()