• Home
  • Raw
  • Download

Lines Matching refs:flags

80 	unsigned long flags;  in notifiers_decrement()  local
82 spin_lock_irqsave(&hmm->ranges_lock, flags); in notifiers_decrement()
94 spin_unlock_irqrestore(&hmm->ranges_lock, flags); in notifiers_decrement()
103 unsigned long flags; in hmm_invalidate_range_start() local
106 spin_lock_irqsave(&hmm->ranges_lock, flags); in hmm_invalidate_range_start()
114 spin_unlock_irqrestore(&hmm->ranges_lock, flags); in hmm_invalidate_range_start()
220 unsigned int flags; member
226 unsigned int flags = FAULT_FLAG_REMOTE; in hmm_vma_do_fault() local
235 if (hmm_vma_walk->flags & HMM_FAULT_ALLOW_RETRY) in hmm_vma_do_fault()
236 flags |= FAULT_FLAG_ALLOW_RETRY; in hmm_vma_do_fault()
238 flags |= FAULT_FLAG_WRITE; in hmm_vma_do_fault()
240 ret = handle_mm_fault(vma, addr, flags); in hmm_vma_do_fault()
319 if (hmm_vma_walk->flags & HMM_FAULT_SNAPSHOT) in hmm_pte_need_fault()
335 if (!(pfns & range->flags[HMM_PFN_VALID])) in hmm_pte_need_fault()
338 if ((cpu_flags & range->flags[HMM_PFN_DEVICE_PRIVATE])) { in hmm_pte_need_fault()
340 if (pfns & range->flags[HMM_PFN_DEVICE_PRIVATE]) { in hmm_pte_need_fault()
341 *write_fault = pfns & range->flags[HMM_PFN_WRITE]; in hmm_pte_need_fault()
348 *fault = !(cpu_flags & range->flags[HMM_PFN_VALID]); in hmm_pte_need_fault()
350 if ((pfns & range->flags[HMM_PFN_WRITE]) && in hmm_pte_need_fault()
351 !(cpu_flags & range->flags[HMM_PFN_WRITE])) { in hmm_pte_need_fault()
364 if (hmm_vma_walk->flags & HMM_FAULT_SNAPSHOT) { in hmm_range_need_fault()
399 return pmd_write(pmd) ? range->flags[HMM_PFN_VALID] | in pmd_to_hmm_pfn_flags()
400 range->flags[HMM_PFN_WRITE] : in pmd_to_hmm_pfn_flags()
401 range->flags[HMM_PFN_VALID]; in pmd_to_hmm_pfn_flags()
449 return pte_write(pte) ? range->flags[HMM_PFN_VALID] | in pte_to_hmm_pfn_flags()
450 range->flags[HMM_PFN_WRITE] : in pte_to_hmm_pfn_flags()
451 range->flags[HMM_PFN_VALID]; in pte_to_hmm_pfn_flags()
493 cpu_flags = range->flags[HMM_PFN_VALID] | in hmm_vma_handle_pte()
494 range->flags[HMM_PFN_DEVICE_PRIVATE]; in hmm_vma_handle_pte()
496 range->flags[HMM_PFN_WRITE] : 0; in hmm_vma_handle_pte()
651 return pud_write(pud) ? range->flags[HMM_PFN_VALID] | in pud_to_hmm_pfn_flags()
652 range->flags[HMM_PFN_WRITE] : in pud_to_hmm_pfn_flags()
653 range->flags[HMM_PFN_VALID]; in pud_to_hmm_pfn_flags()
794 unsigned long flags; in hmm_range_register() local
809 spin_lock_irqsave(&hmm->ranges_lock, flags); in hmm_range_register()
820 spin_unlock_irqrestore(&hmm->ranges_lock, flags); in hmm_range_register()
836 unsigned long flags; in hmm_range_unregister() local
838 spin_lock_irqsave(&hmm->ranges_lock, flags); in hmm_range_unregister()
840 spin_unlock_irqrestore(&hmm->ranges_lock, flags); in hmm_range_unregister()
890 long hmm_range_fault(struct hmm_range *range, unsigned int flags) in hmm_range_fault() argument
923 hmm_vma_walk.flags = flags; in hmm_range_fault()
965 dma_addr_t *daddrs, unsigned int flags) in hmm_range_dma_map() argument
970 ret = hmm_range_fault(range, flags); in hmm_range_dma_map()
1000 if (range->pfns[i] & range->flags[HMM_PFN_WRITE]) in hmm_range_dma_map()
1027 if (range->pfns[i] & range->flags[HMM_PFN_WRITE]) in hmm_range_dma_map()
1077 if (range->pfns[i] & range->flags[HMM_PFN_WRITE]) { in hmm_range_dma_unmap()