/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_mn.c | 63 static bool amdgpu_mn_invalidate_gfx(struct mmu_interval_notifier *mni, in amdgpu_mn_invalidate_gfx() argument 67 struct amdgpu_bo *bo = container_of(mni, struct amdgpu_bo, notifier); in amdgpu_mn_invalidate_gfx() 76 mmu_interval_set_seq(mni, cur_seq); in amdgpu_mn_invalidate_gfx() 100 static bool amdgpu_mn_invalidate_hsa(struct mmu_interval_notifier *mni, in amdgpu_mn_invalidate_hsa() argument 104 struct amdgpu_bo *bo = container_of(mni, struct amdgpu_bo, notifier); in amdgpu_mn_invalidate_hsa() 112 mmu_interval_set_seq(mni, cur_seq); in amdgpu_mn_invalidate_hsa()
|
/drivers/infiniband/hw/hfi1/ |
D | user_exp_rcv.c | 23 static bool tid_rb_invalidate(struct mmu_interval_notifier *mni, 26 static bool tid_cover_invalidate(struct mmu_interval_notifier *mni, 897 static bool tid_rb_invalidate(struct mmu_interval_notifier *mni, in tid_rb_invalidate() argument 902 container_of(mni, struct tid_rb_node, notifier); in tid_rb_invalidate() 948 static bool tid_cover_invalidate(struct mmu_interval_notifier *mni, in tid_cover_invalidate() argument 953 container_of(mni, struct tid_user_buf, notifier); in tid_cover_invalidate() 958 mmu_interval_set_seq(mni, cur_seq); in tid_cover_invalidate()
|
/drivers/gpu/drm/i915/gem/ |
D | i915_gem_userptr.c | 60 static bool i915_gem_userptr_invalidate(struct mmu_interval_notifier *mni, in i915_gem_userptr_invalidate() argument 64 struct drm_i915_gem_object *obj = container_of(mni, struct drm_i915_gem_object, userptr.notifier); in i915_gem_userptr_invalidate() 73 mmu_interval_set_seq(mni, cur_seq); in i915_gem_userptr_invalidate()
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_svm.c | 62 svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni, 2396 svm_range_cpu_invalidate_pagetables(struct mmu_interval_notifier *mni, in svm_range_cpu_invalidate_pagetables() argument 2406 if (!mmget_not_zero(mni->mm)) in svm_range_cpu_invalidate_pagetables() 2409 start = mni->interval_tree.start; in svm_range_cpu_invalidate_pagetables() 2410 last = mni->interval_tree.last; in svm_range_cpu_invalidate_pagetables() 2416 mni->interval_tree.start >> PAGE_SHIFT, in svm_range_cpu_invalidate_pagetables() 2417 mni->interval_tree.last >> PAGE_SHIFT, range->event); in svm_range_cpu_invalidate_pagetables() 2419 prange = container_of(mni, struct svm_range, notifier); in svm_range_cpu_invalidate_pagetables() 2422 mmu_interval_set_seq(mni, cur_seq); in svm_range_cpu_invalidate_pagetables() 2426 svm_range_unmap_from_cpu(mni->mm, prange, start, last); in svm_range_cpu_invalidate_pagetables() [all …]
|
/drivers/gpu/drm/nouveau/ |
D | nouveau_svm.c | 508 static bool nouveau_svm_range_invalidate(struct mmu_interval_notifier *mni, in nouveau_svm_range_invalidate() argument 513 container_of(mni, struct svm_notifier, notifier); in nouveau_svm_range_invalidate() 530 mmu_interval_set_seq(mni, cur_seq); in nouveau_svm_range_invalidate()
|
/drivers/video/fbdev/sis/ |
D | sis_main.c | 607 switch(sisbios_mode[mode_idx].mode_no[ivideo->mni]) { in sisfb_verify_rate() 631 sisbios_mode[mode_idx].mode_no[ivideo->mni], in sisfb_verify_rate() 1265 ivideo->mode_no = sisbios_mode[ivideo->sisfb_mode_idx].mode_no[ivideo->mni]; in sisfb_do_set_var() 1286 ivideo->mode_no = sisbios_mode[ivideo->sisfb_mode_idx].mode_no[ivideo->mni]; in sisfb_do_set_var() 1563 if(ivideo->sisfb_lastrates[sisbios_mode[search_idx].mode_no[ivideo->mni]]) { in sisfb_check_var() 1565 ivideo->sisfb_lastrates[sisbios_mode[search_idx].mode_no[ivideo->mni]]; in sisfb_check_var() 1591 sisbios_mode[search_idx].mode_no[ivideo->mni], in sisfb_check_var() 1594 sisbios_mode[search_idx].mode_no[ivideo->mni], in sisfb_check_var() 5915 ivideo->mni = chipinfo->mni; in sisfb_probe() 6206 ((sisbios_mode[ivideo->sisfb_mode_idx].mode_no[ivideo->mni]) != 0xFF)) { in sisfb_probe() [all …]
|
D | sis_main.h | 84 int mni; member
|
D | sis.h | 510 int mni; /* Mode number index */ member
|
/drivers/infiniband/hw/mlx5/ |
D | odp.c | 226 static bool mlx5_ib_invalidate_range(struct mmu_interval_notifier *mni, in mlx5_ib_invalidate_range() argument 231 container_of(mni, struct ib_umem_odp, notifier); in mlx5_ib_invalidate_range() 246 mmu_interval_set_seq(mni, cur_seq); in mlx5_ib_invalidate_range()
|