/drivers/infiniband/core/ |
D | umem.c | 88 struct vm_area_struct **vma_list; in ib_umem_get() local 154 vma_list = (struct vm_area_struct **) __get_free_page(GFP_KERNEL); in ib_umem_get() 155 if (!vma_list) in ib_umem_get() 191 gup_flags, page_list, vma_list); in ib_umem_get() 201 if (vma_list && !is_vm_hugetlb_page(vma_list[i])) in ib_umem_get() 234 if (vma_list) in ib_umem_get() 235 free_page((unsigned long) vma_list); in ib_umem_get()
|
/drivers/staging/android/ion/ |
D | ion.c | 950 struct ion_vma_list *vma_list; in ion_buffer_sync_for_device() local 967 list_for_each_entry(vma_list, &buffer->vmas, list) { in ion_buffer_sync_for_device() 968 struct vm_area_struct *vma = vma_list->vma; in ion_buffer_sync_for_device() 998 struct ion_vma_list *vma_list; in ion_vm_open() local 1000 vma_list = kmalloc(sizeof(struct ion_vma_list), GFP_KERNEL); in ion_vm_open() 1001 if (!vma_list) in ion_vm_open() 1003 vma_list->vma = vma; in ion_vm_open() 1005 list_add(&vma_list->list, &buffer->vmas); in ion_vm_open() 1012 struct ion_vma_list *vma_list, *tmp; in ion_vm_close() local 1015 list_for_each_entry_safe(vma_list, tmp, &buffer->vmas, list) { in ion_vm_close() [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_gem_shrinker.c | 136 &obj->vma_list, vma_link) in i915_gem_shrink() 196 list_for_each_entry(vma, &obj->vma_list, vma_link) { in num_vma_bound()
|
D | i915_gem.c | 275 list_for_each_entry_safe(vma, next, &obj->vma_list, vma_link) in drop_pages() 2542 list_for_each_entry(vma, &obj->vma_list, vma_link) { in i915_gem_object_retire__read() 3413 if (list_empty(&obj->vma_list)) in __i915_vma_unbind() 3835 list_for_each_entry_safe(vma, next, &obj->vma_list, vma_link) { in i915_gem_object_set_cache_level() 3898 list_for_each_entry(vma, &obj->vma_list, vma_link) { in i915_gem_object_set_cache_level() 3908 list_for_each_entry(vma, &obj->vma_list, vma_link) in i915_gem_object_set_cache_level() 4469 INIT_LIST_HEAD(&obj->vma_list); in i915_gem_object_init() 4573 list_for_each_entry_safe(vma, next, &obj->vma_list, vma_link) { in i915_gem_free_object() 4630 list_for_each_entry(vma, &obj->vma_list, vma_link) { in i915_gem_obj_to_vma() 4649 list_for_each_entry(vma, &obj->vma_list, vma_link) in i915_gem_obj_to_ggtt_view() [all …]
|
D | i915_gpu_error.c | 747 list_for_each_entry(vma, &obj->vma_list, vma_link) in capture_pinned_bo() 1101 list_for_each_entry(vma, &obj->vma_list, vma_link) in i915_gem_capture_vm()
|
D | i915_gem_userptr.c | 84 list_for_each_entry_safe(vma, tmp, &obj->vma_list, vma_link) { in __cancel_userptr__worker()
|
D | i915_debugfs.c | 120 list_for_each_entry(vma, &obj->vma_list, vma_link) { in i915_gem_obj_total_ggtt_size() 158 list_for_each_entry(vma, &obj->vma_list, vma_link) { in describe_obj() 167 list_for_each_entry(vma, &obj->vma_list, vma_link) { in describe_obj() 345 list_for_each_entry(vma, &obj->vma_list, vma_link) { in per_file_stats()
|
D | i915_gem_gtt.c | 3182 list_for_each_entry(vma, &obj->vma_list, vma_link) { in i915_gem_restore_gtt_mappings() 3247 list_add_tail(&vma->vma_link, &obj->vma_list); in __i915_gem_vma_create()
|
D | i915_drv.h | 2052 struct list_head vma_list; member
|
/drivers/gpu/drm/nouveau/ |
D | nouveau_bo.h | 26 struct list_head vma_list; member
|
D | nouveau_bo.c | 207 INIT_LIST_HEAD(&nvbo->vma_list); in nouveau_bo_new() 1236 list_for_each_entry(vma, &nvbo->vma_list, head) { in nouveau_bo_move_ntfy() 1620 list_for_each_entry(vma, &nvbo->vma_list, head) { in nouveau_bo_vma_find() 1645 list_add_tail(&vma->head, &nvbo->vma_list); in nouveau_bo_vma_add()
|
/drivers/misc/mic/scif/ |
D | scif_mmap.c | 83 list_for_each(item, &ep->rma_info.vma_list) { in __scif_zap_mmaps() 181 list_add_tail(&info->list, &ep->rma_info.vma_list); in scif_insert_vma() 194 list_for_each(item, &ep->rma_info.vma_list) { in scif_delete_vma()
|
D | scif_rma.h | 113 struct list_head vma_list; member
|
D | scif_rma.c | 53 INIT_LIST_HEAD(&rma->vma_list); in scif_rma_ep_init()
|