/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
D | radeon_vm.c | 296 struct radeon_bo_va *bo_va; in radeon_vm_bo_find() local 298 list_for_each_entry(bo_va, &bo->va, bo_list) { in radeon_vm_bo_find() 299 if (bo_va->vm == vm) in radeon_vm_bo_find() 300 return bo_va; in radeon_vm_bo_find() 323 struct radeon_bo_va *bo_va; in radeon_vm_bo_add() local 325 bo_va = kzalloc(sizeof(struct radeon_bo_va), GFP_KERNEL); in radeon_vm_bo_add() 326 if (bo_va == NULL) in radeon_vm_bo_add() 329 bo_va->vm = vm; in radeon_vm_bo_add() 330 bo_va->bo = bo; in radeon_vm_bo_add() 331 bo_va->it.start = 0; in radeon_vm_bo_add() [all …]
|
D | radeon_gem.c | 155 struct radeon_bo_va *bo_va; in radeon_gem_object_open() local 168 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open() 169 if (!bo_va) { in radeon_gem_object_open() 170 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open() 172 ++bo_va->ref_count; in radeon_gem_object_open() 186 struct radeon_bo_va *bo_va; in radeon_gem_object_close() local 200 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close() 201 if (bo_va) { in radeon_gem_object_close() 202 if (--bo_va->ref_count == 0) { in radeon_gem_object_close() 203 radeon_vm_bo_rmv(rdev, bo_va); in radeon_gem_object_close() [all …]
|
D | radeon_trace.h | 66 TP_PROTO(struct radeon_bo_va *bo_va), 67 TP_ARGS(bo_va), 75 __entry->soffset = bo_va->it.start; 76 __entry->eoffset = bo_va->it.last + 1; 77 __entry->flags = bo_va->flags;
|
D | radeon_cs.c | 501 struct radeon_bo_va *bo_va; in radeon_bo_vm_update_pte() local 526 bo_va = radeon_vm_bo_find(vm, bo); in radeon_bo_vm_update_pte() 527 if (bo_va == NULL) { in radeon_bo_vm_update_pte() 532 r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem); in radeon_bo_vm_update_pte() 536 radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); in radeon_bo_vm_update_pte()
|
D | radeon.h | 2862 struct radeon_bo_va *bo_va, 2872 struct radeon_bo_va *bo_va, 2876 struct radeon_bo_va *bo_va);
|
/kernel/linux/linux-5.10/drivers/gpu/drm/lima/ |
D | lima_vm.c | 79 struct lima_bo_va *bo_va, *ret = NULL; in lima_vm_bo_find() local 81 list_for_each_entry(bo_va, &bo->va, list) { in lima_vm_bo_find() 82 if (bo_va->vm == vm) { in lima_vm_bo_find() 83 ret = bo_va; in lima_vm_bo_find() 93 struct lima_bo_va *bo_va; in lima_vm_bo_add() local 99 bo_va = lima_vm_bo_find(vm, bo); in lima_vm_bo_add() 100 if (bo_va) { in lima_vm_bo_add() 101 bo_va->ref_count++; in lima_vm_bo_add() 112 bo_va = kzalloc(sizeof(*bo_va), GFP_KERNEL); in lima_vm_bo_add() 113 if (!bo_va) { in lima_vm_bo_add() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vm.c | 1778 int amdgpu_vm_bo_update(struct amdgpu_device *adev, struct amdgpu_bo_va *bo_va, in amdgpu_vm_bo_update() argument 1781 struct amdgpu_bo *bo = bo_va->base.bo; in amdgpu_vm_bo_update() 1782 struct amdgpu_vm *vm = bo_va->base.vm; in amdgpu_vm_bo_update() 1802 if (obj->import_attach && bo_va->is_xgmi) { in amdgpu_vm_bo_update() 1833 last_update = &bo_va->last_pt_update; in amdgpu_vm_bo_update() 1835 if (!clear && bo_va->base.moved) { in amdgpu_vm_bo_update() 1836 bo_va->base.moved = false; in amdgpu_vm_bo_update() 1837 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() 1839 } else if (bo_va->cleared != clear) { in amdgpu_vm_bo_update() 1840 list_splice_init(&bo_va->valids, &bo_va->invalids); in amdgpu_vm_bo_update() [all …]
|
D | amdgpu_csa.c | 66 struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va, in amdgpu_map_static_csa() argument 89 *bo_va = amdgpu_vm_bo_add(adev, vm, bo); in amdgpu_map_static_csa() 90 if (!*bo_va) { in amdgpu_map_static_csa() 96 r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, size, in amdgpu_map_static_csa() 102 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa()
|
D | amdgpu_gem.c | 115 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_open() local 131 bo_va = amdgpu_vm_bo_find(vm, abo); in amdgpu_gem_object_open() 132 if (!bo_va) { in amdgpu_gem_object_open() 133 bo_va = amdgpu_vm_bo_add(adev, vm, abo); in amdgpu_gem_object_open() 135 ++bo_va->ref_count; in amdgpu_gem_object_open() 154 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_close() local 172 bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_gem_object_close() 173 if (!bo_va || --bo_va->ref_count) in amdgpu_gem_object_close() 176 amdgpu_vm_bo_rmv(adev, bo_va); in amdgpu_gem_object_close() 528 struct amdgpu_bo_va *bo_va, in amdgpu_gem_va_update_vm() argument [all …]
|
D | amdgpu_vm.h | 396 struct amdgpu_bo_va *bo_va, 408 struct amdgpu_bo_va *bo_va, 412 struct amdgpu_bo_va *bo_va, 416 struct amdgpu_bo_va *bo_va, 425 struct amdgpu_bo_va *bo_va);
|
D | amdgpu_amdkfd_gpuvm.c | 76 if (entry->bo_va->base.vm == avm) in check_if_add_bo_to_vm() 462 bo_va_entry->bo_va = amdgpu_vm_bo_add(adev, vm, bo); in add_bo_to_vm() 463 if (!bo_va_entry->bo_va) { in add_bo_to_vm() 488 amdgpu_vm_bo_rmv(adev, bo_va_entry->bo_va); in add_bo_to_vm() 501 amdgpu_vm_bo_rmv(adev, entry->bo_va); in remove_bo_from_vm() 691 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms() 713 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms() 718 amdgpu_vm_get_pd_bo(entry->bo_va->base.vm, &ctx->list, in reserve_bo_and_cond_vms() 770 struct amdgpu_bo_va *bo_va = entry->bo_va; in unmap_bo_from_gpuvm() local 771 struct amdgpu_vm *vm = bo_va->base.vm; in unmap_bo_from_gpuvm() [all …]
|
D | amdgpu_trace.h | 244 TP_PROTO(struct amdgpu_bo_va *bo_va, 246 TP_ARGS(bo_va, mapping), 256 __entry->bo = bo_va ? bo_va->base.bo : NULL; 268 TP_PROTO(struct amdgpu_bo_va *bo_va, 270 TP_ARGS(bo_va, mapping), 280 __entry->bo = bo_va ? bo_va->base.bo : NULL;
|
D | amdgpu_cs.c | 625 e->bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_cs_parser_bos() 727 struct amdgpu_bo_va *bo_va; in amdgpu_cs_vm_handling() local 809 bo_va = fpriv->csa_va; in amdgpu_cs_vm_handling() 810 BUG_ON(!bo_va); in amdgpu_cs_vm_handling() 811 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_cs_vm_handling() 815 r = amdgpu_sync_vm_fence(&p->job->sync, bo_va->last_pt_update); in amdgpu_cs_vm_handling() 826 bo_va = e->bo_va; in amdgpu_cs_vm_handling() 827 if (bo_va == NULL) in amdgpu_cs_vm_handling() 830 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_cs_vm_handling() 834 r = amdgpu_sync_vm_fence(&p->job->sync, bo_va->last_pt_update); in amdgpu_cs_vm_handling() [all …]
|
D | amdgpu_csa.h | 35 struct amdgpu_bo *bo, struct amdgpu_bo_va **bo_va,
|
D | amdgpu_bo_list.h | 36 struct amdgpu_bo_va *bo_va; member
|
D | amdgpu_object.h | 53 struct amdgpu_bo_va *bo_va; member
|
D | amdgpu_amdkfd.h | 43 struct amdgpu_bo_va *bo_va; member
|
D | gmc_v9_0.c | 1074 mapping->bo_va->is_xgmi) in gmc_v9_0_get_vm_pte()
|