/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
D | radeon_object.c | 101 void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain) in radeon_ttm_placement_from_domain() argument 105 rbo->placement.placement = rbo->placements; in radeon_ttm_placement_from_domain() 106 rbo->placement.busy_placement = rbo->placements; in radeon_ttm_placement_from_domain() 111 if ((rbo->flags & RADEON_GEM_NO_CPU_ACCESS) && in radeon_ttm_placement_from_domain() 112 rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size) { in radeon_ttm_placement_from_domain() 113 rbo->placements[c].fpfn = in radeon_ttm_placement_from_domain() 114 rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT; in radeon_ttm_placement_from_domain() 115 rbo->placements[c].mem_type = TTM_PL_VRAM; in radeon_ttm_placement_from_domain() 116 rbo->placements[c++].flags = TTM_PL_FLAG_WC | in radeon_ttm_placement_from_domain() 120 rbo->placements[c].fpfn = 0; in radeon_ttm_placement_from_domain() [all …]
|
D | radeon_fb.c | 113 struct radeon_bo *rbo = gem_to_radeon_bo(gobj); in radeonfb_destroy_pinned_object() local 116 ret = radeon_bo_reserve(rbo, false); in radeonfb_destroy_pinned_object() 118 radeon_bo_kunmap(rbo); in radeonfb_destroy_pinned_object() 119 radeon_bo_unpin(rbo); in radeonfb_destroy_pinned_object() 120 radeon_bo_unreserve(rbo); in radeonfb_destroy_pinned_object() 132 struct radeon_bo *rbo = NULL; in radeonfb_create_pinned_object() local 158 rbo = gem_to_radeon_bo(gobj); in radeonfb_create_pinned_object() 176 ret = radeon_bo_set_tiling_flags(rbo, in radeonfb_create_pinned_object() 184 ret = radeon_bo_reserve(rbo, false); in radeonfb_create_pinned_object() 188 ret = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRAM, in radeonfb_create_pinned_object() [all …]
|
D | radeon_gem.c | 151 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_open() local 152 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_open() 163 r = radeon_bo_reserve(rbo, false); in radeon_gem_object_open() 168 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open() 170 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open() 174 radeon_bo_unreserve(rbo); in radeon_gem_object_open() 182 struct radeon_bo *rbo = gem_to_radeon_bo(obj); in radeon_gem_object_close() local 183 struct radeon_device *rdev = rbo->rdev; in radeon_gem_object_close() 194 r = radeon_bo_reserve(rbo, true); in radeon_gem_object_close() 200 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close() [all …]
|
D | radeon_ttm.c | 95 struct radeon_bo *rbo; in radeon_evict_flags() local 104 rbo = container_of(bo, struct radeon_bo, tbo); in radeon_evict_flags() 107 if (rbo->rdev->ring[radeon_copy_ring_index(rbo->rdev)].ready == false) in radeon_evict_flags() 108 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_CPU); in radeon_evict_flags() 109 else if (rbo->rdev->mc.visible_vram_size < rbo->rdev->mc.real_vram_size && in radeon_evict_flags() 110 bo->mem.start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) { in radeon_evict_flags() 111 unsigned fpfn = rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT; in radeon_evict_flags() 119 radeon_ttm_placement_from_domain(rbo, RADEON_GEM_DOMAIN_VRAM | in radeon_evict_flags() 121 rbo->placement.num_busy_placement = 0; in radeon_evict_flags() 122 for (i = 0; i < rbo->placement.num_placement; i++) { in radeon_evict_flags() [all …]
|
D | radeon_legacy_crtc.c | 382 struct radeon_bo *rbo; in radeon_crtc_do_set_base() local 425 rbo = gem_to_radeon_bo(obj); in radeon_crtc_do_set_base() 427 r = radeon_bo_reserve(rbo, false); in radeon_crtc_do_set_base() 431 r = radeon_bo_pin_restricted(rbo, RADEON_GEM_DOMAIN_VRAM, 1 << 27, in radeon_crtc_do_set_base() 434 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 454 nsize = radeon_bo_size(rbo); in radeon_crtc_do_set_base() 464 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in radeon_crtc_do_set_base() 465 radeon_bo_unreserve(rbo); in radeon_crtc_do_set_base() 559 rbo = gem_to_radeon_bo(fb->obj[0]); in radeon_crtc_do_set_base() 560 r = radeon_bo_reserve(rbo, false); in radeon_crtc_do_set_base() [all …]
|
D | atombios_crtc.c | 1152 struct radeon_bo *rbo; in dce4_crtc_do_set_base() local 1177 rbo = gem_to_radeon_bo(obj); in dce4_crtc_do_set_base() 1178 r = radeon_bo_reserve(rbo, false); in dce4_crtc_do_set_base() 1183 fb_location = radeon_bo_gpu_offset(rbo); in dce4_crtc_do_set_base() 1185 r = radeon_bo_pin(rbo, RADEON_GEM_DOMAIN_VRAM, &fb_location); in dce4_crtc_do_set_base() 1187 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1192 radeon_bo_get_tiling_flags(rbo, &tiling_flags, NULL); in dce4_crtc_do_set_base() 1193 radeon_bo_unreserve(rbo); in dce4_crtc_do_set_base() 1451 rbo = gem_to_radeon_bo(fb->obj[0]); in dce4_crtc_do_set_base() 1452 r = radeon_bo_reserve(rbo, false); in dce4_crtc_do_set_base() [all …]
|
D | radeon_uvd.c | 302 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, in radeon_uvd_force_into_uvd_segment() argument 307 for (i = 0; i < rbo->placement.num_placement; ++i) { in radeon_uvd_force_into_uvd_segment() 308 rbo->placements[i].fpfn = 0 >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 309 rbo->placements[i].lpfn = (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 317 if (rbo->placement.num_placement > 1) in radeon_uvd_force_into_uvd_segment() 321 rbo->placements[1] = rbo->placements[0]; in radeon_uvd_force_into_uvd_segment() 322 rbo->placements[1].fpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 323 rbo->placements[1].lpfn += (256 * 1024 * 1024) >> PAGE_SHIFT; in radeon_uvd_force_into_uvd_segment() 324 rbo->placement.num_placement++; in radeon_uvd_force_into_uvd_segment() 325 rbo->placement.num_busy_placement++; in radeon_uvd_force_into_uvd_segment()
|
D | radeon.h | 1694 void radeon_uvd_force_into_uvd_segment(struct radeon_bo *rbo, 2817 extern void radeon_ttm_placement_from_domain(struct radeon_bo *rbo, u32 domain);
|
/kernel/linux/linux-5.10/drivers/gpu/drm/qxl/ |
D | qxl_release.c | 323 struct qxl_bo **rbo) in qxl_alloc_release_reserved() argument 349 if (rbo) in qxl_alloc_release_reserved() 350 *rbo = NULL; in qxl_alloc_release_reserved() 375 if (rbo) in qxl_alloc_release_reserved() 376 *rbo = bo; in qxl_alloc_release_reserved()
|
D | qxl_drv.h | 405 struct qxl_bo **rbo);
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/display/amdgpu_dm/ |
D | amdgpu_dm.c | 3774 struct amdgpu_bo *rbo; in get_fb_info() local 3783 rbo = gem_to_amdgpu_bo(amdgpu_fb->base.obj[0]); in get_fb_info() 3784 r = amdgpu_bo_reserve(rbo, false); in get_fb_info() 3794 amdgpu_bo_get_tiling_flags(rbo, tiling_flags); in get_fb_info() 3797 *tmz_surface = amdgpu_bo_encrypted(rbo); in get_fb_info() 3799 amdgpu_bo_unreserve(rbo); in get_fb_info() 5864 struct amdgpu_bo *rbo; in dm_plane_helper_prepare_fb() local 5879 rbo = gem_to_amdgpu_bo(obj); in dm_plane_helper_prepare_fb() 5880 adev = amdgpu_ttm_adev(rbo->tbo.bdev); in dm_plane_helper_prepare_fb() 5883 tv.bo = &rbo->tbo; in dm_plane_helper_prepare_fb() [all …]
|