/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_page_dirty.c | 83 static void vmw_bo_dirty_scan_pagetable(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan_pagetable() argument 85 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable() 86 pgoff_t offset = drm_vma_node_start(&vbo->base.base.vma_node); in vmw_bo_dirty_scan_pagetable() 87 struct address_space *mapping = vbo->base.bdev->dev_mapping; in vmw_bo_dirty_scan_pagetable() 121 static void vmw_bo_dirty_scan_mkwrite(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan_mkwrite() argument 123 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_mkwrite() 124 unsigned long offset = drm_vma_node_start(&vbo->base.base.vma_node); in vmw_bo_dirty_scan_mkwrite() 125 struct address_space *mapping = vbo->base.bdev->dev_mapping; in vmw_bo_dirty_scan_mkwrite() 131 num_marked = wp_shared_mapping_range(vbo->base.bdev->dev_mapping, in vmw_bo_dirty_scan_mkwrite() 165 void vmw_bo_dirty_scan(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan() argument [all …]
|
D | vmwgfx_bo.c | 43 struct vmw_buffer_object vbo; member 75 return container_of(vmw_bo, struct vmw_user_buffer_object, vbo); in vmw_user_buffer_object() 335 void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin) in vmw_bo_pin_reserved() argument 340 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_pin_reserved() 347 if (vbo->pin_count++ > 0) in vmw_bo_pin_reserved() 350 WARN_ON(vbo->pin_count <= 0); in vmw_bo_pin_reserved() 351 if (--vbo->pin_count > 0) in vmw_bo_pin_reserved() 389 void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo) in vmw_bo_map_and_cache() argument 391 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_map_and_cache() 396 virtual = ttm_kmap_obj_virtual(&vbo->map, ¬_used); in vmw_bo_map_and_cache() [all …]
|
D | vmwgfx_validation.c | 178 struct vmw_buffer_object *vbo) in vmw_validation_find_bo_dup() argument 188 if (!drm_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) in vmw_validation_find_bo_dup() 194 if (entry->base.bo == &vbo->base) { in vmw_validation_find_bo_dup() 259 struct vmw_buffer_object *vbo, in vmw_validation_add_bo() argument 265 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo() 281 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo() 290 val_buf->bo = ttm_bo_get_unless_zero(&vbo->base); in vmw_validation_add_bo() 416 struct vmw_buffer_object *vbo, in vmw_validation_res_switch_backup() argument 427 val->new_backup = vbo; in vmw_validation_res_switch_backup() 457 struct vmw_buffer_object *vbo = res->backup; in vmw_validation_res_reserve() local [all …]
|
D | vmwgfx_resource.c | 780 void vmw_resource_unbind_list(struct vmw_buffer_object *vbo) in vmw_resource_unbind_list() argument 783 .bo = &vbo->base, in vmw_resource_unbind_list() 787 dma_resv_assert_held(vbo->base.base.resv); in vmw_resource_unbind_list() 788 while (!RB_EMPTY_ROOT(&vbo->res_tree)) { in vmw_resource_unbind_list() 789 struct rb_node *node = vbo->res_tree.rb_node; in vmw_resource_unbind_list() 801 (void) ttm_bo_wait(&vbo->base, false, false); in vmw_resource_unbind_list() 999 struct vmw_buffer_object *vbo = NULL; in vmw_resource_pin() local 1002 vbo = res->backup; in vmw_resource_pin() 1004 ttm_bo_reserve(&vbo->base, interruptible, false, NULL); in vmw_resource_pin() 1005 if (!vbo->pin_count) { in vmw_resource_pin() [all …]
|
D | vmwgfx_drv.h | 797 extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo); 802 int vmw_resources_clean(struct vmw_buffer_object *vbo, pgoff_t start, 876 extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo); 877 extern void vmw_bo_unmap(struct vmw_buffer_object *vbo); 898 static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo) in vmw_bo_prio_adjust() argument 900 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust() 903 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust() 904 vbo->base.priority = i; in vmw_bo_prio_adjust() 909 vbo->base.priority = 3; in vmw_bo_prio_adjust() 921 static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio) in vmw_bo_prio_add() argument [all …]
|
D | vmwgfx_drv.c | 360 struct vmw_buffer_object *vbo; in vmw_dummy_query_bo_create() local 370 vbo = kzalloc(sizeof(*vbo), GFP_KERNEL); in vmw_dummy_query_bo_create() 371 if (!vbo) in vmw_dummy_query_bo_create() 374 ret = vmw_bo_init(dev_priv, vbo, PAGE_SIZE, in vmw_dummy_query_bo_create() 380 ret = ttm_bo_reserve(&vbo->base, false, true, NULL); in vmw_dummy_query_bo_create() 382 vmw_bo_pin_reserved(vbo, true); in vmw_dummy_query_bo_create() 384 ret = ttm_bo_kmap(&vbo->base, 0, 1, &map); in vmw_dummy_query_bo_create() 392 vmw_bo_pin_reserved(vbo, false); in vmw_dummy_query_bo_create() 393 ttm_bo_unreserve(&vbo->base); in vmw_dummy_query_bo_create() 397 vmw_bo_unreference(&vbo); in vmw_dummy_query_bo_create() [all …]
|
D | vmwgfx_validation.h | 221 struct vmw_buffer_object *vbo, 241 struct vmw_buffer_object *vbo,
|
D | vmwgfx_fb.c | 187 struct vmw_buffer_object *vbo = par->vmw_bo; in vmw_fb_dirty_flush() local 199 (void) ttm_bo_reserve(&vbo->base, false, false, NULL); in vmw_fb_dirty_flush() 200 virtual = vmw_bo_map_and_cache(vbo); in vmw_fb_dirty_flush() 256 ttm_bo_unreserve(&vbo->base); in vmw_fb_dirty_flush()
|
D | vmwgfx_execbuf.c | 67 struct vmw_buffer_object *vbo; member 1185 reloc->vbo = vmw_bo; in vmw_translate_mob_ptr() 1240 reloc->vbo = vmw_bo; in vmw_translate_guest_ptr() 1717 struct vmw_buffer_object *vbo; in vmw_cmd_res_switch_backup() local 1725 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &vbo); in vmw_cmd_res_switch_backup() 1729 vmw_validation_res_switch_backup(sw_context->ctx, info, vbo, in vmw_cmd_res_switch_backup() 3708 bo = &reloc->vbo->base; in vmw_apply_relocations()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/imx/ |
D | ipuv3-plane.c | 354 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local 436 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_check() 438 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check() 443 if (vbo != old_vbo) in ipu_plane_atomic_check() 553 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local 652 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_update() 656 swap(ubo, vbo); in ipu_plane_atomic_update() 659 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update() 662 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/vc4/ |
D | vc4_validate.c | 887 struct drm_gem_cma_object *vbo = in validate_gl_shader_rec() local 896 to_vc4_bo(&vbo->base)->write_seqno); in validate_gl_shader_rec() 901 if (vbo->base.size < offset || in validate_gl_shader_rec() 902 vbo->base.size - offset < attr_size) { in validate_gl_shader_rec() 904 offset, attr_size, vbo->base.size); in validate_gl_shader_rec() 909 max_index = ((vbo->base.size - offset - attr_size) / in validate_gl_shader_rec() 919 *(uint32_t *)(pkt_v + o) = vbo->paddr + offset; in validate_gl_shader_rec()
|