/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_bo.c | 43 struct vmw_buffer_object vbo; member 75 return container_of(vmw_bo, struct vmw_user_buffer_object, vbo); in vmw_user_buffer_object() 335 void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin) in vmw_bo_pin_reserved() argument 340 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_pin_reserved() 347 if (vbo->pin_count++ > 0) in vmw_bo_pin_reserved() 350 WARN_ON(vbo->pin_count <= 0); in vmw_bo_pin_reserved() 351 if (--vbo->pin_count > 0) in vmw_bo_pin_reserved() 387 void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo) in vmw_bo_map_and_cache() argument 389 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_map_and_cache() 394 virtual = ttm_kmap_obj_virtual(&vbo->map, ¬_used); in vmw_bo_map_and_cache() [all …]
|
D | vmwgfx_validation.c | 175 struct vmw_buffer_object *vbo) in vmw_validation_find_bo_dup() argument 185 if (!drm_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) in vmw_validation_find_bo_dup() 191 if (entry->base.bo == &vbo->base) { in vmw_validation_find_bo_dup() 256 struct vmw_buffer_object *vbo, in vmw_validation_add_bo() argument 262 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo() 278 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo() 287 val_buf->bo = ttm_bo_get_unless_zero(&vbo->base); in vmw_validation_add_bo() 413 struct vmw_buffer_object *vbo, in vmw_validation_res_switch_backup() argument 424 val->new_backup = vbo; in vmw_validation_res_switch_backup() 454 struct vmw_buffer_object *vbo = res->backup; in vmw_validation_res_reserve() local [all …]
|
D | vmwgfx_resource.c | 713 void vmw_resource_unbind_list(struct vmw_buffer_object *vbo) in vmw_resource_unbind_list() argument 718 .bo = &vbo->base, in vmw_resource_unbind_list() 722 dma_resv_assert_held(vbo->base.base.resv); in vmw_resource_unbind_list() 723 list_for_each_entry_safe(res, next, &vbo->res_list, mob_head) { in vmw_resource_unbind_list() 733 (void) ttm_bo_wait(&vbo->base, false, false); in vmw_resource_unbind_list() 931 struct vmw_buffer_object *vbo = NULL; in vmw_resource_pin() local 934 vbo = res->backup; in vmw_resource_pin() 936 ttm_bo_reserve(&vbo->base, interruptible, false, NULL); in vmw_resource_pin() 937 if (!vbo->pin_count) { in vmw_resource_pin() 939 (&vbo->base, in vmw_resource_pin() [all …]
|
D | vmwgfx_drv.h | 716 extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo); 791 extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo); 792 extern void vmw_bo_unmap(struct vmw_buffer_object *vbo); 813 static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo) in vmw_bo_prio_adjust() argument 815 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust() 818 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust() 819 vbo->base.priority = i; in vmw_bo_prio_adjust() 824 vbo->base.priority = 3; in vmw_bo_prio_adjust() 836 static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio) in vmw_bo_prio_add() argument 838 if (vbo->res_prios[prio]++ == 0) in vmw_bo_prio_add() [all …]
|
D | vmwgfx_drv.c | 351 struct vmw_buffer_object *vbo; in vmw_dummy_query_bo_create() local 361 vbo = kzalloc(sizeof(*vbo), GFP_KERNEL); in vmw_dummy_query_bo_create() 362 if (!vbo) in vmw_dummy_query_bo_create() 365 ret = vmw_bo_init(dev_priv, vbo, PAGE_SIZE, in vmw_dummy_query_bo_create() 371 ret = ttm_bo_reserve(&vbo->base, false, true, NULL); in vmw_dummy_query_bo_create() 373 vmw_bo_pin_reserved(vbo, true); in vmw_dummy_query_bo_create() 375 ret = ttm_bo_kmap(&vbo->base, 0, 1, &map); in vmw_dummy_query_bo_create() 383 vmw_bo_pin_reserved(vbo, false); in vmw_dummy_query_bo_create() 384 ttm_bo_unreserve(&vbo->base); in vmw_dummy_query_bo_create() 388 vmw_bo_unreference(&vbo); in vmw_dummy_query_bo_create() [all …]
|
D | vmwgfx_validation.h | 235 struct vmw_buffer_object *vbo, 255 struct vmw_buffer_object *vbo,
|
D | vmwgfx_fb.c | 187 struct vmw_buffer_object *vbo = par->vmw_bo; in vmw_fb_dirty_flush() local 199 (void) ttm_bo_reserve(&vbo->base, false, false, NULL); in vmw_fb_dirty_flush() 200 virtual = vmw_bo_map_and_cache(vbo); in vmw_fb_dirty_flush() 256 ttm_bo_unreserve(&vbo->base); in vmw_fb_dirty_flush()
|
D | vmwgfx_execbuf.c | 67 struct vmw_buffer_object *vbo; member 1181 reloc->vbo = vmw_bo; in vmw_translate_mob_ptr() 1236 reloc->vbo = vmw_bo; in vmw_translate_guest_ptr() 1713 struct vmw_buffer_object *vbo; in vmw_cmd_res_switch_backup() local 1721 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &vbo); in vmw_cmd_res_switch_backup() 1725 vmw_validation_res_switch_backup(sw_context->ctx, info, vbo, in vmw_cmd_res_switch_backup() 3305 bo = &reloc->vbo->base; in vmw_apply_relocations()
|
/drivers/gpu/drm/imx/ |
D | ipuv3-plane.c | 354 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local 436 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_check() 438 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check() 443 if (vbo != old_vbo) in ipu_plane_atomic_check() 553 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local 652 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_update() 656 swap(ubo, vbo); in ipu_plane_atomic_update() 659 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update() 662 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update()
|
/drivers/gpu/drm/vc4/ |
D | vc4_validate.c | 887 struct drm_gem_cma_object *vbo = in validate_gl_shader_rec() local 896 to_vc4_bo(&vbo->base)->write_seqno); in validate_gl_shader_rec() 901 if (vbo->base.size < offset || in validate_gl_shader_rec() 902 vbo->base.size - offset < attr_size) { in validate_gl_shader_rec() 904 offset, attr_size, vbo->base.size); in validate_gl_shader_rec() 909 max_index = ((vbo->base.size - offset - attr_size) / in validate_gl_shader_rec() 919 *(uint32_t *)(pkt_v + o) = vbo->paddr + offset; in validate_gl_shader_rec()
|