Home
last modified time | relevance | path

Searched refs:vbo (Results 1 – 10 of 10) sorted by relevance

/drivers/gpu/drm/vmwgfx/
Dvmwgfx_bo.c43 struct vmw_buffer_object vbo; member
75 return container_of(vmw_bo, struct vmw_user_buffer_object, vbo); in vmw_user_buffer_object()
335 void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin) in vmw_bo_pin_reserved() argument
340 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_pin_reserved()
347 if (vbo->pin_count++ > 0) in vmw_bo_pin_reserved()
350 WARN_ON(vbo->pin_count <= 0); in vmw_bo_pin_reserved()
351 if (--vbo->pin_count > 0) in vmw_bo_pin_reserved()
387 void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo) in vmw_bo_map_and_cache() argument
389 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_map_and_cache()
394 virtual = ttm_kmap_obj_virtual(&vbo->map, &not_used); in vmw_bo_map_and_cache()
[all …]
Dvmwgfx_validation.c175 struct vmw_buffer_object *vbo) in vmw_validation_find_bo_dup() argument
185 if (!drm_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) in vmw_validation_find_bo_dup()
191 if (entry->base.bo == &vbo->base) { in vmw_validation_find_bo_dup()
256 struct vmw_buffer_object *vbo, in vmw_validation_add_bo() argument
262 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo()
278 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo()
287 val_buf->bo = ttm_bo_get_unless_zero(&vbo->base); in vmw_validation_add_bo()
413 struct vmw_buffer_object *vbo, in vmw_validation_res_switch_backup() argument
424 val->new_backup = vbo; in vmw_validation_res_switch_backup()
454 struct vmw_buffer_object *vbo = res->backup; in vmw_validation_res_reserve() local
[all …]
Dvmwgfx_resource.c713 void vmw_resource_unbind_list(struct vmw_buffer_object *vbo) in vmw_resource_unbind_list() argument
718 .bo = &vbo->base, in vmw_resource_unbind_list()
722 dma_resv_assert_held(vbo->base.base.resv); in vmw_resource_unbind_list()
723 list_for_each_entry_safe(res, next, &vbo->res_list, mob_head) { in vmw_resource_unbind_list()
733 (void) ttm_bo_wait(&vbo->base, false, false); in vmw_resource_unbind_list()
931 struct vmw_buffer_object *vbo = NULL; in vmw_resource_pin() local
934 vbo = res->backup; in vmw_resource_pin()
936 ttm_bo_reserve(&vbo->base, interruptible, false, NULL); in vmw_resource_pin()
937 if (!vbo->pin_count) { in vmw_resource_pin()
939 (&vbo->base, in vmw_resource_pin()
[all …]
Dvmwgfx_drv.h716 extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo);
791 extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo);
792 extern void vmw_bo_unmap(struct vmw_buffer_object *vbo);
813 static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo) in vmw_bo_prio_adjust() argument
815 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust()
818 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust()
819 vbo->base.priority = i; in vmw_bo_prio_adjust()
824 vbo->base.priority = 3; in vmw_bo_prio_adjust()
836 static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio) in vmw_bo_prio_add() argument
838 if (vbo->res_prios[prio]++ == 0) in vmw_bo_prio_add()
[all …]
Dvmwgfx_drv.c351 struct vmw_buffer_object *vbo; in vmw_dummy_query_bo_create() local
361 vbo = kzalloc(sizeof(*vbo), GFP_KERNEL); in vmw_dummy_query_bo_create()
362 if (!vbo) in vmw_dummy_query_bo_create()
365 ret = vmw_bo_init(dev_priv, vbo, PAGE_SIZE, in vmw_dummy_query_bo_create()
371 ret = ttm_bo_reserve(&vbo->base, false, true, NULL); in vmw_dummy_query_bo_create()
373 vmw_bo_pin_reserved(vbo, true); in vmw_dummy_query_bo_create()
375 ret = ttm_bo_kmap(&vbo->base, 0, 1, &map); in vmw_dummy_query_bo_create()
383 vmw_bo_pin_reserved(vbo, false); in vmw_dummy_query_bo_create()
384 ttm_bo_unreserve(&vbo->base); in vmw_dummy_query_bo_create()
388 vmw_bo_unreference(&vbo); in vmw_dummy_query_bo_create()
[all …]
Dvmwgfx_validation.h235 struct vmw_buffer_object *vbo,
255 struct vmw_buffer_object *vbo,
Dvmwgfx_fb.c187 struct vmw_buffer_object *vbo = par->vmw_bo; in vmw_fb_dirty_flush() local
199 (void) ttm_bo_reserve(&vbo->base, false, false, NULL); in vmw_fb_dirty_flush()
200 virtual = vmw_bo_map_and_cache(vbo); in vmw_fb_dirty_flush()
256 ttm_bo_unreserve(&vbo->base); in vmw_fb_dirty_flush()
Dvmwgfx_execbuf.c67 struct vmw_buffer_object *vbo; member
1181 reloc->vbo = vmw_bo; in vmw_translate_mob_ptr()
1236 reloc->vbo = vmw_bo; in vmw_translate_guest_ptr()
1713 struct vmw_buffer_object *vbo; in vmw_cmd_res_switch_backup() local
1721 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &vbo); in vmw_cmd_res_switch_backup()
1725 vmw_validation_res_switch_backup(sw_context->ctx, info, vbo, in vmw_cmd_res_switch_backup()
3305 bo = &reloc->vbo->base; in vmw_apply_relocations()
/drivers/gpu/drm/imx/
Dipuv3-plane.c354 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local
436 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_check()
438 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check()
443 if (vbo != old_vbo) in ipu_plane_atomic_check()
553 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local
652 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_update()
656 swap(ubo, vbo); in ipu_plane_atomic_update()
659 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update()
662 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update()
/drivers/gpu/drm/vc4/
Dvc4_validate.c887 struct drm_gem_cma_object *vbo = in validate_gl_shader_rec() local
896 to_vc4_bo(&vbo->base)->write_seqno); in validate_gl_shader_rec()
901 if (vbo->base.size < offset || in validate_gl_shader_rec()
902 vbo->base.size - offset < attr_size) { in validate_gl_shader_rec()
904 offset, attr_size, vbo->base.size); in validate_gl_shader_rec()
909 max_index = ((vbo->base.size - offset - attr_size) / in validate_gl_shader_rec()
919 *(uint32_t *)(pkt_v + o) = vbo->paddr + offset; in validate_gl_shader_rec()