Home
last modified time | relevance | path

Searched refs:vbo (Results 1 – 11 of 11) sorted by relevance

/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/
Dvmwgfx_page_dirty.c83 static void vmw_bo_dirty_scan_pagetable(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan_pagetable() argument
85 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_pagetable()
86 pgoff_t offset = drm_vma_node_start(&vbo->base.base.vma_node); in vmw_bo_dirty_scan_pagetable()
87 struct address_space *mapping = vbo->base.bdev->dev_mapping; in vmw_bo_dirty_scan_pagetable()
121 static void vmw_bo_dirty_scan_mkwrite(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan_mkwrite() argument
123 struct vmw_bo_dirty *dirty = vbo->dirty; in vmw_bo_dirty_scan_mkwrite()
124 unsigned long offset = drm_vma_node_start(&vbo->base.base.vma_node); in vmw_bo_dirty_scan_mkwrite()
125 struct address_space *mapping = vbo->base.bdev->dev_mapping; in vmw_bo_dirty_scan_mkwrite()
131 num_marked = wp_shared_mapping_range(vbo->base.bdev->dev_mapping, in vmw_bo_dirty_scan_mkwrite()
165 void vmw_bo_dirty_scan(struct vmw_buffer_object *vbo) in vmw_bo_dirty_scan() argument
[all …]
Dvmwgfx_bo.c43 struct vmw_buffer_object vbo; member
75 return container_of(vmw_bo, struct vmw_user_buffer_object, vbo); in vmw_user_buffer_object()
335 void vmw_bo_pin_reserved(struct vmw_buffer_object *vbo, bool pin) in vmw_bo_pin_reserved() argument
340 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_pin_reserved()
347 if (vbo->pin_count++ > 0) in vmw_bo_pin_reserved()
350 WARN_ON(vbo->pin_count <= 0); in vmw_bo_pin_reserved()
351 if (--vbo->pin_count > 0) in vmw_bo_pin_reserved()
389 void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo) in vmw_bo_map_and_cache() argument
391 struct ttm_buffer_object *bo = &vbo->base; in vmw_bo_map_and_cache()
396 virtual = ttm_kmap_obj_virtual(&vbo->map, &not_used); in vmw_bo_map_and_cache()
[all …]
Dvmwgfx_validation.c178 struct vmw_buffer_object *vbo) in vmw_validation_find_bo_dup() argument
188 if (!drm_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) in vmw_validation_find_bo_dup()
194 if (entry->base.bo == &vbo->base) { in vmw_validation_find_bo_dup()
259 struct vmw_buffer_object *vbo, in vmw_validation_add_bo() argument
265 bo_node = vmw_validation_find_bo_dup(ctx, vbo); in vmw_validation_add_bo()
281 bo_node->hash.key = (unsigned long) vbo; in vmw_validation_add_bo()
290 val_buf->bo = ttm_bo_get_unless_zero(&vbo->base); in vmw_validation_add_bo()
416 struct vmw_buffer_object *vbo, in vmw_validation_res_switch_backup() argument
427 val->new_backup = vbo; in vmw_validation_res_switch_backup()
457 struct vmw_buffer_object *vbo = res->backup; in vmw_validation_res_reserve() local
[all …]
Dvmwgfx_resource.c780 void vmw_resource_unbind_list(struct vmw_buffer_object *vbo) in vmw_resource_unbind_list() argument
783 .bo = &vbo->base, in vmw_resource_unbind_list()
787 dma_resv_assert_held(vbo->base.base.resv); in vmw_resource_unbind_list()
788 while (!RB_EMPTY_ROOT(&vbo->res_tree)) { in vmw_resource_unbind_list()
789 struct rb_node *node = vbo->res_tree.rb_node; in vmw_resource_unbind_list()
801 (void) ttm_bo_wait(&vbo->base, false, false); in vmw_resource_unbind_list()
999 struct vmw_buffer_object *vbo = NULL; in vmw_resource_pin() local
1002 vbo = res->backup; in vmw_resource_pin()
1004 ttm_bo_reserve(&vbo->base, interruptible, false, NULL); in vmw_resource_pin()
1005 if (!vbo->pin_count) { in vmw_resource_pin()
[all …]
Dvmwgfx_drv.h797 extern void vmw_resource_unbind_list(struct vmw_buffer_object *vbo);
802 int vmw_resources_clean(struct vmw_buffer_object *vbo, pgoff_t start,
876 extern void *vmw_bo_map_and_cache(struct vmw_buffer_object *vbo);
877 extern void vmw_bo_unmap(struct vmw_buffer_object *vbo);
898 static inline void vmw_bo_prio_adjust(struct vmw_buffer_object *vbo) in vmw_bo_prio_adjust() argument
900 int i = ARRAY_SIZE(vbo->res_prios); in vmw_bo_prio_adjust()
903 if (vbo->res_prios[i]) { in vmw_bo_prio_adjust()
904 vbo->base.priority = i; in vmw_bo_prio_adjust()
909 vbo->base.priority = 3; in vmw_bo_prio_adjust()
921 static inline void vmw_bo_prio_add(struct vmw_buffer_object *vbo, int prio) in vmw_bo_prio_add() argument
[all …]
Dvmwgfx_drv.c360 struct vmw_buffer_object *vbo; in vmw_dummy_query_bo_create() local
370 vbo = kzalloc(sizeof(*vbo), GFP_KERNEL); in vmw_dummy_query_bo_create()
371 if (!vbo) in vmw_dummy_query_bo_create()
374 ret = vmw_bo_init(dev_priv, vbo, PAGE_SIZE, in vmw_dummy_query_bo_create()
380 ret = ttm_bo_reserve(&vbo->base, false, true, NULL); in vmw_dummy_query_bo_create()
382 vmw_bo_pin_reserved(vbo, true); in vmw_dummy_query_bo_create()
384 ret = ttm_bo_kmap(&vbo->base, 0, 1, &map); in vmw_dummy_query_bo_create()
392 vmw_bo_pin_reserved(vbo, false); in vmw_dummy_query_bo_create()
393 ttm_bo_unreserve(&vbo->base); in vmw_dummy_query_bo_create()
397 vmw_bo_unreference(&vbo); in vmw_dummy_query_bo_create()
[all …]
Dvmwgfx_validation.h221 struct vmw_buffer_object *vbo,
241 struct vmw_buffer_object *vbo,
Dvmwgfx_fb.c187 struct vmw_buffer_object *vbo = par->vmw_bo; in vmw_fb_dirty_flush() local
199 (void) ttm_bo_reserve(&vbo->base, false, false, NULL); in vmw_fb_dirty_flush()
200 virtual = vmw_bo_map_and_cache(vbo); in vmw_fb_dirty_flush()
256 ttm_bo_unreserve(&vbo->base); in vmw_fb_dirty_flush()
Dvmwgfx_execbuf.c67 struct vmw_buffer_object *vbo; member
1185 reloc->vbo = vmw_bo; in vmw_translate_mob_ptr()
1240 reloc->vbo = vmw_bo; in vmw_translate_guest_ptr()
1717 struct vmw_buffer_object *vbo; in vmw_cmd_res_switch_backup() local
1725 ret = vmw_translate_mob_ptr(dev_priv, sw_context, buf_id, &vbo); in vmw_cmd_res_switch_backup()
1729 vmw_validation_res_switch_backup(sw_context->ctx, info, vbo, in vmw_cmd_res_switch_backup()
3708 bo = &reloc->vbo->base; in vmw_apply_relocations()
/kernel/linux/linux-5.10/drivers/gpu/drm/imx/
Dipuv3-plane.c354 unsigned long eba, ubo, vbo, old_ubo, old_vbo, alpha_eba; in ipu_plane_atomic_check() local
436 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_check()
438 if (vbo & 0x7 || vbo > 0xfffff8) in ipu_plane_atomic_check()
443 if (vbo != old_vbo) in ipu_plane_atomic_check()
553 unsigned long eba, ubo, vbo; in ipu_plane_atomic_update() local
652 vbo = drm_plane_state_to_vbo(state); in ipu_plane_atomic_update()
656 swap(ubo, vbo); in ipu_plane_atomic_update()
659 fb->pitches[1], ubo, vbo); in ipu_plane_atomic_update()
662 "phy = %lu %lu %lu, x = %d, y = %d", eba, ubo, vbo, in ipu_plane_atomic_update()
/kernel/linux/linux-5.10/drivers/gpu/drm/vc4/
Dvc4_validate.c887 struct drm_gem_cma_object *vbo = in validate_gl_shader_rec() local
896 to_vc4_bo(&vbo->base)->write_seqno); in validate_gl_shader_rec()
901 if (vbo->base.size < offset || in validate_gl_shader_rec()
902 vbo->base.size - offset < attr_size) { in validate_gl_shader_rec()
904 offset, attr_size, vbo->base.size); in validate_gl_shader_rec()
909 max_index = ((vbo->base.size - offset - attr_size) / in validate_gl_shader_rec()
919 *(uint32_t *)(pkt_v + o) = vbo->paddr + offset; in validate_gl_shader_rec()