Lines Matching refs:backup
890 &res->backup, in vmw_surface_define_ioctl()
1235 if (res->backup->dirty && res->backup_dirty) { in vmw_gb_surface_bind()
1543 &res->backup, in vmw_gb_surface_define_internal()
1546 if (res->backup->base.base.size < res->backup_size) { in vmw_gb_surface_define_internal()
1548 vmw_bo_unreference(&res->backup); in vmw_gb_surface_define_internal()
1563 &res->backup, in vmw_gb_surface_define_internal()
1572 struct vmw_buffer_object *backup = res->backup; in vmw_gb_surface_define_internal() local
1574 ttm_bo_reserve(&backup->base, false, false, NULL); in vmw_gb_surface_define_internal()
1578 ret = vmw_bo_dirty_add(backup); in vmw_gb_surface_define_internal()
1583 ttm_bo_unreserve(&backup->base); in vmw_gb_surface_define_internal()
1606 if (res->backup) { in vmw_gb_surface_define_internal()
1608 drm_vma_node_offset_addr(&res->backup->base.base.vma_node); in vmw_gb_surface_define_internal()
1609 rep->buffer_size = res->backup->base.base.size; in vmw_gb_surface_define_internal()
1654 if (!srf->res.backup) { in vmw_gb_surface_reference_internal()
1661 ret = vmw_user_bo_reference(tfile, srf->res.backup, &backup_handle); in vmw_gb_surface_reference_internal()
1685 drm_vma_node_offset_addr(&srf->res.backup->base.base.vma_node); in vmw_gb_surface_reference_internal()
1686 rep->crep.buffer_size = srf->res.backup->base.base.size; in vmw_gb_surface_reference_internal()