Searched refs:num_fences (Results 1 – 12 of 12) sorted by relevance
1261 const unsigned num_fences = bo->num_fences; in amdgpu_add_bo_fence_dependencies() local1263 for (unsigned j = 0; j < num_fences; ++j) { in amdgpu_add_bo_fence_dependencies()1278 for (unsigned j = new_num_fences; j < num_fences; ++j) in amdgpu_add_bo_fence_dependencies()1281 bo->num_fences = new_num_fences; in amdgpu_add_bo_fence_dependencies()1289 unsigned num_fences, in amdgpu_add_fences() argument1292 if (bo->num_fences + num_fences > bo->max_fences) { in amdgpu_add_fences()1293 unsigned new_max_fences = MAX2(bo->num_fences + num_fences, bo->max_fences * 2); in amdgpu_add_fences()1296 bo->num_fences * sizeof(*new_fences), in amdgpu_add_fences()1308 if (!bo->num_fences) in amdgpu_add_fences()1311 bo->num_fences--; /* prefer to keep the most recent fence if possible */ in amdgpu_add_fences()[all …]
91 for (idle_fences = 0; idle_fences < bo->num_fences; ++idle_fences) { in amdgpu_bo_wait()101 (bo->num_fences - idle_fences) * sizeof(*bo->fences)); in amdgpu_bo_wait()102 bo->num_fences -= idle_fences; in amdgpu_bo_wait()104 buffer_idle = !bo->num_fences; in amdgpu_bo_wait()112 while (bo->num_fences && buffer_idle) { in amdgpu_bo_wait()129 if (fence_idle && bo->num_fences && bo->fences[0] == fence) { in amdgpu_bo_wait()132 (bo->num_fences - 1) * sizeof(*bo->fences)); in amdgpu_bo_wait()133 bo->num_fences--; in amdgpu_bo_wait()158 for (unsigned i = 0; i < bo->num_fences; ++i) in amdgpu_bo_remove_fences()162 bo->num_fences = 0; in amdgpu_bo_remove_fences()[all …]
106 uint16_t num_fences; member
260 unsigned num_fences,
551 for (unsigned src = 0; src < bo->u.slab.num_fences; ++src) { in radeon_bo_slab_fence()559 bo->u.slab.num_fences = dst; in radeon_bo_slab_fence()562 if (bo->u.slab.num_fences >= bo->u.slab.max_fences) { in radeon_bo_slab_fence()577 bo->u.slab.fences[bo->u.slab.num_fences] = NULL; in radeon_bo_slab_fence()578 radeon_ws_bo_reference(&bo->u.slab.fences[bo->u.slab.num_fences], fence); in radeon_bo_slab_fence()579 bo->u.slab.num_fences++; in radeon_bo_slab_fence()
80 for (num_idle = 0; num_idle < bo->u.slab.num_fences; ++num_idle) { in radeon_bo_is_busy()88 (bo->u.slab.num_fences - num_idle) * sizeof(bo->u.slab.fences[0])); in radeon_bo_is_busy()89 bo->u.slab.num_fences -= num_idle; in radeon_bo_is_busy()110 while (bo->u.slab.num_fences) { in radeon_bo_wait_idle()119 if (bo->u.slab.num_fences && fence == bo->u.slab.fences[0]) { in radeon_bo_wait_idle()122 (bo->u.slab.num_fences - 1) * sizeof(bo->u.slab.fences[0])); in radeon_bo_wait_idle()123 bo->u.slab.num_fences--; in radeon_bo_wait_idle()834 for (unsigned j = 0; j < bo->u.slab.num_fences; ++j) in radeon_bo_slab_free()
50 unsigned num_fences; member
82 __u32 num_fences; member
74 num_fences(struct crocus_batch *batch) in num_fences() function86 fprintf(stderr, "Fence list (length %u): ", num_fences(batch)); in dump_fence_list()885 if (num_fences(batch)) { in submit_batch()887 execbuf.num_cliprects = num_fences(batch); in submit_batch()
90 uint32_t num_fences;
73 num_fences(struct iris_batch *batch) in num_fences() function85 fprintf(stderr, "Fence list (length %u): ", num_fences(batch)); in dump_fence_list()985 if (num_fences(batch)) { in submit_batch()987 execbuf.num_cliprects = num_fences(batch); in submit_batch()
68764 unsigned int num_fences; member68812 __u32 num_fences; member