/external/angle/src/tests/gl_tests/ |
D | FenceSyncTests.cpp | 113 GLuint fences[kFenceCount] = {0}; in TEST_P() local 114 glGenFencesNV(static_cast<GLsizei>(ArraySize(fences)), fences); in TEST_P() 117 for (GLuint fence : fences) in TEST_P() 124 glFinishFenceNV(fences[kFenceCount - 1]); in TEST_P() 126 for (GLuint fence : fences) in TEST_P()
|
/external/mesa3d/src/vulkan/wsi/ |
D | wsi_common.c | 277 if (chain->fences) { in wsi_swapchain_finish() 279 chain->wsi->DestroyFence(chain->device, chain->fences[i], &chain->alloc); in wsi_swapchain_finish() 281 vk_free(&chain->alloc, chain->fences); in wsi_swapchain_finish() 472 swapchain->fences = vk_zalloc(pAllocator, in wsi_common_create_swapchain() 473 sizeof (*swapchain->fences) * swapchain->image_count, in wsi_common_create_swapchain() 474 sizeof (*swapchain->fences), in wsi_common_create_swapchain() 476 if (!swapchain->fences) { in wsi_common_create_swapchain() 569 if (swapchain->fences[image_index] == VK_NULL_HANDLE) { in wsi_common_queue_present() 577 &swapchain->fences[image_index]); in wsi_common_queue_present() 582 wsi->WaitForFences(device, 1, &swapchain->fences[image_index], in wsi_common_queue_present() [all …]
|
/external/libdrm/amdgpu/ |
D | amdgpu_cs.c | 467 static int amdgpu_ioctl_wait_fences(struct amdgpu_cs_fence *fences, in amdgpu_ioctl_wait_fences() argument 475 amdgpu_device_handle dev = fences[0].context->dev; in amdgpu_ioctl_wait_fences() 482 drm_fences[i].ctx_id = fences[i].context->id; in amdgpu_ioctl_wait_fences() 483 drm_fences[i].ip_type = fences[i].ip_type; in amdgpu_ioctl_wait_fences() 484 drm_fences[i].ip_instance = fences[i].ip_instance; in amdgpu_ioctl_wait_fences() 485 drm_fences[i].ring = fences[i].ring; in amdgpu_ioctl_wait_fences() 486 drm_fences[i].seq_no = fences[i].fence; in amdgpu_ioctl_wait_fences() 490 args.in.fences = (uint64_t)(uintptr_t)drm_fences; in amdgpu_ioctl_wait_fences() 507 drm_public int amdgpu_cs_wait_fences(struct amdgpu_cs_fence *fences, in amdgpu_cs_wait_fences() argument 517 if (!fences || !status || !fence_count) in amdgpu_cs_wait_fences() [all …]
|
/external/igt-gpu-tools/tests/i915/ |
D | gem_exec_nop.c | 722 int *fences, n; in fence_signal() local 737 fences = malloc(sizeof(*fences) * NFENCES); in fence_signal() 738 igt_assert(fences); in fence_signal() 739 memset(fences, -1, sizeof(*fences) * NFENCES); in fence_signal() 758 if (fences[n] != -1) { in fence_signal() 759 igt_assert(fence_wait(fences[n])); in fence_signal() 760 close(fences[n]); in fence_signal() 768 fences[n] = execbuf.rsvd2 >> 32; in fence_signal() 769 signal += fence_enable_signaling(fences[n]); in fence_signal() 781 if (fences[n] != -1) in fence_signal() [all …]
|
D | gem_eio.c | 680 int fences[16]; in test_inflight_internal() local 705 fences[nfence] = execbuf.rsvd2 >> 32; in test_inflight_internal() 706 igt_assert(fences[nfence] != -1); in test_inflight_internal() 713 igt_assert_eq(sync_fence_status(fences[nfence]), -EIO); in test_inflight_internal() 714 close(fences[nfence]); in test_inflight_internal()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AVR/atomics/ |
D | fence.ll | 3 ; Checks that atomic fences are simply removed from IR. 4 ; AVR is always singlethreaded so fences do nothing.
|
/external/mesa3d/src/gallium/winsys/amdgpu/drm/ |
D | amdgpu_cs.c | 897 static void cleanup_fence_list(struct amdgpu_fence_list *fences) in cleanup_fence_list() argument 899 for (unsigned i = 0; i < fences->num; i++) in cleanup_fence_list() 900 amdgpu_fence_reference(&fences->list[i], NULL); in cleanup_fence_list() 901 fences->num = 0; in cleanup_fence_list() 1204 static void add_fence_to_list(struct amdgpu_fence_list *fences, in add_fence_to_list() argument 1207 unsigned idx = fences->num++; in add_fence_to_list() 1209 if (idx >= fences->max) { in add_fence_to_list() 1213 fences->max = idx + increment; in add_fence_to_list() 1214 size = fences->max * sizeof(fences->list[0]); in add_fence_to_list() 1215 fences->list = realloc(fences->list, size); in add_fence_to_list() [all …]
|
D | amdgpu_bo.c | 91 if (!amdgpu_fence_wait(bo->fences[idle_fences], 0, false)) in amdgpu_bo_wait() 97 amdgpu_fence_reference(&bo->fences[i], NULL); in amdgpu_bo_wait() 99 memmove(&bo->fences[0], &bo->fences[idle_fences], in amdgpu_bo_wait() 100 (bo->num_fences - idle_fences) * sizeof(*bo->fences)); in amdgpu_bo_wait() 115 amdgpu_fence_reference(&fence, bo->fences[0]); in amdgpu_bo_wait() 128 if (fence_idle && bo->num_fences && bo->fences[0] == fence) { in amdgpu_bo_wait() 129 amdgpu_fence_reference(&bo->fences[0], NULL); in amdgpu_bo_wait() 130 memmove(&bo->fences[0], &bo->fences[1], in amdgpu_bo_wait() 131 (bo->num_fences - 1) * sizeof(*bo->fences)); in amdgpu_bo_wait() 158 amdgpu_fence_reference(&bo->fences[i], NULL); in amdgpu_bo_remove_fences() [all …]
|
/external/mesa3d/src/util/ |
D | u_queue.c | 658 struct util_queue_fence *fences; in util_queue_finish() local 672 fences = malloc(queue->num_threads * sizeof(*fences)); in util_queue_finish() 676 util_queue_fence_init(&fences[i]); in util_queue_finish() 677 util_queue_add_job(queue, &barrier, &fences[i], in util_queue_finish() 682 util_queue_fence_wait(&fences[i]); in util_queue_finish() 683 util_queue_fence_destroy(&fences[i]); in util_queue_finish() 689 free(fences); in util_queue_finish()
|
/external/deqp/external/vulkancts/modules/vulkan/wsi/ |
D | vktWsiDisplayControlTests.cpp | 296 std::vector<VkFence>& fences) in initFences() argument 298 for (VkFence& fence : fences) in initFences() 304 std::vector<VkFence>& fences) in deinitFences() argument 306 for (VkFence& fence : fences) in deinitFences() 315 fences.clear(); in deinitFences() 1047 std::vector<VkFence> fences = std::vector<VkFence>(availableDisplays.size(), (VkFence)0); in testDisplayEvent() local 1050 initFences(vkd, device, fences); in testDisplayEvent() 1062 VkFence& fence = fences[i]; in testDisplayEvent() 1070 deinitFences (vkd, device, fences); in testDisplayEvent() 1082 std::vector<VkFence> fences = std::vector<VkFence>(1, (VkFence)0); in testDeviceEvent() local [all …]
|
D | vktWsiSharedPresentableImageTests.cpp | 229 std::vector<vk::VkFence>& fences) in initFences() argument 231 for (size_t ndx = 0; ndx < fences.size(); ndx++) in initFences() 232 fences[ndx] = createFence(vkd, device).disown(); in initFences() 237 std::vector<vk::VkFence>& fences) in deinitFences() argument 239 for (size_t ndx = 0; ndx < fences.size(); ndx++) in deinitFences() 241 if (fences[ndx] != (vk::VkFence)0) in deinitFences() 242 vkd.destroyFence(device, fences[ndx], DE_NULL); in deinitFences() 244 fences[ndx] = (vk::VkFence)0; in deinitFences() 247 fences.clear(); in deinitFences()
|
D | vktWsiIncrementalPresentTests.cpp | 217 std::vector<vk::VkFence>& fences) in initFences() argument 219 for (size_t ndx = 0; ndx < fences.size(); ndx++) in initFences() 220 fences[ndx] = createFence(vkd, device).disown(); in initFences() 225 std::vector<vk::VkFence>& fences) in deinitFences() argument 227 for (size_t ndx = 0; ndx < fences.size(); ndx++) in deinitFences() 229 if (fences[ndx] != (vk::VkFence)0) in deinitFences() 230 vkd.destroyFence(device, fences[ndx], DE_NULL); in deinitFences() 232 fences[ndx] = (vk::VkFence)0; in deinitFences() 235 fences.clear(); in deinitFences()
|
D | vktWsiDisplayTimingTests.cpp | 215 std::vector<vk::VkFence>& fences) in initFences() argument 217 for (size_t ndx = 0; ndx < fences.size(); ndx++) in initFences() 218 fences[ndx] = createFence(vkd, device).disown(); in initFences() 223 std::vector<vk::VkFence>& fences) in deinitFences() argument 225 for (size_t ndx = 0; ndx < fences.size(); ndx++) in deinitFences() 227 if (fences[ndx] != (vk::VkFence)0) in deinitFences() 228 vkd.destroyFence(device, fences[ndx], DE_NULL); in deinitFences() 230 fences[ndx] = (vk::VkFence)0; in deinitFences() 233 fences.clear(); in deinitFences()
|
/external/llvm/lib/Target/WebAssembly/ |
D | WebAssemblyInstrAtomics.td | 18 // Atomic fences 21 // TODO: add atomic fences here...
|
/external/igt-gpu-tools/benchmarks/wsim/ |
D | README | 86 Sync (fd) fences 89 Sync fences are also supported as dependencies. 129 Submit fences 132 Submit fences are a type of input fence which are signalled when the originating 133 batch buffer is submitted to the GPU. (In contrary to normal sync fences, which 136 Submit fences have the identical syntax as the sync fences with the lower-case 216 fences. In other words, the rule expression is telling the driver: "If you pick 250 This can also be extended using sync fences to improve chances of the first
|
/external/deqp/external/vulkancts/modules/vulkan/synchronization/ |
D | vktSynchronizationSmokeTests.cpp | 547 VkFence fences[2]; member 578 createFences(vkd, device, false, DE_LENGTH_OF_ARRAY(fences), fences); in TestContext() 583 destroyFences(vkd, device, DE_LENGTH_OF_ARRAY(fences), fences); in ~TestContext() 1021 fenceStatus = deviceInterface.getFenceStatus(device, testContext.fences[0]); in testFences() 1027 fenceStatus = deviceInterface.getFenceStatus(device, testContext.fences[1]); in testFences() 1034 VK_CHECK(deviceInterface.queueSubmit(queue, 1, &submitInfo, testContext.fences[0])); in testFences() 1037 waitStatus = deviceInterface.waitForFences(device, 1, &testContext.fences[0], true, 0u); in testFences() 1046 …waitStatus = deviceInterface.waitForFences(device, 1, &testContext.fences[0], true, DEFAULT_TIMEOU… in testFences() 1055 …waitStatus = deviceInterface.waitForFences(device, 1, &testContext.fences[0], true, std::numeric_l… in testFences() 1063 waitStatus = deviceInterface.waitForFences(device, 1, &testContext.fences[1], true, 1); in testFences() [all …]
|
/external/mesa3d/src/gallium/winsys/radeon/drm/ |
D | radeon_drm_cs.c | 531 if (bo->u.slab.fences[src]->num_cs_references) { in radeon_bo_slab_fence() 532 bo->u.slab.fences[dst] = bo->u.slab.fences[src]; in radeon_bo_slab_fence() 535 radeon_bo_reference(&bo->u.slab.fences[src], NULL); in radeon_bo_slab_fence() 543 struct radeon_bo **new_fences = REALLOC(bo->u.slab.fences, in radeon_bo_slab_fence() 551 bo->u.slab.fences = new_fences; in radeon_bo_slab_fence() 556 bo->u.slab.fences[bo->u.slab.num_fences] = NULL; in radeon_bo_slab_fence() 557 radeon_bo_reference(&bo->u.slab.fences[bo->u.slab.num_fences], fence); in radeon_bo_slab_fence()
|
D | radeon_drm_bo.c | 82 if (radeon_real_bo_is_busy(bo->u.slab.fences[num_idle])) { in radeon_bo_is_busy() 86 radeon_bo_reference(&bo->u.slab.fences[num_idle], NULL); in radeon_bo_is_busy() 88 memmove(&bo->u.slab.fences[0], &bo->u.slab.fences[num_idle], in radeon_bo_is_busy() 89 (bo->u.slab.num_fences - num_idle) * sizeof(bo->u.slab.fences[0])); in radeon_bo_is_busy() 113 radeon_bo_reference(&fence, bo->u.slab.fences[0]); in radeon_bo_wait_idle() 120 if (bo->u.slab.num_fences && fence == bo->u.slab.fences[0]) { in radeon_bo_wait_idle() 121 radeon_bo_reference(&bo->u.slab.fences[0], NULL); in radeon_bo_wait_idle() 122 memmove(&bo->u.slab.fences[0], &bo->u.slab.fences[1], in radeon_bo_wait_idle() 123 (bo->u.slab.num_fences - 1) * sizeof(bo->u.slab.fences[0])); in radeon_bo_wait_idle() 837 radeon_bo_reference(&bo->u.slab.fences[j], NULL); in radeon_bo_slab_free() [all …]
|
D | radeon_drm_bo.h | 52 struct radeon_bo **fences; member
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | consecutive-fences.ll | 3 ; Make sure we collapse the fences in this case 68 !3 = !DIFile(filename: "consecutive-fences.ll", directory: "")
|
/external/deqp/external/vulkancts/modules/vulkan/sparse_resources/ |
D | vktSparseResourcesQueueBindSparseTests.cpp | 134 …aitForFences (const DeviceInterface& vk, const VkDevice device, const std::vector<FenceSp>& fences) in waitForFences() argument 136 …for (std::vector<FenceSp>::const_iterator fenceSpIter = fences.begin(); fenceSpIter != fences.end(… in waitForFences()
|
/external/mesa3d/src/gallium/drivers/zink/ |
D | zink_query.c | 33 unsigned fences; member 153 assert(!p_atomic_read(&query->fences)); in destroy_query() 168 if (p_atomic_read(&query->fences)) { in zink_destroy_query() 182 if (!p_atomic_dec_return(&query->fences)) { in zink_prune_queries() 356 p_atomic_inc(&q->fences); in begin_query()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/DeadStoreElimination/ |
D | fence.ll | 4 ; across release or stronger fences. It's not required 50 ; We DSE stack alloc'ed and byval locations, in the presence of fences.
|
/external/llvm/test/Transforms/DeadStoreElimination/ |
D | fence.ll | 4 ; across release or stronger fences. It's not required 50 ; We DSE stack alloc'ed and byval locations, in the presence of fences.
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Analysis/MemorySSA/ |
D | atomic-clobber.ll | 52 ; CHECK-LABEL: define void @fences 53 define void @fences(i32* %a) {
|