/external/ltp/testcases/kernel/mem/mtest07/ |
D | mallocstress.c | 154 int thread_index; in stress_malloc() local 156 for (thread_index = 0; thread_index < NUM_THREADS; thread_index++) { in stress_malloc() 157 SAFE_PTHREAD_CREATE(&thread_id[thread_index], NULL, alloc_mem, in stress_malloc() 158 (void *)(uintptr_t)thread_index); in stress_malloc() 165 for (thread_index = 0; thread_index < NUM_THREADS; thread_index++) { in stress_malloc() 168 SAFE_PTHREAD_JOIN(thread_id[thread_index], &status); in stress_malloc() 171 thread_index); in stress_malloc()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | bias_op_gpu.cu.cc | 186 int32 thread_index = threadIdx.x; in BiasGradNCHW_SharedAtomics() local 188 if (thread_index < 32) { in BiasGradNCHW_SharedAtomics() 189 AccT data = s_data[thread_index]; in BiasGradNCHW_SharedAtomics() 193 if (thread_index == 0) { in BiasGradNCHW_SharedAtomics() 198 if (thread_index < 16) s_data[thread_index] += s_data[thread_index + 16]; in BiasGradNCHW_SharedAtomics() 199 if (thread_index < 8) s_data[thread_index] += s_data[thread_index + 8]; in BiasGradNCHW_SharedAtomics() 200 if (thread_index < 4) s_data[thread_index] += s_data[thread_index + 4]; in BiasGradNCHW_SharedAtomics() 201 if (thread_index < 2) s_data[thread_index] += s_data[thread_index + 2]; in BiasGradNCHW_SharedAtomics() 202 if (thread_index < 1) s_data[thread_index] += s_data[thread_index + 1]; in BiasGradNCHW_SharedAtomics() 205 if (thread_index == 0) { in BiasGradNCHW_SharedAtomics()
|
/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
D | parallel_interleave_dataset_op.cc | 688 const int64 thread_index) { in WorkerThread() argument 710 auto cleanup = gtl::MakeCleanup([this, thread_index, ctx] { in WorkerThread() 712 workers_[thread_index].cond_var.notify_all(); in WorkerThread() 724 worker_thread_states_[thread_index].iterator == nullptr && in WorkerThread() 725 worker_thread_states_[thread_index].iterator_creation_status.ok(); in WorkerThread() 744 read_new_input = worker_thread_states_[thread_index].input.empty(); in WorkerThread() 749 while (!cancelled_ && !workers_[thread_index].is_producing) { in WorkerThread() 751 workers_[thread_index].cond_var.wait(l); in WorkerThread() 763 worker_thread_states_[thread_index].input.swap( in WorkerThread() 764 workers_[thread_index].input); in WorkerThread() [all …]
|
/external/llvm-project/lldb/test/API/functionalities/breakpoint/hardware_breakpoints/hardware_breakpoint_on_multiple_threads/ |
D | main.cpp | 12 hw_break_function (uint32_t thread_index) { in hw_break_function() argument 13 printf ("%s called by Thread #%u...\n", __FUNCTION__, thread_index); in hw_break_function() 18 thread_func (uint32_t thread_index) { in thread_func() argument 19 printf ("%s (thread index = %u) starting...\n", __FUNCTION__, thread_index); in thread_func() 23 hw_break_function(thread_index); // Call hw_break_function in thread_func()
|
/external/llvm-project/lldb/test/Shell/ExecControl/StopHook/Inputs/ |
D | stop-hook-threads.cpp | 30 thread_func (uint32_t thread_index) in thread_func() argument 33 printf ("%s (thread index = %u) startng...\n", __FUNCTION__, thread_index); in thread_func() 41 printf ("%s (thread = %u) doing a usleep (%d)...\n", __FUNCTION__, thread_index, usec); in thread_func() 49 … %u) after usleep access_pool returns %d (count=%d)...\n", __FUNCTION__, thread_index, val, count); in thread_func() 51 printf ("%s (thread index = %u) exiting...\n", __FUNCTION__, thread_index); in thread_func()
|
/external/llvm-project/lldb/test/API/python_api/lldbutil/iter/ |
D | main.cpp | 50 uint32_t thread_index = *((uint32_t *)arg); in thread_func() local 51 uint32_t thread_mask = (1u << (thread_index)); in thread_func() 52 printf ("%s (thread index = %u) startng...\n", __FUNCTION__, thread_index); in thread_func() 61 printf ("%s (thread = %u) doing a usleep (%d)...\n", __FUNCTION__, thread_index, usec); in thread_func() 65 …printf ("%s (thread = %u) after usleep ...\n", __FUNCTION__, thread_index); // Set break point at … in thread_func() 67 printf ("%s (thread index = %u) exiting...\n", __FUNCTION__, thread_index); in thread_func()
|
/external/llvm-project/lldb/test/API/python_api/watchpoint/watchlocation/ |
D | main.cpp | 48 thread_func (uint32_t thread_index) in thread_func() argument 50 printf ("%s (thread index = %u) startng...\n", __FUNCTION__, thread_index); in thread_func() 60 printf ("%s (thread = %u) doing a usleep (%d)...\n", __FUNCTION__, thread_index, usec); in thread_func() 68 … %u) after usleep access_pool returns %d (count=%d)...\n", __FUNCTION__, thread_index, val, count); in thread_func() 70 printf ("%s (thread index = %u) exiting...\n", __FUNCTION__, thread_index); in thread_func()
|
/external/llvm-project/lldb/test/API/commands/watchpoints/hello_watchlocation/ |
D | main.cpp | 48 thread_func (uint32_t thread_index) in thread_func() argument 50 printf ("%s (thread index = %u) startng...\n", __FUNCTION__, thread_index); in thread_func() 60 printf ("%s (thread = %u) doing a usleep (%d)...\n", __FUNCTION__, thread_index, usec); in thread_func() 68 … %u) after usleep access_pool returns %d (count=%d)...\n", __FUNCTION__, thread_index, val, count); in thread_func() 70 printf ("%s (thread index = %u) exiting...\n", __FUNCTION__, thread_index); in thread_func()
|
/external/llvm-project/lldb/test/API/commands/watchpoints/watchpoint_set_command/ |
D | main.cpp | 56 thread_func (uint32_t thread_index) in thread_func() argument 58 printf ("%s (thread index = %u) startng...\n", __FUNCTION__, thread_index); in thread_func() 66 printf ("%s (thread = %u) sleeping for 1 second...\n", __FUNCTION__, thread_index); in thread_func() 74 …= %u) after sleep access_pool returns %d (count=%d)...\n", __FUNCTION__, thread_index, val, count); in thread_func() 76 printf ("%s (thread index = %u) exiting...\n", __FUNCTION__, thread_index); in thread_func()
|
/external/compiler-rt/lib/sanitizer_common/tests/ |
D | sanitizer_stoptheworld_test.cc | 98 volatile uptr thread_index; member 105 : thread_index(0), in AdvancedCallbackArgument() 115 &callback_argument->thread_index, 1); in AdvancedIncrementerThread() 126 __sync_fetch_and_add(&callback_argument->thread_index, in AdvancedIncrementerThread() 127 kThreadCount - callback_argument->thread_index); in AdvancedIncrementerThread() 175 while (__sync_fetch_and_add(&argument.thread_index, 0) < kStopWorldAfter) in TEST() 182 while (__sync_fetch_and_add(&argument.thread_index, 0) < kThreadCount) in TEST()
|
/external/llvm-project/compiler-rt/lib/sanitizer_common/tests/ |
D | sanitizer_stoptheworld_test.cpp | 97 volatile uptr thread_index; member 104 : thread_index(0), in AdvancedCallbackArgument() 114 &callback_argument->thread_index, 1); in AdvancedIncrementerThread() 125 __sync_fetch_and_add(&callback_argument->thread_index, in AdvancedIncrementerThread() 126 kThreadCount - callback_argument->thread_index); in AdvancedIncrementerThread() 174 while (__sync_fetch_and_add(&argument.thread_index, 0) < kStopWorldAfter) in TEST() 181 while (__sync_fetch_and_add(&argument.thread_index, 0) < kThreadCount) in TEST()
|
/external/ruy/ruy/ |
D | ctx.cc | 181 TuningResolver* Ctx::GetThreadSpecificTuningResolver(int thread_index) const { in GetThreadSpecificTuningResolver() 183 RUY_DCHECK_LT(thread_index, static_cast<int>(resources.size())); in GetThreadSpecificTuningResolver() 184 return &resources[thread_index]->tuning_resolver; in GetThreadSpecificTuningResolver() 187 Allocator* Ctx::GetThreadSpecificAllocator(int thread_index) const { in GetThreadSpecificAllocator() 189 RUY_DCHECK_LT(thread_index, static_cast<int>(resources.size())); in GetThreadSpecificAllocator() 190 return &resources[thread_index]->allocator; in GetThreadSpecificAllocator()
|
D | ctx.h | 76 TuningResolver* GetThreadSpecificTuningResolver(int thread_index) const; 77 Allocator* GetThreadSpecificAllocator(int thread_index) const;
|
/external/mesa3d/src/util/ |
D | u_queue.c | 247 int thread_index; member 254 int thread_index = ((struct thread_input*)input)->thread_index; in util_queue_thread_func() local 277 snprintf(name, sizeof(name), "%s%i", queue->name, thread_index); in util_queue_thread_func() 288 while (thread_index < queue->num_threads && queue->num_queued == 0) in util_queue_thread_func() 292 if (thread_index >= queue->num_threads) { in util_queue_thread_func() 308 job.execute(job.job, thread_index); in util_queue_thread_func() 311 job.cleanup(job.job, thread_index); in util_queue_thread_func() 338 input->thread_index = index; in util_queue_create_thread() 693 util_queue_get_thread_time_nano(struct util_queue *queue, unsigned thread_index) in util_queue_get_thread_time_nano() argument 696 if (thread_index >= queue->num_threads) in util_queue_get_thread_time_nano() [all …]
|
/external/mesa3d/src/gallium/drivers/llvmpipe/ |
D | lp_rast.c | 526 pq->start[task->thread_index] = task->thread_data.vis_counter; in lp_rast_begin_query() 529 pq->start[task->thread_index] = task->thread_data.ps_invocations; in lp_rast_begin_query() 553 pq->end[task->thread_index] += in lp_rast_end_query() 554 task->thread_data.vis_counter - pq->start[task->thread_index]; in lp_rast_end_query() 555 pq->start[task->thread_index] = 0; in lp_rast_end_query() 558 pq->end[task->thread_index] = os_time_get_nano(); in lp_rast_end_query() 561 pq->end[task->thread_index] += in lp_rast_end_query() 562 task->thread_data.ps_invocations - pq->start[task->thread_index]; in lp_rast_end_query() 563 pq->start[task->thread_index] = 0; in lp_rast_end_query() 751 task->thread_index, (long long unsigned)total, in rasterize_scene() [all …]
|
/external/llvm-project/llvm/utils/benchmark/test/ |
D | fixture_test.cc | 10 if (state.thread_index == 0) { in SetUp() 17 if (state.thread_index == 0) { in TearDown() 36 if (st.thread_index == 0) { in BENCHMARK_DEFINE_F()
|
D | benchmark_test.cc | 129 if (state.thread_index == 0) { in BM_SetupTeardown() 142 if (state.thread_index == 0) { in BM_SetupTeardown() 160 int from = thread_size * state.thread_index; in BM_ParallelMemset() 163 if (state.thread_index == 0) { in BM_ParallelMemset() 175 if (state.thread_index == 0) { in BM_ParallelMemset()
|
/external/llvm-project/libcxx/utils/google-benchmark/test/ |
D | fixture_test.cc | 10 if (state.thread_index == 0) { in SetUp() 17 if (state.thread_index == 0) { in TearDown() 36 if (st.thread_index == 0) { in BENCHMARK_DEFINE_F()
|
D | benchmark_test.cc | 129 if (state.thread_index == 0) { in BM_SetupTeardown() 142 if (state.thread_index == 0) { in BM_SetupTeardown() 160 int from = thread_size * state.thread_index; in BM_ParallelMemset() 163 if (state.thread_index == 0) { in BM_ParallelMemset() 175 if (state.thread_index == 0) { in BM_ParallelMemset()
|
/external/libcxx/utils/google-benchmark/test/ |
D | fixture_test.cc | 10 if (state.thread_index == 0) { in SetUp() 17 if (state.thread_index == 0) { in TearDown() 36 if (st.thread_index == 0) { in BENCHMARK_DEFINE_F()
|
D | benchmark_test.cc | 129 if (state.thread_index == 0) { in BM_SetupTeardown() 142 if (state.thread_index == 0) { in BM_SetupTeardown() 160 int from = thread_size * state.thread_index; in BM_ParallelMemset() 163 if (state.thread_index == 0) { in BM_ParallelMemset() 175 if (state.thread_index == 0) { in BM_ParallelMemset()
|
/external/google-benchmark/test/ |
D | fixture_test.cc | 12 if (state.thread_index == 0) { in SetUp() 19 if (state.thread_index == 0) { in TearDown() 38 if (st.thread_index == 0) { in BENCHMARK_DEFINE_F()
|
D | benchmark_test.cc | 129 if (state.thread_index == 0) { in BM_SetupTeardown() 142 if (state.thread_index == 0) { in BM_SetupTeardown() 160 int from = thread_size * state.thread_index; in BM_ParallelMemset() 163 if (state.thread_index == 0) { in BM_ParallelMemset() 175 if (state.thread_index == 0) { in BM_ParallelMemset()
|
/external/llvm-project/lldb/test/API/functionalities/breakpoint/consecutive_breakpoints/ |
D | TestConsecutiveBreakpoints.py | 86 thread_index = self.process.GetNumThreads() + 1 87 self.assertFalse(self.process.GetThreadAtIndex(thread_index).IsValid()) 88 self.breakpoint2.SetThreadIndex(thread_index)
|
/external/mesa3d/src/broadcom/compiler/ |
D | vir_register_allocate.c | 528 int thread_index = ffs(c->threads) - 1; in v3d_register_allocate() local 530 if (thread_index >= 1) in v3d_register_allocate() 531 thread_index--; in v3d_register_allocate() 680 c->compiler->reg_class_phys[thread_index]); in v3d_register_allocate() 683 c->compiler->reg_class_r5[thread_index]); in v3d_register_allocate() 686 c->compiler->reg_class_phys_or_acc[thread_index]); in v3d_register_allocate() 690 c->compiler->reg_class_any[thread_index]); in v3d_register_allocate() 739 if (is_uniform || thread_index == 0) { in v3d_register_allocate()
|