/third_party/ltp/testcases/kernel/mem/mtest07/ |
D | mallocstress.c | 154 int thread_index; in stress_malloc() local 156 for (thread_index = 0; thread_index < NUM_THREADS; thread_index++) { in stress_malloc() 157 SAFE_PTHREAD_CREATE(&thread_id[thread_index], NULL, alloc_mem, in stress_malloc() 158 (void *)(uintptr_t)thread_index); in stress_malloc() 165 for (thread_index = 0; thread_index < NUM_THREADS; thread_index++) { in stress_malloc() 168 SAFE_PTHREAD_JOIN(thread_id[thread_index], &status); in stress_malloc() 171 thread_index); in stress_malloc()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ |
D | gather_cpu_kernel.cc | 91 int thread_index = 0; in ParallelRun() local 92 while (thread_index < thread_num) { in ParallelRun() 93 int count = SizeToInt(MSMIN(stride, outer_size - stride * IntToSize(thread_index))); in ParallelRun() 97 auto thread_stride = static_cast<size_t>(stride * thread_index); in ParallelRun() 100 …k = [this, in, indices_data, count, inner_size, limit, indices_element_size, out, thread_index]() { in ParallelRun() 103 … MS_LOG(ERROR) << "GatherRun error task_id[" << thread_index << "] error_code[" << ret << "]"; in ParallelRun() 109 thread_index++; in ParallelRun()
|
D | assign_cpu_kernel.cc | 79 size_t thread_index = 0; in Launch() local 84 while (thread_index < thread_num) { in Launch() 85 auto thread_stride = stride * thread_index; in Launch() 87 if (thread_index == thread_num - 1) { in Launch() 107 thread_index++; in Launch()
|
D | stridedslice_cpu_kernel.cc | 195 int thread_index = 0; in ParallelRun() local 206 while (thread_index < thread_num) { in ParallelRun() 208 std::bind(execute_func, this, input_addr, output_addr, thread_index * cal_num_per_thread_)); in ParallelRun() 209 thread_index++; in ParallelRun()
|
/third_party/mesa3d/src/util/ |
D | u_queue.c | 247 int thread_index; member 254 int thread_index = ((struct thread_input*)input)->thread_index; in util_queue_thread_func() local 282 snprintf(name, sizeof(name), "%s%i", queue->name, thread_index); in util_queue_thread_func() 293 while (thread_index < queue->num_threads && queue->num_queued == 0) in util_queue_thread_func() 297 if (thread_index >= queue->num_threads) { in util_queue_thread_func() 313 job.execute(job.job, job.global_data, thread_index); in util_queue_thread_func() 317 job.cleanup(job.job, job.global_data, thread_index); in util_queue_thread_func() 345 input->thread_index = index; in util_queue_create_thread() 714 util_queue_get_thread_time_nano(struct util_queue *queue, unsigned thread_index) in util_queue_get_thread_time_nano() argument 717 if (thread_index >= queue->num_threads) in util_queue_get_thread_time_nano() [all …]
|
D | disk_cache.c | 314 destroy_put_job(void *job, void *gdata, int thread_index) in destroy_put_job() argument 324 destroy_put_job_nocopy(void *job, void *gdata, int thread_index) in destroy_put_job_nocopy() argument 328 destroy_put_job(job, gdata, thread_index); in destroy_put_job_nocopy() 332 cache_put(void *job, void *gdata, int thread_index) in cache_put() argument
|
D | u_queue.h | 194 typedef void (*util_queue_execute_func)(void *job, void *gdata, int thread_index); 255 unsigned thread_index);
|
/third_party/benchmark/test/ |
D | fixture_test.cc | 12 if (state.thread_index() == 0) { in SetUp() 19 if (state.thread_index() == 0) { in TearDown() 38 if (st.thread_index() == 0) { in BENCHMARK_DEFINE_F()
|
D | benchmark_setup_teardown_test.cc | 21 assert(state.thread_index() == 0); in DoSetup1() 26 assert(state.thread_index() == 0); in DoTeardown1() 51 assert(state.thread_index() == 0); in DoSetup2() 56 assert(state.thread_index() == 0); in DoTeardown2()
|
D | benchmark_test.cc | 130 if (state.thread_index() == 0) { in BM_SetupTeardown() 143 if (state.thread_index() == 0) { in BM_SetupTeardown() 161 int from = thread_size * state.thread_index(); in BM_ParallelMemset() 164 if (state.thread_index() == 0) { in BM_ParallelMemset() 176 if (state.thread_index() == 0) { in BM_ParallelMemset()
|
D | skip_with_error_test.cc | 100 if (state.range(0) == 1 && state.thread_index() <= (state.threads() / 2)) { in BM_error_during_running() 146 if (state.thread_index() <= (state.threads() / 2)) in BM_error_after_running() 158 if (state.range(0) == 1 && state.thread_index() <= (state.threads() / 2)) { in BM_error_while_paused()
|
/third_party/mesa3d/src/gallium/drivers/llvmpipe/ |
D | lp_rast.c | 646 pq->start[task->thread_index] = task->thread_data.vis_counter; in lp_rast_begin_query() 649 pq->start[task->thread_index] = task->thread_data.ps_invocations; in lp_rast_begin_query() 652 pq->start[task->thread_index] = os_time_get_nano(); in lp_rast_begin_query() 676 pq->end[task->thread_index] += in lp_rast_end_query() 677 task->thread_data.vis_counter - pq->start[task->thread_index]; in lp_rast_end_query() 678 pq->start[task->thread_index] = 0; in lp_rast_end_query() 682 pq->end[task->thread_index] = os_time_get_nano(); in lp_rast_end_query() 685 pq->end[task->thread_index] += in lp_rast_end_query() 686 task->thread_data.ps_invocations - pq->start[task->thread_index]; in lp_rast_end_query() 687 pq->start[task->thread_index] = 0; in lp_rast_end_query() [all …]
|
/third_party/mesa3d/src/broadcom/compiler/ |
D | vir_register_allocate.c | 572 tmu_spilling_allowed(struct v3d_compile *c, int thread_index) in tmu_spilling_allowed() argument 574 return thread_index == 0 && c->tmu_spilling_allowed; in tmu_spilling_allowed() 615 int thread_index = ffs(c->threads) - 1; in v3d_register_allocate() local 617 if (thread_index >= 1) in v3d_register_allocate() 618 thread_index--; in v3d_register_allocate() 767 c->compiler->reg_class_phys[thread_index]); in v3d_register_allocate() 770 c->compiler->reg_class_r5[thread_index]); in v3d_register_allocate() 773 c->compiler->reg_class_phys_or_acc[thread_index]); in v3d_register_allocate() 777 c->compiler->reg_class_any[thread_index]); in v3d_register_allocate() 823 if (is_uniform || tmu_spilling_allowed(c, thread_index)) { in v3d_register_allocate()
|
/third_party/boost/boost/math/quadrature/ |
D | naive_monte_carlo.hpp | 368 void m_thread_monte(uint64_t thread_index, uint64_t seed) in m_thread_monte() argument 376 Real M1 = m_thread_averages[thread_index].load(boost::memory_order::consume); in m_thread_monte() 377 Real S = m_thread_Ss[thread_index].load(boost::memory_order::consume); in m_thread_monte() 383 uint64_t k = m_thread_calls[thread_index].load(boost::memory_order::consume); in m_thread_monte() 421 m_thread_averages[thread_index].store(M1, boost::memory_order::release); in m_thread_monte() 422 m_thread_Ss[thread_index].store(S, boost::memory_order::release); in m_thread_monte() 423 m_thread_calls[thread_index].store(k, boost::memory_order::release); in m_thread_monte()
|
/third_party/boost/libs/histogram/benchmark/ |
D | histogram_parallel_filling.cpp | 53 if (state.thread_index == 0) { in AtomicStorage() 58 std::default_random_engine gen(state.thread_index); in AtomicStorage()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/mkldnn/ |
D | matmul_cpu_kernel.cc | 230 size_t thread_index = 0; in ParallelRun() local 231 while (thread_index < thread_count_) { in ParallelRun() 232 (void)tasks.emplace_back(std::bind(&MatMulCPUKernel::FloatRun, this, thread_index)); in ParallelRun() 233 thread_index++; in ParallelRun()
|
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/ |
D | mutex_benchmark.cc | 100 (multiple_priorities && state.thread_index != 0) ? 1 : 0); in BM_MutexEnqueue() 179 int priority = state.thread_index % state.range(1); in BM_Contended()
|
/third_party/grpc/test/cpp/microbenchmarks/ |
D | bm_threadpool.cc | 155 if (state.thread_index == 0) { in BM_ThreadPoolExternalAdd() 169 if (state.thread_index == 0) { in BM_ThreadPoolExternalAdd()
|
D | bm_cq_multiple_threads.cc | 161 auto thd_idx = state.thread_index; in BM_Cq_Throughput()
|
/third_party/mesa3d/src/mesa/main/ |
D | glthread.c | 45 glthread_unmarshal_batch(void *job, void *gdata, int thread_index) in glthread_unmarshal_batch() argument 83 glthread_thread_initialization(void *job, void *gdata, int thread_index) in glthread_thread_initialization() argument
|
/third_party/mesa3d/src/gallium/winsys/radeon/drm/ |
D | radeon_drm_cs.h | 140 void radeon_drm_cs_emit_ioctl_oneshot(void *job, void *gdata, int thread_index);
|
/third_party/mesa3d/src/gallium/drivers/radeonsi/ |
D | si_state_shaders.c | 2200 static void si_build_shader_variant(struct si_shader *shader, int thread_index, bool low_priority) in si_build_shader_variant() argument 2207 if (thread_index >= 0) { in si_build_shader_variant() 2209 assert(thread_index < ARRAY_SIZE(sscreen->compiler_lowp)); in si_build_shader_variant() 2210 compiler = &sscreen->compiler_lowp[thread_index]; in si_build_shader_variant() 2212 assert(thread_index < ARRAY_SIZE(sscreen->compiler)); in si_build_shader_variant() 2213 compiler = &sscreen->compiler[thread_index]; in si_build_shader_variant() 2242 static void si_build_shader_variant_low_priority(void *job, void *gdata, int thread_index) in si_build_shader_variant_low_priority() argument 2246 assert(thread_index >= 0); in si_build_shader_variant_low_priority() 2248 si_build_shader_variant(shader, thread_index, true); in si_build_shader_variant_low_priority() 2304 const struct si_shader_key *key, int thread_index, in si_shader_select_with_key() argument [all …]
|
D | si_compute.c | 110 static void si_create_compute_state_async(void *job, void *gdata, int thread_index) in si_create_compute_state_async() argument 120 assert(thread_index >= 0); in si_create_compute_state_async() 121 assert(thread_index < ARRAY_SIZE(sscreen->compiler)); in si_create_compute_state_async() 122 compiler = &sscreen->compiler[thread_index]; in si_create_compute_state_async()
|
/third_party/mesa3d/src/util/perf/ |
D | u_trace.c | 322 process_chunk(void *job, void *gdata, int thread_index) in process_chunk() argument 390 cleanup_chunk(void *job, void *gdata, int thread_index) in cleanup_chunk() argument
|
/third_party/mesa3d/src/gallium/drivers/freedreno/ir3/ |
D | ir3_gallium.c | 245 create_initial_variants_async(void *job, void *gdata, int thread_index) in create_initial_variants_async() argument 254 create_initial_compute_variants_async(void *job, void *gdata, int thread_index) in create_initial_compute_variants_async() argument
|