Home
last modified time | relevance | path

Searched refs:thread_index (Results 1 – 25 of 44) sorted by relevance

12

/third_party/ltp/testcases/kernel/mem/mtest07/
Dmallocstress.c154 int thread_index; in stress_malloc() local
156 for (thread_index = 0; thread_index < NUM_THREADS; thread_index++) { in stress_malloc()
157 SAFE_PTHREAD_CREATE(&thread_id[thread_index], NULL, alloc_mem, in stress_malloc()
158 (void *)(uintptr_t)thread_index); in stress_malloc()
165 for (thread_index = 0; thread_index < NUM_THREADS; thread_index++) { in stress_malloc()
168 SAFE_PTHREAD_JOIN(thread_id[thread_index], &status); in stress_malloc()
171 thread_index); in stress_malloc()
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/
Dgather_cpu_kernel.cc91 int thread_index = 0; in ParallelRun() local
92 while (thread_index < thread_num) { in ParallelRun()
93 int count = SizeToInt(MSMIN(stride, outer_size - stride * IntToSize(thread_index))); in ParallelRun()
97 auto thread_stride = static_cast<size_t>(stride * thread_index); in ParallelRun()
100 …k = [this, in, indices_data, count, inner_size, limit, indices_element_size, out, thread_index]() { in ParallelRun()
103 … MS_LOG(ERROR) << "GatherRun error task_id[" << thread_index << "] error_code[" << ret << "]"; in ParallelRun()
109 thread_index++; in ParallelRun()
Dassign_cpu_kernel.cc79 size_t thread_index = 0; in Launch() local
84 while (thread_index < thread_num) { in Launch()
85 auto thread_stride = stride * thread_index; in Launch()
87 if (thread_index == thread_num - 1) { in Launch()
107 thread_index++; in Launch()
Dstridedslice_cpu_kernel.cc195 int thread_index = 0; in ParallelRun() local
206 while (thread_index < thread_num) { in ParallelRun()
208 std::bind(execute_func, this, input_addr, output_addr, thread_index * cal_num_per_thread_)); in ParallelRun()
209 thread_index++; in ParallelRun()
/third_party/mesa3d/src/util/
Du_queue.c247 int thread_index; member
254 int thread_index = ((struct thread_input*)input)->thread_index; in util_queue_thread_func() local
282 snprintf(name, sizeof(name), "%s%i", queue->name, thread_index); in util_queue_thread_func()
293 while (thread_index < queue->num_threads && queue->num_queued == 0) in util_queue_thread_func()
297 if (thread_index >= queue->num_threads) { in util_queue_thread_func()
313 job.execute(job.job, job.global_data, thread_index); in util_queue_thread_func()
317 job.cleanup(job.job, job.global_data, thread_index); in util_queue_thread_func()
345 input->thread_index = index; in util_queue_create_thread()
714 util_queue_get_thread_time_nano(struct util_queue *queue, unsigned thread_index) in util_queue_get_thread_time_nano() argument
717 if (thread_index >= queue->num_threads) in util_queue_get_thread_time_nano()
[all …]
Ddisk_cache.c314 destroy_put_job(void *job, void *gdata, int thread_index) in destroy_put_job() argument
324 destroy_put_job_nocopy(void *job, void *gdata, int thread_index) in destroy_put_job_nocopy() argument
328 destroy_put_job(job, gdata, thread_index); in destroy_put_job_nocopy()
332 cache_put(void *job, void *gdata, int thread_index) in cache_put() argument
Du_queue.h194 typedef void (*util_queue_execute_func)(void *job, void *gdata, int thread_index);
255 unsigned thread_index);
/third_party/benchmark/test/
Dfixture_test.cc12 if (state.thread_index() == 0) { in SetUp()
19 if (state.thread_index() == 0) { in TearDown()
38 if (st.thread_index() == 0) { in BENCHMARK_DEFINE_F()
Dbenchmark_setup_teardown_test.cc21 assert(state.thread_index() == 0); in DoSetup1()
26 assert(state.thread_index() == 0); in DoTeardown1()
51 assert(state.thread_index() == 0); in DoSetup2()
56 assert(state.thread_index() == 0); in DoTeardown2()
Dbenchmark_test.cc130 if (state.thread_index() == 0) { in BM_SetupTeardown()
143 if (state.thread_index() == 0) { in BM_SetupTeardown()
161 int from = thread_size * state.thread_index(); in BM_ParallelMemset()
164 if (state.thread_index() == 0) { in BM_ParallelMemset()
176 if (state.thread_index() == 0) { in BM_ParallelMemset()
Dskip_with_error_test.cc100 if (state.range(0) == 1 && state.thread_index() <= (state.threads() / 2)) { in BM_error_during_running()
146 if (state.thread_index() <= (state.threads() / 2)) in BM_error_after_running()
158 if (state.range(0) == 1 && state.thread_index() <= (state.threads() / 2)) { in BM_error_while_paused()
/third_party/mesa3d/src/gallium/drivers/llvmpipe/
Dlp_rast.c646 pq->start[task->thread_index] = task->thread_data.vis_counter; in lp_rast_begin_query()
649 pq->start[task->thread_index] = task->thread_data.ps_invocations; in lp_rast_begin_query()
652 pq->start[task->thread_index] = os_time_get_nano(); in lp_rast_begin_query()
676 pq->end[task->thread_index] += in lp_rast_end_query()
677 task->thread_data.vis_counter - pq->start[task->thread_index]; in lp_rast_end_query()
678 pq->start[task->thread_index] = 0; in lp_rast_end_query()
682 pq->end[task->thread_index] = os_time_get_nano(); in lp_rast_end_query()
685 pq->end[task->thread_index] += in lp_rast_end_query()
686 task->thread_data.ps_invocations - pq->start[task->thread_index]; in lp_rast_end_query()
687 pq->start[task->thread_index] = 0; in lp_rast_end_query()
[all …]
/third_party/mesa3d/src/broadcom/compiler/
Dvir_register_allocate.c572 tmu_spilling_allowed(struct v3d_compile *c, int thread_index) in tmu_spilling_allowed() argument
574 return thread_index == 0 && c->tmu_spilling_allowed; in tmu_spilling_allowed()
615 int thread_index = ffs(c->threads) - 1; in v3d_register_allocate() local
617 if (thread_index >= 1) in v3d_register_allocate()
618 thread_index--; in v3d_register_allocate()
767 c->compiler->reg_class_phys[thread_index]); in v3d_register_allocate()
770 c->compiler->reg_class_r5[thread_index]); in v3d_register_allocate()
773 c->compiler->reg_class_phys_or_acc[thread_index]); in v3d_register_allocate()
777 c->compiler->reg_class_any[thread_index]); in v3d_register_allocate()
823 if (is_uniform || tmu_spilling_allowed(c, thread_index)) { in v3d_register_allocate()
/third_party/boost/boost/math/quadrature/
Dnaive_monte_carlo.hpp368 void m_thread_monte(uint64_t thread_index, uint64_t seed) in m_thread_monte() argument
376 Real M1 = m_thread_averages[thread_index].load(boost::memory_order::consume); in m_thread_monte()
377 Real S = m_thread_Ss[thread_index].load(boost::memory_order::consume); in m_thread_monte()
383 uint64_t k = m_thread_calls[thread_index].load(boost::memory_order::consume); in m_thread_monte()
421 m_thread_averages[thread_index].store(M1, boost::memory_order::release); in m_thread_monte()
422 m_thread_Ss[thread_index].store(S, boost::memory_order::release); in m_thread_monte()
423 m_thread_calls[thread_index].store(k, boost::memory_order::release); in m_thread_monte()
/third_party/boost/libs/histogram/benchmark/
Dhistogram_parallel_filling.cpp53 if (state.thread_index == 0) { in AtomicStorage()
58 std::default_random_engine gen(state.thread_index); in AtomicStorage()
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/mkldnn/
Dmatmul_cpu_kernel.cc230 size_t thread_index = 0; in ParallelRun() local
231 while (thread_index < thread_count_) { in ParallelRun()
232 (void)tasks.emplace_back(std::bind(&MatMulCPUKernel::FloatRun, this, thread_index)); in ParallelRun()
233 thread_index++; in ParallelRun()
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/
Dmutex_benchmark.cc100 (multiple_priorities && state.thread_index != 0) ? 1 : 0); in BM_MutexEnqueue()
179 int priority = state.thread_index % state.range(1); in BM_Contended()
/third_party/grpc/test/cpp/microbenchmarks/
Dbm_threadpool.cc155 if (state.thread_index == 0) { in BM_ThreadPoolExternalAdd()
169 if (state.thread_index == 0) { in BM_ThreadPoolExternalAdd()
Dbm_cq_multiple_threads.cc161 auto thd_idx = state.thread_index; in BM_Cq_Throughput()
/third_party/mesa3d/src/mesa/main/
Dglthread.c45 glthread_unmarshal_batch(void *job, void *gdata, int thread_index) in glthread_unmarshal_batch() argument
83 glthread_thread_initialization(void *job, void *gdata, int thread_index) in glthread_thread_initialization() argument
/third_party/mesa3d/src/gallium/winsys/radeon/drm/
Dradeon_drm_cs.h140 void radeon_drm_cs_emit_ioctl_oneshot(void *job, void *gdata, int thread_index);
/third_party/mesa3d/src/gallium/drivers/radeonsi/
Dsi_state_shaders.c2200 static void si_build_shader_variant(struct si_shader *shader, int thread_index, bool low_priority) in si_build_shader_variant() argument
2207 if (thread_index >= 0) { in si_build_shader_variant()
2209 assert(thread_index < ARRAY_SIZE(sscreen->compiler_lowp)); in si_build_shader_variant()
2210 compiler = &sscreen->compiler_lowp[thread_index]; in si_build_shader_variant()
2212 assert(thread_index < ARRAY_SIZE(sscreen->compiler)); in si_build_shader_variant()
2213 compiler = &sscreen->compiler[thread_index]; in si_build_shader_variant()
2242 static void si_build_shader_variant_low_priority(void *job, void *gdata, int thread_index) in si_build_shader_variant_low_priority() argument
2246 assert(thread_index >= 0); in si_build_shader_variant_low_priority()
2248 si_build_shader_variant(shader, thread_index, true); in si_build_shader_variant_low_priority()
2304 const struct si_shader_key *key, int thread_index, in si_shader_select_with_key() argument
[all …]
Dsi_compute.c110 static void si_create_compute_state_async(void *job, void *gdata, int thread_index) in si_create_compute_state_async() argument
120 assert(thread_index >= 0); in si_create_compute_state_async()
121 assert(thread_index < ARRAY_SIZE(sscreen->compiler)); in si_create_compute_state_async()
122 compiler = &sscreen->compiler[thread_index]; in si_create_compute_state_async()
/third_party/mesa3d/src/util/perf/
Du_trace.c322 process_chunk(void *job, void *gdata, int thread_index) in process_chunk() argument
390 cleanup_chunk(void *job, void *gdata, int thread_index) in cleanup_chunk() argument
/third_party/mesa3d/src/gallium/drivers/freedreno/ir3/
Dir3_gallium.c245 create_initial_variants_async(void *job, void *gdata, int thread_index) in create_initial_variants_async() argument
254 create_initial_compute_variants_async(void *job, void *gdata, int thread_index) in create_initial_compute_variants_async() argument

12