/external/llvm-project/openmp/runtime/test/api/ |
D | omp_pause_resource.c | 6 int fails, nthreads, my_dev; in test_omp_pause_resource() local 9 nthreads = 0; in test_omp_pause_resource() 14 nthreads = omp_get_num_threads(); in test_omp_pause_resource() 19 #pragma omp parallel shared(nthreads) in test_omp_pause_resource() 21 nthreads = omp_get_num_threads(); in test_omp_pause_resource() 23 if (nthreads == 0) in test_omp_pause_resource() 27 nthreads = 0; in test_omp_pause_resource() 29 #pragma omp parallel shared(nthreads) in test_omp_pause_resource() 31 nthreads = omp_get_num_threads(); in test_omp_pause_resource() 33 if (nthreads == 0) in test_omp_pause_resource() [all …]
|
D | kmp_set_defaults_lock_bug.c | 15 int nthreads = 0; in test_kmp_set_defaults_lock_bug() local 22 nthreads++; in test_kmp_set_defaults_lock_bug() 33 nthreads++; in test_kmp_set_defaults_lock_bug() 37 return (nthreads == 2*nthreads_lib); in test_kmp_set_defaults_lock_bug()
|
D | omp_get_num_threads.c | 10 int nthreads = 0; in test_omp_get_num_threads() local 18 nthreads++; in test_omp_get_num_threads() 25 return (nthreads == nthreads_lib); in test_omp_get_num_threads()
|
/external/python/cpython3/Tools/ccbench/ |
D | ccbench.py | 187 def run_throughput_test(func, args, nthreads): argument 188 assert nthreads >= 1 197 if nthreads == 1: 221 for i in range(nthreads): 229 while len(ready) < nthreads: 234 start_cond.notify(nthreads) 245 nthreads = 1 247 while nthreads <= max_threads: 248 results = run_throughput_test(func, args, nthreads) 252 print("threads=%d: %d" % (nthreads, speed), end="") [all …]
|
/external/ltp/testcases/realtime/perf/latency/ |
D | pthread_cond_many.c | 55 int nthreads = 0; variable 186 void test_signal(long iter, long nthreads) in test_signal() argument 197 stats_container_init(&dat, iter * nthreads); in test_signal() 199 pt = malloc(sizeof(*pt) * nthreads); in test_signal() 204 for (j = 0; j < nthreads; j++) { in test_signal() 208 for (i = 0; i < (iter - 1) * nthreads; i += nthreads) { in test_signal() 209 for (j = 0, k = i; j < nthreads; j++, k++) { in test_signal() 219 for (j = 0; j < nthreads; j++) { in test_signal() 231 for (i = 0; i < iter * nthreads; i++) { in test_signal() 271 nthreads = atoi(v); in parse_args() [all …]
|
/external/python/cpython2/Tools/ccbench/ |
D | ccbench.py | 196 def run_throughput_test(func, args, nthreads): argument 197 assert nthreads >= 1 206 if nthreads == 1: 230 for i in range(nthreads): 238 while len(ready) < nthreads: 243 start_cond.notify(nthreads) 254 nthreads = 1 256 while nthreads <= max_threads: 257 results = run_throughput_test(func, args, nthreads) 261 print("threads=%d: %d" % (nthreads, speed), end="") [all …]
|
/external/eigen/test/ |
D | cuda_basic.cu | 146 int nthreads = 100; in test_cuda_basic() local 150 int data_size = nthreads * 512; in test_cuda_basic() 155 CALL_SUBTEST( run_and_compare_to_cuda(coeff_wise<Vector3f>(), nthreads, in, out) ); in test_cuda_basic() 156 CALL_SUBTEST( run_and_compare_to_cuda(coeff_wise<Array44f>(), nthreads, in, out) ); in test_cuda_basic() 158 CALL_SUBTEST( run_and_compare_to_cuda(replicate<Array4f>(), nthreads, in, out) ); in test_cuda_basic() 159 CALL_SUBTEST( run_and_compare_to_cuda(replicate<Array33f>(), nthreads, in, out) ); in test_cuda_basic() 161 CALL_SUBTEST( run_and_compare_to_cuda(redux<Array4f>(), nthreads, in, out) ); in test_cuda_basic() 162 CALL_SUBTEST( run_and_compare_to_cuda(redux<Matrix3f>(), nthreads, in, out) ); in test_cuda_basic() 164 CALL_SUBTEST( run_and_compare_to_cuda(prod_test<Matrix3f,Matrix3f>(), nthreads, in, out) ); in test_cuda_basic() 165 CALL_SUBTEST( run_and_compare_to_cuda(prod_test<Matrix4f,Vector4f>(), nthreads, in, out) ); in test_cuda_basic() [all …]
|
/external/ltp/testcases/kernel/fs/fs_fill/ |
D | fs_fill.c | 23 static unsigned int nthreads; variable 66 pthread_t threads[nthreads]; in testrun() 70 for (i = 0; i < nthreads; i++) in testrun() 84 for (i = 0; i < nthreads; i++) in testrun() 94 nthreads = tst_ncpus_conf() + 2; in setup() 95 workers = SAFE_MALLOC(sizeof(struct worker) * nthreads); in setup() 97 for (i = 0; i < nthreads; i++) { in setup() 103 tst_res(TINFO, "Running %i writer threads", nthreads); in setup()
|
/external/grpc-grpc/test/core/gpr/ |
D | cpu_test.cc | 59 int nthreads; member 94 ct->nthreads--; in worker_thread() 95 if (ct->nthreads == 0) { in worker_thread() 108 ct.nthreads = static_cast<int>(ct.ncores) * 3; in cpu_test() 115 uint32_t nthreads = ct.ncores * 3; in cpu_test() local 117 static_cast<grpc_core::Thread*>(gpr_malloc(sizeof(*thd) * nthreads)); in cpu_test() 119 for (i = 0; i < nthreads; i++) { in cpu_test() 128 for (i = 0; i < nthreads; i++) { in cpu_test()
|
D | sync_test.cc | 138 int nthreads; /* number of threads */ member 163 static struct test* test_new(int nthreads, int64_t iterations, int incr_step) { in test_new() argument 165 m->nthreads = nthreads; in test_new() 167 gpr_malloc(sizeof(*m->threads) * nthreads)); in test_new() 171 m->done = nthreads; in test_new() 179 gpr_ref_init(&m->thread_refcount, nthreads); in test_new() 197 for (i = 0; i != m->nthreads; i++) { in test_create_threads() 210 for (int i = 0; i != m->nthreads; i++) { in test_wait() 268 if (m->counter != m->nthreads * m->iterations * m->incr_step) { in test() 270 static_cast<long>(m->counter), m->nthreads, in test() [all …]
|
/external/ltp/testcases/open_posix_testsuite/stress/threads/pthread_create/ |
D | s-c1.c | 104 int nthreads; member 147 int nthreads, ctl, i, tmp; in main() local 197 nthreads = 0; in main() 247 pthread_create(&th[nthreads], in main() 258 nthreads++; in main() 268 if (nthreads > my_max) { in main() 298 if ((nthreads % RESOLUTION) == 0) { in main() 307 m_tmp->nthreads = in main() 308 nthreads; in main() 331 sc, nthreads, in main() [all …]
|
/external/compiler-rt/test/asan/TestCases/Posix/ |
D | halt_on_error-torture.cc | 29 size_t nthreads = 10; variable 61 nthreads = (size_t)strtoul(argv[1], 0, 0); in main() 64 pthread_t *tids = new pthread_t[nthreads]; in main() 66 for (size_t i = 0; i < nthreads; ++i) { in main() 73 for (size_t i = 0; i < nthreads; ++i) { in main()
|
/external/llvm-project/compiler-rt/test/asan/TestCases/Posix/ |
D | halt_on_error-torture.cpp | 24 size_t nthreads = 10; variable 55 nthreads = (size_t)strtoul(argv[1], 0, 0); in main() 58 pthread_t *tids = new pthread_t[nthreads]; in main() 60 for (size_t i = 0; i < nthreads; ++i) { in main() 67 for (size_t i = 0; i < nthreads; ++i) { in main()
|
/external/llvm-project/openmp/runtime/test/master/ |
D | omp_master.c | 7 int nthreads; in test_omp_master() local 10 nthreads = 0; in test_omp_master() 19 nthreads++; in test_omp_master() 24 return ((nthreads == 1) && (executing_thread == 0)); in test_omp_master()
|
D | omp_master_3.c | 7 int nthreads; in test_omp_master_3() local 11 nthreads = 0; in test_omp_master_3() 25 nthreads++; in test_omp_master_3() 30 return ((nthreads == 1) && (executing_thread == 0) && (tid_result == 0)); in test_omp_master_3()
|
/external/llvm-project/openmp/runtime/test/parallel/ |
D | omp_parallel_num_threads.c | 9 int nthreads; in test_omp_parallel_num_threads() local 23 nthreads = 0; in test_omp_parallel_num_threads() 28 nthreads += 1; in test_omp_parallel_num_threads() 30 num_failed = num_failed + !(nthreads == threads); in test_omp_parallel_num_threads()
|
/external/llvm-project/lldb/test/API/commands/process/attach-resume/ |
D | main.cpp | 26 static const size_t nthreads = 16; in main() local 27 std::thread threads[nthreads]; in main() 30 for (i=0; i<nthreads; i++) in main() 33 for (i=0; i<nthreads; i++) in main()
|
/external/llvm-project/openmp/runtime/test/teams/ |
D | teams.c | 29 int nthreads = omp_get_num_threads(); in main() local 30 if (nthreads != NUM_THREADS_PER_TEAM) { in main() 32 nthreads, NUM_THREADS_PER_TEAM); in main() 35 if (tid < 0 || tid >= nthreads) { in main()
|
/external/tensorflow/tensorflow/core/kernels/image/ |
D | resize_nearest_neighbor_op_gpu.cu.cc | 36 const int nthreads, const T* __restrict__ bottom_data, const int in_height, in ResizeNearestNeighborNHWC() argument 40 GPU_1D_KERNEL_LOOP(index, nthreads) { in ResizeNearestNeighborNHWC() 67 const int nthreads, const T* __restrict__ bottom_data, const int in_height, in LegacyResizeNearestNeighborNHWC() argument 71 GPU_1D_KERNEL_LOOP(index, nthreads) { in LegacyResizeNearestNeighborNHWC() 96 const int nthreads, const T* __restrict__ top_diff, const int in_height, in ResizeNearestNeighborBackwardNHWC() argument 100 GPU_1D_KERNEL_LOOP(index, nthreads) { in ResizeNearestNeighborBackwardNHWC() 127 const int nthreads, const T* __restrict__ top_diff, const int in_height, in LegacyResizeNearestNeighborBackwardNHWC() argument 131 GPU_1D_KERNEL_LOOP(index, nthreads) { in LegacyResizeNearestNeighborBackwardNHWC() 176 void (*kernel)(const int nthreads, const T* __restrict__ bottom_data, in operator ()() 229 void (*kernel)(const int nthreads, const T* __restrict__ top_diff, in operator ()()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | bincount_op_gpu.cu.cc | 105 __global__ void BincountReduceKernel(const Tidx* in, T* out, const int nthreads, in BincountReduceKernel() argument 107 GPU_1D_KERNEL_LOOP(index, nthreads) { in BincountReduceKernel() 122 const int nthreads = arr.dimension(0); in Compute() local 125 GpuLaunchConfig config = GetGpuLaunchConfig(nthreads, d); in Compute() 128 output.data(), nthreads, num_bins); in Compute() 137 const int nthreads = num_rows * num_cols; in BincountColReduceKernel() local 138 GPU_1D_KERNEL_LOOP(index, nthreads) { in BincountColReduceKernel() 167 const int nthreads = num_rows * num_cols; in BincountColReduceSharedKernel() local 168 GPU_1D_KERNEL_LOOP(index, nthreads) { in BincountColReduceSharedKernel()
|
D | maxpooling_op_gpu.cu.cc | 68 const int nthreads, const dtype* __restrict__ bottom_data, in MaxPoolForwardNCHW() argument 74 GPU_1D_KERNEL_LOOP(index, nthreads) { in MaxPoolForwardNCHW() 112 const int nthreads, const int32* __restrict__ bottom_data, const int height, in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() argument 119 GPU_1D_KERNEL_LOOP(index, nthreads) { in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() 145 const int nthreads, const dtype* __restrict__ bottom_data, const int height, in MaxPoolForwardNHWC() argument 151 GPU_1D_KERNEL_LOOP(index, nthreads) { in MaxPoolForwardNHWC() 203 __global__ void MaxPoolBackward(const int nthreads, in MaxPoolBackward() argument 209 GPU_1D_KERNEL_LOOP(index, nthreads) { in MaxPoolBackward() 231 const int nthreads, const dtype* __restrict__ bottom_data, in MaxPoolGradBackwardNoMaskNCHW() argument 237 GPU_1D_KERNEL_LOOP(index, nthreads) { in MaxPoolGradBackwardNoMaskNCHW() [all …]
|
D | bias_op_gpu.cu.cc | 54 __global__ void BiasNHWCKernel(int32 nthreads, const T* __restrict__ input, in BiasNHWCKernel() argument 57 GPU_1D_KERNEL_LOOP(index, nthreads) { in BiasNHWCKernel() 64 __global__ void BiasNCHWKernel(int32 nthreads, const T* __restrict__ input, in BiasNCHWKernel() argument 68 GPU_1D_KERNEL_LOOP(index, nthreads) { in BiasNCHWKernel() 103 __global__ void BiasGradNHWC_Naive(int32 nthreads, in BiasGradNHWC_Naive() argument 107 GPU_1D_KERNEL_LOOP(index, nthreads) { in BiasGradNHWC_Naive() 115 __global__ void BiasGradNCHW_Naive(int32 nthreads, in BiasGradNCHW_Naive() argument 119 GPU_1D_KERNEL_LOOP(index, nthreads) { in BiasGradNCHW_Naive() 128 int32 nthreads, const T* __restrict__ output_backprop, in BiasGradNHWC_SharedAtomics() argument 138 for (int32 index = blockIdx.x * blockDim.x + threadIdx.x; index < nthreads; in BiasGradNHWC_SharedAtomics()
|
/external/jemalloc_new/test/unit/ |
D | retained.c | 109 unsigned nthreads = ncpus * 2; in TEST_BEGIN() local 110 VARIABLE_ARRAY(thd_t, threads, nthreads); in TEST_BEGIN() 111 for (unsigned i = 0; i < nthreads; i++) { in TEST_BEGIN() 121 while (atomic_load_u(&nfinished, ATOMIC_ACQUIRE) < nthreads) { in TEST_BEGIN() 132 size_t allocated = esz * nthreads * PER_THD_NALLOCS; in TEST_BEGIN() 169 for (unsigned i = 0; i < nthreads; i++) { in TEST_BEGIN()
|
/external/autotest/client/tests/monotonic_time/src/ |
D | threads.c | 86 int create_threads(int nthreads, thread_func_t func, void *arg) in create_threads() argument 88 if (nthreads > MAX_THREADS) in create_threads() 89 nthreads = MAX_THREADS; in create_threads() 91 while (--nthreads >= 0) { in create_threads()
|
/external/ltp/testcases/open_posix_testsuite/stress/threads/pthread_cond_timedwait/ |
D | s-c.c | 175 int nthreads; member 290 long do_threads_test(int nthreads, mes_t * measure) in do_threads_test() argument 317 th = (pthread_t *) calloc(nthreads, sizeof(pthread_t)); in do_threads_test() 322 output("%d", nthreads); in do_threads_test() 396 for (i = 0; i < nthreads; i++) { in do_threads_test() 444 while (tnum < nthreads) { in do_threads_test() 495 for (i = 0; i < nthreads; i++) { in do_threads_test() 538 output("%5d threads; %d.%09d s (%i loops)\n", nthreads, ts_cumul.tv_sec, in do_threads_test() 617 m_tmp->nthreads = nth; in main() 737 Xavg += (double)cur->nthreads; in parse_measure() [all …]
|