/external/tensorflow/tensorflow/core/kernels/ |
D | bincount_op_gpu.cu.cc | 105 __global__ void BincountReduceKernel(const Tidx* in, T* out, const int nthreads, in BincountReduceKernel() 122 const int nthreads = arr.dimension(0); in Compute() local 137 const int nthreads = num_rows * num_cols; in BincountColReduceKernel() local 167 const int nthreads = num_rows * num_cols; in BincountColReduceSharedKernel() local
|
D | bias_op_gpu.cu.cc | 54 __global__ void BiasNHWCKernel(int32 nthreads, const T* __restrict__ input, in BiasNHWCKernel() 64 __global__ void BiasNCHWKernel(int32 nthreads, const T* __restrict__ input, in BiasNCHWKernel() 103 __global__ void BiasGradNHWC_Naive(int32 nthreads, in BiasGradNHWC_Naive() 115 __global__ void BiasGradNCHW_Naive(int32 nthreads, in BiasGradNCHW_Naive() 128 int32 nthreads, const T* __restrict__ output_backprop, in BiasGradNHWC_SharedAtomics()
|
D | maxpooling_op_gpu.cu.cc | 68 const int nthreads, const dtype* __restrict__ bottom_data, in MaxPoolForwardNCHW() 112 const int nthreads, const int32* __restrict__ bottom_data, const int height, in MaxPoolForwardNoMaskKernel_NCHW_VECT_C() 145 const int nthreads, const dtype* __restrict__ bottom_data, const int height, in MaxPoolForwardNHWC() 203 __global__ void MaxPoolBackward(const int nthreads, in MaxPoolBackward() 231 const int nthreads, const dtype* __restrict__ bottom_data, in MaxPoolGradBackwardNoMaskNCHW() 272 const int nthreads, const dtype* __restrict__ bottom_data, in MaxPoolGradBackwardNoMaskNHWC() 332 __global__ void MaxPoolGradBackward(const int nthreads, in MaxPoolGradBackward()
|
D | depthtospace_op_gpu.cu.cc | 33 __global__ void D2S_NHWC(const int32 nthreads, in D2S_NHWC() 64 __global__ void D2S_NCHW(const int32 nthreads, in D2S_NCHW() 101 __global__ void D2S_NCHW_LOOP(const int32 nthreads, in D2S_NCHW_LOOP()
|
D | spacetodepth_op_gpu.cu.cc | 32 __global__ void S2D_NHWC(const int32 nthreads, in S2D_NHWC() 64 __global__ void S2D_NCHW(const int32 nthreads, in S2D_NCHW() 102 __global__ void S2D_NCHW_LOOP(const int32 nthreads, in S2D_NCHW_LOOP()
|
D | dilation_ops_gpu.cu.cc | 39 const int32 nthreads, const T* __restrict__ input_ptr, in DilationKernel() 78 const int32 nthreads, const T* __restrict__ input_ptr, in DilationBackpropInputKernel() 128 const int32 nthreads, const T* __restrict__ input_ptr, in DilationBackpropFilterKernel()
|
D | inplace_ops_functor_gpu.cu.cc | 30 __global__ void DoParallelConcatOpKernel(int nthreads, const int64 rows, in DoParallelConcatOpKernel() 83 __global__ void DoInplaceOpKernel(int nthreads, const int64 rows, in DoInplaceOpKernel()
|
/external/grpc-grpc/test/core/gpr/ |
D | cpu_test.cc | 59 int nthreads; member 115 uint32_t nthreads = ct.ncores * 3; in cpu_test() local
|
/external/tensorflow/tensorflow/core/kernels/image/ |
D | resize_nearest_neighbor_op_gpu.cu.cc | 36 const int nthreads, const T* __restrict__ bottom_data, const int in_height, in ResizeNearestNeighborNHWC() 67 const int nthreads, const T* __restrict__ bottom_data, const int in_height, in LegacyResizeNearestNeighborNHWC() 96 const int nthreads, const T* __restrict__ top_diff, const int in_height, in ResizeNearestNeighborBackwardNHWC() 127 const int nthreads, const T* __restrict__ top_diff, const int in_height, in LegacyResizeNearestNeighborBackwardNHWC()
|
D | resize_bilinear_op_gpu.cu.cc | 118 const int32 nthreads, const T* __restrict__ images, float height_scale, in ResizeBilinearKernel() 168 __global__ void ResizeBilinearGradKernel(const int32 nthreads, in ResizeBilinearGradKernel() 234 const int32 nthreads, const float* __restrict__ input_grad, in ResizeBilinearDeterministicGradKernel() 287 const int32 nthreads, const T* __restrict__ images, float height_scale, in LegacyResizeBilinearKernel() 337 const int32 nthreads, const float* __restrict__ input_grad, in LegacyResizeBilinearGradKernel()
|
/external/llvm-project/openmp/runtime/test/api/ |
D | omp_get_num_threads.c | 10 int nthreads = 0; in test_omp_get_num_threads() local
|
D | kmp_set_defaults_lock_bug.c | 15 int nthreads = 0; in test_kmp_set_defaults_lock_bug() local
|
D | omp_pause_resource.c | 6 int fails, nthreads, my_dev; in test_omp_pause_resource() local
|
/external/llvm-project/openmp/runtime/test/master/ |
D | omp_master.c | 7 int nthreads; in test_omp_master() local
|
D | omp_master_3.c | 7 int nthreads; in test_omp_master_3() local
|
/external/llvm-project/openmp/runtime/test/parallel/ |
D | omp_parallel_num_threads.c | 9 int nthreads; in test_omp_parallel_num_threads() local
|
/external/python/cpython2/Tools/ccbench/ |
D | ccbench.py | 196 def run_throughput_test(func, args, nthreads): argument 300 def run_latency_test(func, args, nthreads): argument 432 def run_bandwidth_test(func, args, nthreads): argument
|
/external/python/cpython3/Tools/ccbench/ |
D | ccbench.py | 187 def run_throughput_test(func, args, nthreads): argument 294 def run_latency_test(func, args, nthreads): argument 427 def run_bandwidth_test(func, args, nthreads): argument
|
/external/ltp/testcases/realtime/perf/latency/ |
D | pthread_cond_many.c | 55 int nthreads = 0; variable 186 void test_signal(long iter, long nthreads) in test_signal()
|
/external/llvm-project/openmp/runtime/test/teams/ |
D | teams.c | 29 int nthreads = omp_get_num_threads(); in main() local
|
/external/llvm-project/lldb/test/API/commands/process/attach-resume/ |
D | main.cpp | 26 static const size_t nthreads = 16; in main() local
|
/external/compiler-rt/test/asan/TestCases/Posix/ |
D | halt_on_error-torture.cc | 29 size_t nthreads = 10; variable
|
/external/llvm-project/compiler-rt/test/asan/TestCases/Posix/ |
D | halt_on_error-torture.cpp | 24 size_t nthreads = 10; variable
|
/external/ltp/testcases/open_posix_testsuite/stress/threads/pthread_create/ |
D | s-c1.c | 104 int nthreads; member 147 int nthreads, ctl, i, tmp; in main() local
|
/external/ltp/testcases/kernel/fs/fs_fill/ |
D | fs_fill.c | 23 static unsigned int nthreads; variable
|