/external/gwp_asan/gwp_asan/tests/ |
D | thread_contention.cpp | 45 void runThreadContentionTest(unsigned NumThreads, unsigned NumIterations, in runThreadContentionTest() argument 50 for (unsigned i = 0; i < NumThreads; ++i) { in runThreadContentionTest() 61 unsigned NumThreads = 4; in TEST_F() local 63 InitNumSlots(NumThreads); in TEST_F() 64 runThreadContentionTest(NumThreads, NumIterations, &GPA); in TEST_F()
|
D | mutex_test.cpp | 66 static void runSynchronisedTest(unsigned NumThreads, unsigned CounterMax) { in runSynchronisedTest() argument 69 ASSERT_TRUE(CounterMax % NumThreads == 0); in runSynchronisedTest() 75 for (unsigned i = 0; i < NumThreads; ++i) in runSynchronisedTest() 77 CounterMax / NumThreads); in runSynchronisedTest()
|
/external/scudo/standalone/tests/ |
D | condition_variable_test.cpp | 18 constexpr scudo::u32 NumThreads = 2; in simpleWaitAndNotifyAll() local 20 std::thread Threads[NumThreads]; in simpleWaitAndNotifyAll() 27 for (scudo::u32 I = 0; I < NumThreads; ++I) { in simpleWaitAndNotifyAll() 32 if (Counter % NumThreads != Id && Counter < CounterMax) in simpleWaitAndNotifyAll()
|
/external/lzma/CPP/7zip/Archive/7z/ |
D | 7zCompressionMode.h | 17 UInt32 NumThreads; member 20 CMethodFull(): CodecIndex(-1), NumThreads(1), Set_NumThreads(false) {} in CMethodFull() 61 UInt32 NumThreads; member 76 , NumThreads(1) in CCompressionMethodMode()
|
/external/swiftshader/third_party/subzero/src/ |
D | LinuxMallocProfiling.cpp | 67 LinuxMallocProfiling::LinuxMallocProfiling(size_t NumThreads, Ostream *Ls) in LinuxMallocProfiling() argument 69 if (NumThreads != 0) { in LinuxMallocProfiling() 102 LinuxMallocProfiling::LinuxMallocProfiling(size_t NumThreads, Ostream *Ls) { in LinuxMallocProfiling() argument 103 (void)NumThreads; in LinuxMallocProfiling()
|
D | IceBrowserCompileServer.cpp | 116 char *onInitCallback(uint32_t NumThreads, int *ObjFileFDs, in onInitCallback() argument 152 gCompileServer->getParsedFlags(UseNumThreadsFromBrowser, NumThreads, in onInitCallback() 242 uint32_t NumThreads, int argc, in getParsedFlags() argument 248 ClFlags::Flags.setNumTranslationThreads(NumThreads); in getParsedFlags()
|
D | LinuxMallocProfiling.h | 32 LinuxMallocProfiling(size_t NumThreads, Ostream *Ls);
|
/external/swiftshader/third_party/subzero/crosstest/ |
D | test_sync_atomic_main.cpp | 241 const size_t NumThreads = 4; in testAtomicRMWThreads() local 242 pthread_t t[NumThreads]; in testAtomicRMWThreads() 243 pthread_attr_t attr[NumThreads]; in testAtomicRMWThreads() 247 for (size_t m = 0; m < NumThreads; ++m) { in testAtomicRMWThreads() 255 for (size_t m = 0; m < NumThreads; ++m) { in testAtomicRMWThreads() 262 for (size_t m = 0; m < NumThreads; ++m) { in testAtomicRMWThreads() 273 for (size_t m = 0; m < NumThreads; ++m) { in testAtomicRMWThreads()
|
/external/swiftshader/third_party/llvm-16.0/llvm/lib/Target/DirectX/ |
D | DXILMetadata.cpp | 95 unsigned NumThreads[3]; member 111 SmallVector<StringRef> NumThreads; in EntryProps() local 112 NumThreadsStr.split(NumThreads, ','); in EntryProps() 113 assert(NumThreads.size() == 3 && "invalid numthreads"); in EntryProps() 115 llvm::zip(NumThreads, MutableArrayRef<unsigned>(CS.NumThreads)); in EntryProps() 181 for (auto Num : ArrayRef<unsigned>(CS.NumThreads)) in appendNumThreads()
|
/external/lzma/CPP/7zip/Compress/ |
D | BZip2Encoder.cpp | 104 if (++Encoder->NextBlockIndex == Encoder->NumThreads) in ThreadFunc() 152 NumThreads = 1; in CEncoder() 171 if (ThreadsInfo && m_NumThreadsPrev == NumThreads) in Create() 176 MtMode = (NumThreads > 1); in Create() 177 m_NumThreadsPrev = NumThreads; in Create() 178 ThreadsInfo = new CThreadInfo[NumThreads]; in Create() 183 for (UInt32 t = 0; t < NumThreads; t++) in Create() 192 NumThreads = t; in Create() 207 for (UInt32 t = 0; t < NumThreads; t++) in Free() 721 if (blockIndex == Encoder->NumThreads) in EncodeBlock3() [all …]
|
/external/tensorflow/tensorflow/core/common_runtime/ |
D | process_util_test.cc | 22 TEST(ProcessUtilTest, NumThreads) { in TEST() argument 33 EXPECT_EQ(10, pool->NumThreads()); in TEST()
|
D | renamed_device.cc | 62 eigen_worker_threads_.num_threads = underlying_threadpool->NumThreads(); in RenamedDevice() 67 underlying_threadpool, underlying_threadpool->NumThreads()); in RenamedDevice()
|
/external/tensorflow/tensorflow/core/platform/ |
D | threadpool.cc | 111 underlying_threadpool_, underlying_threadpool_->NumThreads(), nullptr)); in ThreadPool() 124 NumThreads() == 1) { in NumShardsUsedByFixedBlockSizeScheduling() 191 if (num_shards_used <= NumThreads()) { in ParallelForFixedBlockSizeScheduling() 244 int ThreadPool::NumThreads() const { in NumThreads() function in tensorflow::thread::ThreadPool 245 return underlying_threadpool_->NumThreads(); in NumThreads()
|
/external/tensorflow/tensorflow/core/tfrt/utils/ |
D | test_util.h | 41 int NumThreads() const override { in NumThreads() function 42 return underlying_threadpool_.NumThreads(); in NumThreads()
|
D | thread_pool.h | 42 int NumThreads() const override { in NumThreads() function 43 return underlying_threadpool_.NumThreads(); in NumThreads()
|
/external/tensorflow/tensorflow/compiler/aot/ |
D | test.cc | 63 Eigen::ThreadPoolDevice device(&pool, pool.NumThreads()); in TEST() 75 Eigen::ThreadPoolDevice device(&pool, pool.NumThreads()); in BM_NAME()
|
/external/tensorflow/tensorflow/core/kernels/batching_util/ |
D | bounded_executor.cc | 58 for (int i = 0; i < NumThreads(); i++) { in ~BoundedExecutor() 77 int BoundedExecutor::NumThreads() const { return options_.num_threads; } in NumThreads() function in tensorflow::serving::BoundedExecutor
|
/external/tensorflow/tensorflow/compiler/aot/tests/ |
D | tfcompile_test.cc | 264 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 335 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 361 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 467 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 495 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 519 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 533 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 555 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 582 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST() 682 Eigen::ThreadPoolDevice device(&tp, tp.NumThreads()); in TEST()
|
/external/tensorflow/tensorflow/core/tfrt/eager/ |
D | tfrt_context.h | 60 int NumThreads() const override { return thread_pool().NumThreads(); } in NumThreads() function
|
D | tfrt_context.cc | 62 << "# threads: " << eager_ctx_thread_pool_->NumThreads() in TfrtContext() 64 << local_thread_pool_wrapper_->NumThreads(); in TfrtContext()
|
/external/lzma/CPP/7zip/UI/Common/ |
D | ZipRegistry.h | 90 UInt32 NumThreads; member 122 BlockLogSize = NumThreads = Level = Dictionary = Order = (UInt32)(Int32)-1; in ResetForLevelChange()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_2d_gpu.h | 274 template <typename T, int NumThreads, int TileSizeI, int TileSizeJ, 279 eigen_assert(blockDim.x == NumThreads); 285 constexpr int ReadRowPerPass = NumThreads / TileSizeJ; 286 constexpr int WriteRowPerPass = NumThreads / TileSizeI; 345 constexpr int in_effective_thread_num = NumThreads / TileSizeJ * TileSizeJ; 391 constexpr int out_effective_thread_num = NumThreads / TileSizeI * TileSizeI; 702 constexpr int NumThreads = TileLongSide; 705 SwapDimension1And2InTensor3UsingTiles<T, NumThreads, TileLongSide, 707 total_tiles_count, NumThreads, 0, d.stream(), input, input_dims, 711 SwapDimension1And2InTensor3UsingTiles<T, NumThreads, TileShortSide, [all …]
|
/external/swiftshader/third_party/marl/tools/bench/ |
D | bench.go | 32 NumThreads uint member 50 t.NumThreads = uint(n)
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | eigen_support.cc | 83 int NumThreads() const override { return pool_ ? pool_->NumThreads() : 1; } in NumThreads() function in tflite::eigen_support::__anon143c18630111::EigenThreadPoolWrapper
|
/external/eigen/unsupported/Eigen/CXX11/src/ThreadPool/ |
D | ThreadPoolInterface.h | 37 virtual int NumThreads() const = 0;
|