/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
D | parallel_interleave_dataset_op.cc | 241 workers_(dataset()->num_threads()), in Iterator() 248 for (auto& worker : workers_) { in ~Iterator() 280 WorkerState* current_worker = &workers_[current_worker_index]; in GetNextInternal() 358 workers_[interleave_indices_[next_index_]].cond_var.wait(l); in GetNextInternal() 389 writer->WriteScalar(full_name("workers_size"), workers_.size())); in SaveInternal() 390 for (int i = 0; i < workers_.size(); ++i) { in SaveInternal() 593 workers_[i].SetInputs(s, std::move(args)); in EnsureWorkerThreadsStarted() 637 workers_[thread_index].cond_var.notify_all(); in WorkerThread() 675 while (!cancelled_ && !workers_[thread_index].is_producing) { in WorkerThread() 677 workers_[thread_index].cond_var.wait(l); in WorkerThread() [all …]
|
D | numa_map_and_batch_dataset_op.cc | 214 for (size_t i = 0; i < workers_.size(); ++i) { in ~Iterator() 215 workers_[i]->manager.Cancel(); in ~Iterator() 247 worker = workers_[cur_block_].get(); in GetNextInternal() 248 cur_block_ = (cur_block_ + 1) % workers_.size(); in GetNextInternal() 272 for (size_t i = 0; i < workers_.size(); ++i) { in SaveInternal() 273 if (!workers_[i]->manager.Quiesce()) { in SaveInternal() 282 writer->WriteScalar(full_name("num_workers"), workers_.size())); in SaveInternal() 284 for (size_t i = 0; i < workers_.size(); ++i) { in SaveInternal() 285 size_t index = (cur_block_ + i) % workers_.size(); in SaveInternal() 286 TF_RETURN_IF_ERROR(workers_[index]->manager.Save(writer, this, i)); in SaveInternal() [all …]
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | test_utils.h | 128 workers_[target] = wi; in AddWorker() 137 for (auto it : workers_) { in ListWorkers() 145 for (auto it : workers_) { in ListWorkersInJob() 156 auto it = workers_.find(target); in CreateWorker() 157 if (it != workers_.end()) { in CreateWorker() 187 std::unordered_map<string, WorkerInterface*> workers_;
|
D | worker_session.cc | 29 for (auto& p : workers_) { in ~WorkerFreeListCache() 45 auto p = workers_.find(target); in CreateWorker() 46 if (p != workers_.end()) { in CreateWorker() 52 workers_.insert(std::make_pair(target, state)); in CreateWorker() 90 std::unordered_map<string, WorkerState> workers_ GUARDED_BY(mu_);
|
D | collective_rma_distributed_test.cc | 142 auto it = workers_.find(task_name); in GetDeviceLocalityAsync() 143 if (it == workers_.end()) { in GetDeviceLocalityAsync() 177 for (FakeWorker* w : workers_) { in ~CollRMADistTest() 230 workers_.push_back(fw); in DefineWorker() 247 std::vector<FakeWorker*> workers_; member in tensorflow::__anon7e5c43490111::CollRMADistTest 264 FakeWorker* wi = workers_[1]; in TEST_F() 299 FakeWorker* wi = workers_[1]; in TEST_F()
|
D | device_resolver_distributed_test.cc | 104 auto it = workers_.find(task_name); in GetDeviceLocalityAsync() 105 if (it == workers_.end()) { in GetDeviceLocalityAsync() 139 for (FakeWorker* w : workers_) { in ~DeviceResDistTest() 170 workers_.push_back(fw); in DefineWorker() 178 std::vector<FakeWorker*> workers_; member in tensorflow::__anon0903cdc10111::DeviceResDistTest
|
D | collective_param_resolver_distributed_test.cc | 102 auto it = workers_.find(task_name); in GetDeviceLocalityAsync() 103 if (it == workers_.end()) { in GetDeviceLocalityAsync() 140 for (FakeWorker* w : workers_) { in ~DeviceResDistTest() 179 workers_.push_back(fw); in DefineWorker() 289 std::vector<FakeWorker*> workers_; member in tensorflow::__anonc2f8b6eb0111::DeviceResDistTest
|
/external/gemmlowp/internal/ |
D | multi_thread_gemm.h | 370 for (auto w : workers_) { in ~WorkersPool() 380 assert(workers_count <= workers_.size()); in Execute() 384 [this, &n](Task* task) { workers_[n++]->StartWork(task); }); in Execute() 401 if (workers_.size() >= workers_count) { in CreateWorkers() 404 counter_to_decrement_when_ready_.Reset(workers_count - workers_.size()); in CreateWorkers() 405 while (workers_.size() < workers_count) { in CreateWorkers() 406 workers_.push_back(new Worker(&counter_to_decrement_when_ready_)); in CreateWorkers() 416 std::vector<Worker*> workers_; variable
|
/external/tensorflow/tensorflow/core/kernels/ |
D | quantization_utils.h | 904 : workers_(workers) {} in TensorflowGemmlowpWorkersPool() 915 assert(workers_ != nullptr); in Execute() 918 workers_->Schedule([this, task]() { in Execute() 934 thread::ThreadPool* const workers_;
|
/external/v8/src/ |
D | d8.h | 555 static std::vector<Worker*> workers_; variable
|
D | d8.cc | 412 std::vector<Worker*> Shell::workers_; member in v8::Shell 1415 if (workers_.size() >= kMaxWorkers) { in WorkerNew() 1429 workers_.push_back(worker); in WorkerNew() 3302 workers_copy.swap(workers_); in CleanupWorkers()
|