| /external/tensorflow/tensorflow/python/distribute/v1/ |
| D | all_reduce_test.py | 89 def _buildInput(self, num_workers, num_gpus): argument 144 def _buildRing(self, num_workers, num_gpus, subdiv): argument 149 def _testAllReduce(self, num_workers, num_gpus, shape, build_f): argument 165 def _testRingAllReduce(self, num_workers, num_gpus, shape, subdiv): argument 189 def _buildShuffle(self, num_workers, num_gpus, num_shards): argument 196 def _testShuffleAllReduce(self, num_workers, num_gpus, shape, num_shards): argument 216 def _buildRecursiveHD(self, num_workers, num_gpus): argument 222 def _testRecursiveHDAllReduce(self, num_workers, num_gpus, shape): argument
|
| /external/tensorflow/tensorflow/core/distributed_runtime/ |
| D | collective_param_resolver_distributed_test.cc | 138 void DefineWorkers(int num_workers, int num_devices, in DefineWorkers() 178 void DefineCollectiveParams(int num_workers, int num_devices, in DefineCollectiveParams() 191 CollectiveParams* CreateCollectiveParams(int num_workers, int num_devices, in CreateCollectiveParams() 208 void IssueRequests(int num_workers, int num_devices) { in IssueRequests() 244 void ValidateCollectiveParams(int num_workers, int num_devices) { in ValidateCollectiveParams() 289 void RestartWorker(int worker_idx, int num_workers, int num_devices, in RestartWorker() 326 const int num_workers = 1; in TEST_F() local 335 const int num_workers = 2; in TEST_F() local 344 const int num_workers = 2; in TEST_F() local 385 const int num_workers = 4; in TEST_F() local
|
| /external/rust/crates/grpcio-sys/grpc/src/core/lib/iomgr/poller/ |
| D | eventmanager_libuv.cc | 26 grpc::experimental::LibuvEventManager::Options::Options(int num_workers) in Options() 31 int num_workers = options_.num_workers(); in LibuvEventManager() local
|
| D | eventmanager_libuv.h | 41 int num_workers() const { return num_workers_; } in num_workers() function
|
| /external/libaom/libaom/av1/encoder/ |
| D | ethread.c | 213 MultiThreadHandle *multi_thread_ctxt, int num_tiles, int num_workers) { in assign_tile_to_thread() 389 static AOM_INLINE void create_enc_workers(AV1_COMP *cpi, int num_workers) { in create_enc_workers() 511 static AOM_INLINE void launch_enc_workers(AV1_COMP *cpi, int num_workers) { in launch_enc_workers() 528 static AOM_INLINE void sync_enc_workers(AV1_COMP *cpi, int num_workers) { in sync_enc_workers() 544 int num_workers) { in accumulate_counters_enc_workers() 564 int num_workers) { in prepare_enc_workers() 624 int num_workers = AOMMIN(cpi->oxcf.max_threads, tile_cols * tile_rows); in av1_encode_tiles_mt() local 659 int num_workers = 0; in av1_encode_tiles_row_mt() local
|
| /external/tensorflow/tensorflow/python/distribute/ |
| D | combinations_test.py | 48 def testClusterParams(self, distribution, has_chief, num_workers): argument 57 def testClusterParamsHasDefault(self, distribution, has_chief, num_workers): argument 64 def testClusterParamsNoStrategy(self, v, has_chief, num_workers): argument
|
| D | combinations.py | 260 num_workers=1, argument 484 def decorator(self, has_chief, num_workers, runner, **kwargs): argument 548 def _num_total_workers(has_chief, num_workers): argument
|
| D | multi_worker_test_base.py | 88 def _create_cluster(num_workers, argument 156 def create_in_process_cluster(num_workers, argument 353 def create_multi_process_cluster(num_workers, argument 374 num_workers=1, argument 454 def setUpClass(cls, num_workers=2, num_ps=1): # pylint: disable=g-missing-super-call argument
|
| /external/tensorflow/tensorflow/python/data/experimental/ops/ |
| D | distribute.py | 69 def __init__(self, input_dataset, num_workers, index, num_replicas=None): argument 88 def _AutoShardDatasetV1(input_dataset, num_workers, index, num_replicas=None): # pylint: disable=i… argument 353 def batch_sizes_for_worker(global_batch_size, num_workers, argument
|
| /external/tensorflow/tensorflow/core/grappler/optimizers/data/ |
| D | auto_shard.cc | 141 int64 num_workers, int64 index) { in AddShardNode() 382 int64 num_workers, int64 index) { in ProcessDatasetSourceNode() 423 const NodeDef* node, int64 num_workers, int64 index, in FindFuncAndTensorSliceDataset() 449 Status RecursivelyHandleOp(const NodeDef& node, int64 num_workers, int64 index, in RecursivelyHandleOp() 554 Status ShardByFile(const NodeDef& sink_node, int64 num_workers, int64 index, in ShardByFile() 611 Status ShardByData(const NodeDef& sink_node, int64 num_workers, int64 index, in ShardByData() 627 Status OptimizeGraph(const GrapplerItem& item, int64 num_workers, int64 index, in OptimizeGraph()
|
| /external/tensorflow/tensorflow/core/kernels/data/experimental/ |
| D | auto_shard_dataset_op.cc | 47 int64 index, num_workers, auto_shard_policy, num_replicas; in MakeDataset() local 73 RewriterConfig AutoShardDatasetOp::CreateConfig(int64 num_workers, int64 index, in CreateConfig()
|
| /external/tensorflow/tensorflow/core/kernels/boosted_trees/quantiles/ |
| D | weighted_quantiles_stream_test.cc | 196 int32 num_workers, double eps, int64 max_elements, in TestDistributedStreams() 237 const int32 num_workers = 10; in TEST() local 246 const int32 num_workers = 10; in TEST() local 258 const int32 num_workers = 10; in TEST() local 267 const int32 num_workers = 10; in TEST() local
|
| /external/libvpx/libvpx/vp9/encoder/ |
| D | vp9_ethread.c | 78 static void create_enc_workers(VP9_COMP *cpi, int num_workers) { in create_enc_workers() 140 int num_workers) { in launch_enc_workers() 175 const int num_workers = VPXMIN(cpi->oxcf.max_threads, tile_cols); in vp9_encode_tiles_mt() local 443 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_fp_row_mt() local 529 int num_workers = cpi->num_workers ? cpi->num_workers : 1; in vp9_temporal_filter_row_mt() local 602 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_tiles_row_mt() local
|
| /external/libaom/libaom/av1/common/ |
| D | thread_common.c | 56 int width, int num_workers) { in loop_filter_alloc() 423 const int num_workers = nworkers; in loop_filter_rows_mt() local 484 AVxWorker *workers, int num_workers, in av1_loop_filter_frame_mt() 588 int num_workers, int num_rows_lr, in loop_restoration_alloc() 652 void av1_loop_restoration_dealloc(AV1LrSync *lr_sync, int num_workers) { in av1_loop_restoration_dealloc() 874 const int num_workers = nworkers; in foreach_rest_unit_in_planes_mt() local 917 AVxWorker *workers, int num_workers, in av1_loop_restoration_filter_frame_mt()
|
| /external/libvpx/libvpx/vp9/common/ |
| D | vp9_thread_common.c | 170 const int num_workers = VPXMIN(nworkers, VPXMIN(num_tile_cols, sb_rows)); in loop_filter_rows_mt() local 223 int num_workers, VP9LfSync *lf_sync) { in vp9_loop_filter_frame_mt() 243 int num_workers) { in vp9_lpf_mt_init() 280 int width, int num_workers) { in vp9_loop_filter_alloc()
|
| /external/webrtc/third_party/abseil-cpp/absl/synchronization/ |
| D | blocking_counter_test.cc | 39 const int num_workers = 10; in TEST() local
|
| /external/libtextclassifier/abseil-cpp/absl/synchronization/ |
| D | blocking_counter_test.cc | 39 const int num_workers = 10; in TEST() local
|
| /external/abseil-cpp/absl/synchronization/ |
| D | blocking_counter_test.cc | 39 const int num_workers = 10; in TEST() local
|
| /external/rust/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/synchronization/ |
| D | blocking_counter_test.cc | 39 const int num_workers = 10; in TEST() local
|
| /external/openscreen/third_party/abseil/src/absl/synchronization/ |
| D | blocking_counter_test.cc | 39 const int num_workers = 10; in TEST() local
|
| /external/tensorflow/tensorflow/python/data/experimental/kernel_tests/ |
| D | data_service_test_base.py | 53 num_workers, argument 169 num_workers, argument
|
| /external/tensorflow/tensorflow/core/data/service/ |
| D | test_cluster.cc | 41 TestCluster::TestCluster(int num_workers) : num_workers_(num_workers) {} in TestCluster()
|
| /external/tensorflow/tensorflow/python/training/ |
| D | sync_replicas_optimizer_test.py | 35 def get_workers(num_workers, replicas_to_aggregate, workers): argument
|
| /external/tensorflow/tensorflow/python/keras/distribute/ |
| D | parameter_server_training_test.py | 65 def make_cluster(num_workers, num_ps): argument 74 def make_coordinator(num_workers, num_ps): argument
|
| /external/tensorflow/tensorflow/python/data/experimental/service/ |
| D | server_lib_wrapper.cc | 42 int num_workers; in PYBIND11_MODULE() local
|