/external/tensorflow/tensorflow/python/distribute/ |
D | all_reduce_test.py | 90 def _buildInput(self, num_workers, num_gpus): argument 145 def _buildRing(self, num_workers, num_gpus, subdiv): argument 150 def _testAllReduce(self, num_workers, num_gpus, shape, build_f): argument 166 def _testRingAllReduce(self, num_workers, num_gpus, shape, subdiv): argument 190 def _buildShuffle(self, num_workers, num_gpus, num_shards): argument 197 def _testShuffleAllReduce(self, num_workers, num_gpus, shape, num_shards): argument 217 def _buildRecursiveHD(self, num_workers, num_gpus): argument 223 def _testRecursiveHDAllReduce(self, num_workers, num_gpus, shape): argument
|
D | cross_device_utils.py | 329 num_workers, argument 380 num_workers, argument 420 def sum_gradients_all_reduce(dev_prefixes, replica_grads, num_workers, alg, argument
|
D | all_reduce.py | 193 def _ring_permutations(num_workers, num_subchunks, gpu_perm): argument 254 def build_ring_all_reduce(input_tensors, num_workers, num_subchunks, argument
|
/external/tensorflow/tensorflow/contrib/opt/python/training/ |
D | model_average_optimizer_test.py | 35 def create_local_cluster(num_workers, num_ps, protocol="grpc"): argument 61 def _get_workers(num_workers, steps, workers): argument
|
D | agn_optimizer_test.py | 38 def create_local_cluster(num_workers, num_ps, protocol="grpc"): argument 64 def _get_workers(num_workers, period, workers, num_ps=1): argument
|
D | elastic_average_optimizer_test.py | 41 def create_local_cluster(num_workers, num_ps, protocol="grpc"): argument 67 def _get_workers(num_workers, period, workers, moving_rate, num_ps=1): argument
|
D | drop_stale_gradient_optimizer_test.py | 34 def _get_workers(num_workers, staleness): argument
|
/external/tensorflow/tensorflow/core/distributed_runtime/ |
D | collective_param_resolver_distributed_test.cc | 145 void DefineWorkers(int num_workers, int num_devices, in DefineWorkers() 183 void DefineCollectiveParams(int num_workers, int num_devices) { in DefineCollectiveParams() 205 void IssueRequests(int num_workers, int num_devices) { in IssueRequests() 222 void IssueRequest(int num_workers, int num_devices, int idx) { in IssueRequest() 248 void ValidateCollectiveParams(int num_workers, int num_devices) { in ValidateCollectiveParams() 298 const int num_workers = 1; in TEST_F() local 307 const int num_workers = 2; in TEST_F() local 316 const int num_workers = 4; in TEST_F() local
|
/external/libaom/libaom/av1/encoder/ |
D | ethread.c | 198 int num_tiles, int num_workers) { in assign_tile_to_thread() 377 static void create_enc_workers(AV1_COMP *cpi, int num_workers) { in create_enc_workers() 484 static void launch_enc_workers(AV1_COMP *cpi, int num_workers) { in launch_enc_workers() 501 static void sync_enc_workers(AV1_COMP *cpi, int num_workers) { in sync_enc_workers() 516 static void accumulate_counters_enc_workers(AV1_COMP *cpi, int num_workers) { in accumulate_counters_enc_workers() 534 int num_workers) { in prepare_enc_workers() 591 int num_workers = AOMMIN(cpi->oxcf.max_threads, tile_cols * tile_rows); in av1_encode_tiles_mt() local 626 int num_workers = 0; in av1_encode_tiles_row_mt() local
|
/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
D | rebatch_dataset_op.cc | 36 int64 num_workers; in MakeDataset() local 57 const int64 num_workers, const DataTypeVector& output_types, in Dataset()
|
D | auto_shard_dataset_op.cc | 37 int64 num_workers; in MakeDataset() local 64 const int64 num_workers, const int64 index, in Dataset()
|
/external/tensorflow/tensorflow/core/grappler/optimizers/data/ |
D | auto_shard.cc | 107 int64 num_workers, int64 index) { in AddShardNode() 176 Status RecursivelyHandleOp(const NodeDef& node, int64 num_workers, int64 index, in RecursivelyHandleOp() 227 Status OptimizeGraph(const GrapplerItem& item, int64 num_workers, int64 index, in OptimizeGraph()
|
D | rebatch.cc | 140 Status MutateBatchSize(const NodeDef& node, int64 num_workers, in MutateBatchSize() 186 Status RecursivelyHandleOp(const NodeDef& node, int64 num_workers, in RecursivelyHandleOp() 244 Status OptimizeGraph(const GrapplerItem& item, int64 num_workers, in OptimizeGraph()
|
/external/tensorflow/tensorflow/python/data/experimental/ops/ |
D | distribute.py | 43 def __init__(self, input_dataset, num_workers, index): argument
|
/external/tensorflow/tensorflow/contrib/boosted_trees/lib/quantiles/ |
D | weighted_quantiles_stream_test.cc | 197 int32 num_workers, double eps, int64 max_elements, in TestDistributedStreams() 238 const int32 num_workers = 10; in TEST() local 247 const int32 num_workers = 10; in TEST() local 259 const int32 num_workers = 10; in TEST() local 268 const int32 num_workers = 10; in TEST() local
|
/external/tensorflow/tensorflow/core/kernels/boosted_trees/quantiles/ |
D | weighted_quantiles_stream_test.cc | 196 int32 num_workers, double eps, int64 max_elements, in TestDistributedStreams() 237 const int32 num_workers = 10; in TEST() local 246 const int32 num_workers = 10; in TEST() local 258 const int32 num_workers = 10; in TEST() local 267 const int32 num_workers = 10; in TEST() local
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_ethread.c | 78 static void create_enc_workers(VP9_COMP *cpi, int num_workers) { in create_enc_workers() 140 int num_workers) { in launch_enc_workers() 175 const int num_workers = VPXMIN(cpi->oxcf.max_threads, tile_cols); in vp9_encode_tiles_mt() local 443 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_fp_row_mt() local 529 int num_workers = cpi->num_workers ? cpi->num_workers : 1; in vp9_temporal_filter_row_mt() local 602 int num_workers = VPXMAX(cpi->oxcf.max_threads, 1); in vp9_encode_tiles_row_mt() local
|
/external/libvpx/libvpx/vp9/common/ |
D | vp9_thread_common.c | 170 const int num_workers = VPXMIN(nworkers, VPXMIN(num_tile_cols, sb_rows)); in loop_filter_rows_mt() local 223 int num_workers, VP9LfSync *lf_sync) { in vp9_loop_filter_frame_mt() 243 int num_workers) { in vp9_lpf_mt_init() 280 int width, int num_workers) { in vp9_loop_filter_alloc()
|
D | vp9_thread_common.h | 39 int num_workers; // number of allocated workers. member
|
/external/libaom/libaom/av1/common/ |
D | thread_common.c | 56 int width, int num_workers) { in loop_filter_alloc() 416 const int num_workers = nworkers; in loop_filter_rows_mt() local 477 AVxWorker *workers, int num_workers, in av1_loop_filter_frame_mt() 581 int num_workers, int num_rows_lr, in loop_restoration_alloc() 645 void av1_loop_restoration_dealloc(AV1LrSync *lr_sync, int num_workers) { in av1_loop_restoration_dealloc() 867 const int num_workers = nworkers; in foreach_rest_unit_in_planes_mt() local 910 AVxWorker *workers, int num_workers, in av1_loop_restoration_filter_frame_mt()
|
D | thread_common.h | 47 int num_workers; member 87 int num_workers; member
|
/external/tensorflow/tensorflow/python/training/ |
D | sync_replicas_optimizer_test.py | 35 def get_workers(num_workers, replicas_to_aggregate, workers): argument
|
/external/tensorflow/tensorflow/contrib/distribute/python/ |
D | multi_worker_test_base.py | 69 def _create_cluster(num_workers, argument 133 def create_in_process_cluster(num_workers, argument 176 num_workers=1, argument
|
/external/tensorflow/tensorflow/core/util/ |
D | work_sharder_test.cc | 30 void RunSharding(int64 num_workers, int64 total, int64 cost_per_unit, in RunSharding()
|
/external/v8/tools/testrunner/local/ |
D | pool.py | 104 def __init__(self, num_workers, heartbeat_timeout=1): argument
|