/external/tensorflow/tensorflow/python/distribute/ |
D | collective_all_reduce_strategy_test.py | 331 combinations.combine(mode=['graph'], required_gpus=[0, 1, 2])) 332 def testMinimizeLossGraph(self, required_gpus): argument 334 self._cluster_spec, required_gpus) 337 combinations.combine(mode=['graph'], required_gpus=[0, 1, 2])) 338 def testVariableInitialization(self, required_gpus): argument 342 num_gpus=required_gpus) 346 mode=['graph'], required_gpus=[0, 1, 2], use_dataset=[True, False])) 347 def testMakeInputFnIterator(self, required_gpus, use_dataset): argument 348 def _worker_fn(task_type, task_id, required_gpus): argument 357 devices_per_worker = max(1, required_gpus) [all …]
|
D | cross_device_ops_test.py | 314 required_gpus=[0, 1, 2], 322 def testAllReduceDense(self, num_processes, required_gpus, implementation, argument 324 if (required_gpus == 0 and 333 gpus_per_process=required_gpus, 355 required_gpus=[0, 1, 2], 364 def testAllReduceSparse(self, num_processes, required_gpus, implementation, argument 366 if (required_gpus == 0 and 376 gpus_per_process=required_gpus, 436 required_gpus=[0, 1, 2], 444 def testBatchAllReduceDense(self, num_processes, required_gpus, argument [all …]
|
D | strategy_combinations.py | 128 def _get_multi_worker_mirrored_creator(required_gpus): argument 142 num_accelerators={"GPU": required_gpus}, 207 initializer=_get_multi_worker_mirrored_creator(required_gpus=0)) 211 initializer=_get_multi_worker_mirrored_creator(required_gpus=0)) 218 required_gpus=None) 220 "OneDeviceCPU", lambda: OneDeviceStrategy("/cpu:0"), required_gpus=None) 222 "OneDeviceGPU", lambda: OneDeviceStrategy("/gpu:0"), required_gpus=1) 226 required_gpus=None) 230 required_gpus=1) 258 required_gpus=1) [all …]
|
D | parameter_server_strategy_test.py | 622 combinations.combine(mode=['graph'], required_gpus=[0, 1, 2])) 623 def testLocalSimpleIncrement(self, required_gpus): argument 624 self._test_simple_increment(None, 0, required_gpus) 627 combinations.combine(mode=['graph'], required_gpus=[0, 1, 2])) 628 def testMinimizeLossGraphDistributed(self, required_gpus): argument 630 self._cluster_spec, required_gpus) 633 combinations.combine(mode=['graph'], required_gpus=[0, 1, 2])) 634 def testMinimizeLossGraphLocal(self, required_gpus): argument 635 self._test_minimize_loss_graph(None, None, required_gpus) 640 mode=['graph'], required_gpus=[1, 2], use_dataset=[True, False])) [all …]
|
D | combinations.py | 166 required_gpus = kwargs.get("required_gpus", None) 168 if distributions and required_gpus: 172 number_of_required_gpus = max([required_gpus or 0] + 173 [d.required_gpus or 0 for d in distributions]) 256 required_gpus=None, argument 280 self.required_gpus = required_gpus
|
D | cross_device_utils_test.py | 87 required_gpus=1)) 101 required_gpus=1))
|
D | remote_mirrored_strategy_eager_test.py | 45 required_gpus=1)
|
D | mirrored_strategy_test.py | 647 required_gpus=2) 1155 required_gpus=1) 1254 required_gpus=1)
|
D | mirrored_variable_test.py | 70 required_gpus=0)
|
/external/tensorflow/tensorflow/tools/docs/ |
D | tf_doctest.py | 132 def setup_gpu(required_gpus): argument 145 if required_gpus == 0: 150 if len(available_gpus) >= required_gpus: 151 tf.config.set_visible_devices(available_gpus[:required_gpus]) 156 num_logical_gpus = required_gpus - len(available_gpus) + 1 220 setup_gpu(FLAGS.required_gpus)
|
D | BUILD | 92 "--required_gpus=2",
|
/external/tensorflow/tensorflow/python/distribute/v1/ |
D | cross_device_ops_test.py | 367 required_gpus=1)) 398 required_gpus=1)) 726 required_gpus=[0, 1, 2], 729 def testReductionDistributed(self, required_gpus, use_strategy_object, argument 735 required_gpus, 743 required_gpus=[0, 1, 2], 745 def testReduceIndexedSlicesDistributed(self, required_gpus, variable_length): argument 749 required_gpus, 758 required_gpus=2, 763 def testReductionLocal(self, required_gpus, communication, argument [all …]
|
/external/tensorflow/tensorflow/python/keras/distribute/ |
D | worker_training_state_test.py | 39 required_gpus=[0, 1],
|
D | mirrored_variable_test.py | 53 required_gpus=0)
|
D | multi_worker_test.py | 193 required_gpus=[0, 1]))
|
D | keras_correctness_test_base.py | 86 s for s in all_strategies if s.required_tpu or s.required_gpus or
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | collective_ops_multi_worker_test.py | 62 combinations.combine(device="CPU", communication="RING", required_gpus=0) + 64 device="GPU", communication=["RING", "NCCL"], required_gpus=1))
|
D | collective_ops_test.py | 92 combinations.combine(device='CPU', communication='RING', required_gpus=0) + 94 device='GPU', communication=['RING', 'NCCL'], required_gpus=2))
|