/external/tensorflow/tensorflow/contrib/distribute/python/ |
D | moving_averages_test.py | 44 replica_id = [0] 48 val = constant_op.constant([1.0 + replica_id[0], 2.0 - replica_id[0]]) 49 replica_id[0] += 1 61 average_val = [1.0 + 0.5 * (replica_id[0] - 1), 62 2.0 - 0.5 * (replica_id[0] - 1)] 71 replica_id = [0] 75 val = constant_op.constant([1.0 + replica_id[0], 2.0 - replica_id[0]]) 76 replica_id[0] += 1 87 average_val = [1.0 + 0.5 * (replica_id[0] - 1), 88 2.0 - 0.5 * (replica_id[0] - 1)]
|
D | keras_optimizer_v2_test.py | 66 replica_id = _replica_id() 67 return math_ops.cast(replica_id + 1, dtype=dtypes.float32) * 0.5 * var 138 replica_id = ds_context.get_replica_context().replica_id_in_sync_group 139 if not isinstance(replica_id, ops.Tensor): 140 replica_id = constant_op.constant(replica_id) 141 return replica_id
|
D | parameter_server_strategy_test.py | 64 replica_id = ds_context.get_replica_context().replica_id_in_sync_group 65 if isinstance(replica_id, ops.Tensor): 66 replica_id = tensor_util.constant_value(replica_id) 67 return replica_id 168 replica_id = _get_replica_id_integer() 169 last_part_device = ('device:GPU:%d' % replica_id) 314 replica_id = _get_replica_id_integer() 315 replica_compute_device = ('/device:GPU:%d' % replica_id) 322 replica_id = _get_replica_id_integer() 323 replica_variable_device = ('/device:GPU:%d' % replica_id)
|
D | mirrored_strategy_multigpu_test.py | 94 replica_id = int(self.evaluate(_replica_id())) 97 return list(range(replica_id)) 395 replica_id = self.evaluate(_replica_id()) 396 v = variable_scope.variable(1.0, name="foo_" + str(replica_id)) 663 replica_id = self.evaluate(_replica_id()) 674 updates = [v_sum.assign_add(2.0 + replica_id), 675 v_mean.assign(6.0 * replica_id)] 676 all_v_sum[replica_id] = v_sum 677 all_v_mean[replica_id] = v_mean 680 components_sum[replica_id] = c_sum [all …]
|
D | strategy_test_lib.py | 223 replica_id = ds_context.get_replica_context().replica_id_in_sync_group 224 return summary_ops.scalar("a", replica_id) 255 replica_id = self.evaluate( 257 self.assertLess(replica_id, len(d.extended.worker_devices)) 258 self.assertFalse(expected_devices[replica_id]) 259 expected_devices[replica_id] = True
|
/external/tensorflow/tensorflow/python/distribute/ |
D | summary_op_util.py | 45 replica_id = replica_context.replica_id_in_sync_group 46 if isinstance(replica_id, ops.Tensor): 47 replica_id = tensor_util.constant_value(replica_id) 48 return replica_id and replica_id > 0
|
D | values.py | 91 def is_device_in_replica(self, device, replica_id): argument 153 def is_device_in_replica(self, device, replica_id): argument 201 replica_id = replica_context.replica_id_in_sync_group 202 if not isinstance(replica_id, int): 203 replica_id = tensor_util.constant_value(replica_id) 204 return values[replica_id] 211 replica_id = self._device_to_replica.get(device) 212 if replica_id is None: 215 return values[replica_id] 217 def is_device_in_replica(self, device, replica_id): argument [all …]
|
D | mirrored_strategy.py | 805 def __init__(self, dist, coord, replica_id, device_map, variable_creator_fn, argument 811 self.replica_id = replica_id 858 if self.replica_id > 0: 861 self._name_scope += "replica_%d/" % self.replica_id 877 self.replica_id, dtypes.int32)), \ 879 self.replica_id]), \ 882 self._captured_var_scope, reuse=self.replica_id > 0), \ 962 replica_id = tensor_util.constant_value(self._replica_id_in_sync_group) 963 return [self._strategy.extended.worker_devices_by_replica[replica_id]]
|
D | tpu_strategy.py | 167 def replicated_fn(replica_id, replica_args, replica_kwargs): argument 169 with _TPUReplicaContext(self, replica_id_in_sync_group=replica_id): 325 for replica_id in range(self._num_replicas_in_sync): 326 …select_replica = lambda x: values.select_replica(replica_id, x) # pylint: disable=cell-var-from-l… 639 replica_id = tensor_util.constant_value(self._replica_id_in_sync_group) 641 if replica_id is None: # Non-constant `Tensor` inside `tpu.replicate`. 645 return (ds.extended.worker_devices[replica_id],)
|
/external/tensorflow/tensorflow/python/tpu/ |
D | tpu_context.py | 142 def device_for_replica(self, replica_id): argument 156 return self._internal_ctx.device_for_replica(replica_id) 527 …def _placement_function(_sentinal=None, replica_id=None, host_id=None): # pylint: disable=invalid… argument 530 if replica_id is not None and host_id is not None: 537 if replica_id is not None: 540 replica=replica_id, job=master) 542 host_id = replica_id / self.num_of_cores_per_host 674 def device_for_replica(self, replica_id): argument 689 replica=replica_id, job=master), 690 self.device_assignment.tpu_ordinal(replica=replica_id)) [all …]
|
D | tensor_tracer.py | 1178 for replica_id in range(0, self._num_replicas): 1181 _COMPACT_TRACE_FILE_PREFIX) + '%d'%replica_id 1254 def _make_flush_fun(replica_id): argument 1263 if isinstance(replica_id, str): 1264 replica_id_str = replica_id 1266 replica_id_str = '%d'%replica_id 1298 def _f(replica_id): argument 1299 return _make_flush_fun(replica_id)
|
D | tpu_feed.py | 848 replica_id = self._device_assignment.lookup_replicas( 851 replica=replica_id, logical_core=logical_core)
|
D | tpu_estimator.py | 973 def tpu_ordinal_function_impl(replica_id): argument 975 return ctx.device_assignment.tpu_ordinal(replica=replica_id) 977 return replica_id % num_replicas_per_host 979 def device_function_impl(replica_id): argument 980 return ctx.tpu_host_placement_function(replica_id=replica_id) 1899 with ops.device(self._ctx.tpu_host_placement_function(replica_id=0)):
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | allocation_tracker.cc | 185 const GlobalDataHandle& data, int replica_id) const { in ResolveForReplica() 189 if (replica_id >= replicated_buffers.size()) { in ResolveForReplica() 193 replica_id, replicated_buffers.size()); in ResolveForReplica() 195 return replicated_buffers[replica_id]; in ResolveForReplica()
|
D | service.cc | 1012 if (arg->replica_id() < 0 || arg->replica_id() >= replica_count) { in TransferToInfeed() 1015 StrCat("The replica_id=", arg->replica_id(), in TransferToInfeed() 1024 executor = replicas[arg->replica_id()]; in TransferToInfeed() 1029 executor = replicas[arg->replica_id()]; in TransferToInfeed() 1041 if (arg->replica_id() < 0 || arg->replica_id() >= replica_count) { in TransferFromOutfeed() 1044 arg->replica_id(), replica_count); in TransferFromOutfeed() 1051 executor = replicas[arg->replica_id()]; in TransferFromOutfeed() 1056 executor = replicas[arg->replica_id()]; in TransferFromOutfeed()
|
D | allocation_tracker.h | 74 int replica_id) const;
|
/external/tensorflow/tensorflow/compiler/xla/client/ |
D | client.h | 152 Status TransferToInfeed(const LiteralSlice& literal, int64 replica_id = 0, 161 const Shape* shape_with_layout, int64 replica_id = 0,
|
D | client.cc | 96 Status Client::TransferToInfeed(const LiteralSlice& literal, int64 replica_id, in TransferToInfeed() argument 103 request.set_replica_id(replica_id); in TransferToInfeed() 119 const Shape* shape_with_layout, int64 replica_id, in TransferFromOutfeed() argument 125 request.set_replica_id(replica_id); in TransferFromOutfeed()
|
/external/tensorflow/tensorflow/compiler/xla/ |
D | xla.proto | 339 int64 replica_id = 2; field 350 int64 replica_id = 2; field
|
/external/tensorflow/tensorflow/python/framework/ |
D | ops_test.py | 1372 def __init__(self, graph, replica_id): argument 1375 self._replica_id = replica_id 1429 def __init__(self, graph, replica_id, op_to_colocate_with): argument 1430 super(ColocatingThread, self).__init__(graph, replica_id) 1494 def __init__(self, graph, replica_id, dependency_op): argument 1495 super(DependingThread, self).__init__(graph, replica_id)
|