/external/tensorflow/tensorflow/compiler/mlir/xla/transforms/ |
D | lhlo_legalize_to_gpu.cc | 52 ReduceOp reduce_op, ArrayRef<Value> args, in matchAndRewrite() argument 54 auto loc = reduce_op.getLoc(); in matchAndRewrite() 57 for (auto result : reduce_op.out()) { in matchAndRewrite() 69 auto reducing_dimension = *reduce_op.dimensions().int_value_begin(); in matchAndRewrite() 73 for (auto input : reduce_op.operands()) { in matchAndRewrite() 97 for (auto pair : llvm::zip(reduce_op.init_values(), reduce_op.out())) { in matchAndRewrite() 120 auto output = *reduce_op.out().begin(); in matchAndRewrite() 130 auto input_buffer = *reduce_op.operands().begin(); in matchAndRewrite() 138 auto input = *reduce_op.operand_begin(); in matchAndRewrite() 151 mapping.map(reduce_op.body().front().getArgument(0), accumulator); in matchAndRewrite() [all …]
|
/external/tensorflow/tensorflow/python/distribute/ |
D | cross_device_ops.py | 79 reduce_op, value, destinations, num_replicas_in_graph): argument 93 if reduce_op == reduce_util.ReduceOp.MEAN: 101 "the given reduce op %s." % (value, reduce_op)) 194 reduce_op): argument 208 if reduce_op == reduce_util.ReduceOp.MEAN: 211 elif reduce_op != reduce_util.ReduceOp.SUM: 228 def reduce(self, reduce_op, per_replica_value, destinations): argument 260 return self.reduce_implementation(reduce_op, per_replica_value, 263 def batch_reduce(self, reduce_op, value_destination_pairs): argument 305 return self.batch_reduce_implementation(reduce_op, value_destination_pairs) [all …]
|
D | distribute_lib.py | 820 def reduce(self, reduce_op, value, axis): argument 860 if isinstance(reduce_op, six.string_types): 861 reduce_op = reduce_util.ReduceOp(reduce_op.upper()) 863 return self._extended._reduce(reduce_op, value) # pylint: disable=protected-access 864 if reduce_op == reduce_util.ReduceOp.SUM: 867 return self._extended._reduce(reduce_op, value) # pylint: disable=protected-access 868 if reduce_op != reduce_util.ReduceOp.MEAN: 870 "not: %r" % reduce_op) 1158 def reduce(self, reduce_op, value, axis=None): argument 1159 return super(StrategyV1, self).reduce(reduce_op, value, axis) [all …]
|
D | reduce_util.py | 47 reduce_op = mapping.get(aggregation) 48 if not reduce_op: 51 return reduce_op
|
D | cross_device_ops_test.py | 212 reduce_op, batch_reduce): argument 221 reduce_op, [(per_replica, per_replica)]) 223 result = cross_device_ops_instance.reduce(reduce_op, per_replica, 229 if reduce_op == reduce_util.ReduceOp.SUM: 233 assert reduce_op == reduce_util.ReduceOp.MEAN 368 reduce_op=[reduce_util.ReduceOp.SUM, reduce_util.ReduceOp.MEAN], 372 def testIndexedSlicesAllReduce(self, cross_device_ops_instance, reduce_op, argument 376 reduce_op, batch_reduce) 553 def _reduce(test_object, reduce_op, per_replica, destinations): argument 556 return test_object.extended.reduce_to(reduce_op, per_replica, [all …]
|
D | one_device_strategy.py | 182 def reduce(self, reduce_op, value, axis): # pylint: disable=useless-super-delegation argument 213 return super(OneDeviceStrategy, self).reduce(reduce_op, value, axis) 356 def _reduce_to(self, reduce_op, value, destinations): argument 357 del reduce_op, destinations
|
D | central_storage_strategy.py | 181 def reduce(self, reduce_op, value, axis): # pylint: disable=useless-super-delegation argument 245 return super(CentralStorageStrategy, self).reduce(reduce_op, value, axis)
|
D | tpu_strategy.py | 560 def _reduce_to(self, reduce_op, value, destinations): argument 564 if reduce_op == reduce_util.ReduceOp.MEAN: 567 elif reduce_op != reduce_util.ReduceOp.SUM: 578 reduce_op, value, destinations, self._num_replicas_in_sync) 584 if reduce_op == reduce_util.ReduceOp.MEAN: 886 …for name, reduce_op in ctx._last_step_outputs_reduce_ops.items(): # pylint: disable=protected-acc… 893 if reduce_op is not None:
|
D | mirrored_strategy.py | 707 …for name, reduce_op in ctx._last_step_outputs_reduce_ops.items(): # pylint: disable=protected-acc… 711 if reduce_op is None: 791 def _reduce_to(self, reduce_op, value, destinations): argument 793 reduce_op == reduce_util.ReduceOp.MEAN): 802 reduce_op, value, destinations, self._num_replicas_in_sync) 804 reduce_op, value, destinations=destinations) 806 def _batch_reduce_to(self, reduce_op, value_destination_pairs): argument 807 return self._get_cross_device_ops().batch_reduce(reduce_op,
|
D | parameter_server_strategy.py | 469 def _reduce_to(self, reduce_op, value, destinations): argument 474 reduce_op, value, destinations, self._num_replicas_in_sync) 476 reduce_op, value, destinations=destinations) 478 def _batch_reduce_to(self, reduce_op, value_destination_pairs): argument 481 return self._cross_device_ops.batch_reduce(reduce_op,
|
/external/tensorflow/tensorflow/tools/api/golden/v2/ |
D | tensorflow.distribute.-cross-device-ops.pbtxt | 11 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 15 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 27 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 31 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-nccl-all-reduce.pbtxt | 13 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 17 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 29 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 33 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-hierarchical-copy-all-reduce.pbtxt | 13 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 17 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 29 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 33 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-reduction-to-one-device.pbtxt | 12 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 16 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 28 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 32 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-strategy-extended.pbtxt | 23 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 35 …argspec: "args=[\'self\', \'reduce_op\', \'value\', \'destinations\'], varargs=None, keywords=None…
|
/external/tensorflow/tensorflow/tools/api/golden/v1/ |
D | tensorflow.distribute.-cross-device-ops.pbtxt | 11 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 15 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 27 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 31 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-reduction-to-one-device.pbtxt | 12 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 16 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 28 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 32 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-hierarchical-copy-all-reduce.pbtxt | 13 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 17 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 29 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 33 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-nccl-all-reduce.pbtxt | 13 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 17 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 29 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k… 33 …argspec: "args=[\'self\', \'reduce_op\', \'per_replica_value\', \'destinations\'], varargs=None, k…
|
D | tensorflow.distribute.-strategy-extended.pbtxt | 40 …argspec: "args=[\'self\', \'reduce_op\', \'value_destination_pairs\'], varargs=None, keywords=None… 72 …argspec: "args=[\'self\', \'reduce_op\', \'value\', \'destinations\'], varargs=None, keywords=None…
|
/external/tensorflow/tensorflow/python/ops/ragged/ |
D | ragged_math_ops.py | 425 def ragged_reduce_aggregate(reduce_op, argument 472 return reduce_op(rt_input, axis, name=name) 476 return reduce_op(rt_input, axis, name=name, separator=separator) 490 return reduce_op(rt_input.flat_values, None, name=name) 511 inner_reduced = ragged_reduce_aggregate(reduce_op, unsorted_segment_op, 514 return ragged_reduce_aggregate(reduce_op, unsorted_segment_op, 541 ragged_reduce_aggregate(reduce_op, unsorted_segment_op, 550 reduce_op=math_ops.reduce_sum, 560 reduce_op=math_ops.reduce_prod, 571 reduce_op=math_ops.reduce_min, [all …]
|
/external/mesa3d/src/amd/compiler/ |
D | aco_lower_to_hw_instr.cpp | 485 void emit_reduction(lower_context *ctx, aco_opcode op, ReduceOp reduce_op, unsigned cluster_size, P… in emit_reduction() argument 494 identity[0] = Operand(get_reduction_identity(reduce_op, 0)); in emit_reduction() 495 identity[1] = Operand(get_reduction_identity(reduce_op, 1)); in emit_reduction() 527 if (reduce_op == imin8 || reduce_op == imax8) in emit_reduction() 536 if (reduce_op == imin8 || reduce_op == imax8) in emit_reduction() 546 (reduce_op == iadd16 || reduce_op == imax16 || in emit_reduction() 547 reduce_op == imin16 || reduce_op == umin16 || reduce_op == umax16)) { in emit_reduction() 551 if (reduce_op == imin16 || reduce_op == imax16 || reduce_op == iadd16) in emit_reduction() 560 if (reduce_op == imin16 || reduce_op == imax16 || reduce_op == iadd16) in emit_reduction() 579 emit_op(ctx, tmp, vtmp, tmp, PhysReg{0}, reduce_op, src.size()); in emit_reduction() [all …]
|
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
D | TensorArgMax.h | 174 const ReduceOp& reduce_op, 177 : m_xpr(expr), m_reduce_op(reduce_op), m_return_dim(return_dim), m_reduce_dims(reduce_dims) {} 184 const ReduceOp& reduce_op() const { return m_reduce_op; } 224 m_impl(op.expression().index_tuples().reduce(op.reduce_dims(), op.reduce_op()), device),
|
/external/tensorflow/tensorflow/lite/testing/op_tests/ |
D | reduce.py | 27 def make_reduce_tests(reduce_op, argument 144 out = reduce_op(input_tensor, axis=axis, keepdims=parameters["keepdims"])
|
/external/tensorflow/tensorflow/python/keras/engine/ |
D | training_distributed.py | 108 reduce_op = ds_reduce_util.ReduceOp.SUM 112 reduce_op = ds_reduce_util.ReduceOp.MEAN 113 ctx.set_last_step_output(label, output, reduce_op) 346 reduce_op = ds_reduce_util.ReduceOp.SUM 350 reduce_op = ds_reduce_util.ReduceOp.MEAN 351 output_tensors[label] = current_strategy.reduce(reduce_op, output,
|