/external/tensorflow/tensorflow/compiler/xla/service/ |
D | optimize_input_output_buffer_alias.cc | 52 input_shape, [&](const Shape& subshape, const ShapeIndex& index) { in Build() 63 output_shape, [&](const Shape& subshape, const ShapeIndex& index) { in Build()
|
D | bfloat16_propagation.cc | 53 root->shape(), [&](const Shape& subshape, const ShapeIndex& index) { in DetermineFusionComputationPrecision() 101 inst->shape(), [&](const Shape& subshape, const ShapeIndex& index) { in RevertIfFusionInternalBF16Changes() 178 const Shape& subshape, const ShapeIndex& index) { in DetermineWhileComputationsPrecision() 209 const Shape& subshape = ShapeUtil::GetSubshape(hlo.shape(), index); in AllUsersConsumeBF16() local 797 auto subshape = entry.first; in Run() local 830 Shape* subshape = ShapeUtil::GetMutableSubshape(hlo->mutable_shape(), index); in OutputTypeAfterChange() local
|
D | while_loop_invariant_code_motion.cc | 223 [&input_size](const Shape& subshape, const ShapeIndex& /*index*/) { in TryHoistingInvariantInstructionsFromWhileBody() 231 [&output_size](const Shape& subshape, const ShapeIndex& /*index*/) { in TryHoistingInvariantInstructionsFromWhileBody()
|
D | instruction_fusion.cc | 176 [&output_rank](const Shape& subshape, const ShapeIndex& shape_index) { in EffectivelyAtMostUnary() 275 [&size](const Shape& subshape, const ShapeIndex& shape_index) { in ComputeGloballyUnfusible()
|
D | layout_assignment.cc | 305 const ShapeIndex& index) -> Status { in SetInstructionLayout() 1313 [&](const Shape& subshape, const ShapeIndex& shape_index) { in PropagateOperandConstraint() 1350 user->shape(), [&](const Shape& subshape, const ShapeIndex& shape_index) { in PropagateOperandConstraint() 1491 const Shape& subshape = ShapeUtil::GetSubshape(instruction->shape(), index); in InferArrayLayout() local 1634 [instruction, &constraints](Shape* subshape, const ShapeIndex& index) { in AssignLayouts() 2086 return absl::c_all_of(shape.tuple_shapes(), [](const Shape& subshape) { in IsAtMostRank1()
|
D | hlo_element_type_converter.cc | 70 Shape subshape = ShapeUtil::GetTupleElementShape(shape, i); in GetConvertedTupleShape() local
|
D | shaped_buffer.cc | 86 [this, &s](const Shape& subshape, const ShapeIndex& index) { in ToString()
|
D | generic_transfer_manager.cc | 85 [&](const Shape& subshape, const ShapeIndex& index) -> Status { in TransferLiteralFromDeviceInternal()
|
D | bfloat16_normalization.cc | 239 auto subshape = ShapeUtil::GetMutableSubshape(hlo->mutable_shape(), {i}); in HandleMultipleOutputs() local
|
D | hlo_verifier.cc | 33 shape, [](const Shape& subshape, const ShapeIndex&) -> Status { in VerifyNotSparse() 788 [&](const Shape& subshape, const ShapeIndex& index) { in CheckMixedPrecisionOperands() 1056 shape, [&contains_token](const Shape& subshape, const ShapeIndex&) { in ShapeContainsToken()
|
D | while_loop_simplifier.cc | 593 const Shape& subshape = desired_shape.tuple_shapes(i); in UnflattenTupleInstr() local 631 const Shape& subshape = shape.tuple_shapes(i); in GetFlatTupleElems() local
|
D | hlo_cost_analysis.cc | 584 [&](const Shape& subshape, const ShapeIndex&) { in HandleAllReduce() 625 [this](const Shape& subshape, const ShapeIndex& /*shape_index*/) { in HandleFusion()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | gpu_hlo_support_checker.cc | 31 [&instruction](const Shape& subshape, const ShapeIndex&) { in Run()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | cpu_hlo_support_checker.cc | 31 [&instruction](const Shape& subshape, const ShapeIndex&) { in Run()
|
/external/tensorflow/tensorflow/compiler/xla/ |
D | literal.cc | 113 const Shape& subshape = shape.tuple_shapes(i); in SetPiece() local 582 Shape* subshape = ShapeUtil::GetMutableSubshape(&new_shape, shape_index); in Relayout() local 598 [this, &result](const Shape& subshape, const ShapeIndex& index) { in Relayout() 790 const Shape& subshape = ShapeUtil::GetSubshape(shape(), shape_index); in GetAsString() local 835 const Shape& subshape = ShapeUtil::GetSubshape(shape(), shape_index); in GetSparseElementAsString() local 922 shape(), [&](const Shape& subshape, const ShapeIndex& index) { in Hash() 1054 const Shape& subshape = ShapeUtil::GetSubshape(literal.shape(), shape_index); in TupleToStringHelper() local 1070 const Shape& subshape, bool print_shape, in SparseArrayToStringHelper() 1098 const Shape& subshape = ShapeUtil::GetSubshape(literal.shape(), shape_index); in DenseArrayToStringHelper() local 1171 const Shape& subshape = ShapeUtil::GetSubshape(literal.shape(), shape_index); in ToStringHelper() local [all …]
|
D | literal_util.cc | 51 &result_shape, [](Shape* subshape, const ShapeIndex&) { in ConvertType() 64 [&](const Shape& subshape, const ShapeIndex& shape_index) { in ConvertType()
|
D | literal.h | 360 const Shape& subshape() const { return *subshape_; } in subshape() function 361 void set_subshape(const Shape* subshape) { subshape_ = subshape; } in set_subshape() 962 const Shape& subshape = p.subshape(); in AppendSparseElement() local
|
D | shape.cc | 88 for (const Shape& subshape : tuple_shapes_) { in is_static() local
|
D | shape_util.cc | 771 const Shape* subshape = &shape; in IndexIsValid() local 827 for (const Shape& subshape : shape.tuple_shapes()) { in GetLeafCount() local 897 [&func](const Shape& subshape, const ShapeIndex& index) { in ForEachSubshape() 910 [&func](Shape* subshape, const ShapeIndex& index) { in ForEachMutableSubshape() 1449 for (const Shape& subshape : shape.tuple_shapes()) { in Hash() local
|
D | shape_util_test.cc | 479 shape, [&calls, &shape](const Shape& subshape, const ShapeIndex& index) { in TEST() 494 shape, [&calls, &shape](const Shape& subshape, const ShapeIndex& index) { in TEST() 516 &shape, [&calls, &shape](const Shape* subshape, const ShapeIndex& index) { in TEST()
|
/external/tensorflow/tensorflow/core/ops/ |
D | math_ops.cc | 991 ShapeHandle subshape; in SegmentReductionShapeFn() local 1015 ShapeHandle subshape; in SparseSegmentReductionShapeFn() local 1039 ShapeHandle subshape; in SparseSegmentReductionGradShapeFn() local 1080 ShapeHandle subshape; in SparseSegmentReductionWithNumSegmentsShapeFn() local
|
/external/tensorflow/tensorflow/compiler/jit/ |
D | xla_tensor.cc | 58 xla::Shape subshape = in AllocateShapedBuffer() local
|
/external/tensorflow/tensorflow/compiler/xrt/ |
D | xrt_state.cc | 112 xla::Shape subshape = in AllocateScopedShapedBuffer() local 296 XRTTupleAllocation* parent, const xla::ShapeIndex& subshape, in MakeSubBuffer() 411 xla::Shape subshape = in MakeTuple() local
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | shape_util.cc | 33 const xla::Shape& subshape = in PopulateInfeedLayoutVector() local
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | client_library_test_base.cc | 332 &layout_shape, [&](Shape* subshape, const ShapeIndex& /*index*/) { in ComputeAndCompareLiteralWithStatus() 389 &layout_shape, [&](Shape* subshape, const ShapeIndex& /*index*/) { in ComputeAndCompareLiteralWithStatus() 624 [](Shape* subshape, const ShapeIndex&) { in MaybeConvertShapeToBfloat16()
|