/external/tensorflow/tensorflow/compiler/xla/ |
D | layout_util.cc | 382 if (ShapeUtil::TupleElementCount(src) != in CopyLayoutInternal() 383 ShapeUtil::TupleElementCount(*dst)) { in CopyLayoutInternal() 387 for (int64 i = 0; i < ShapeUtil::TupleElementCount(src); ++i) { in CopyLayoutInternal() 416 if (!rhs.IsTuple() || ShapeUtil::TupleElementCount(lhs) != in LayoutsInShapesEqual() 417 ShapeUtil::TupleElementCount(rhs)) { in LayoutsInShapesEqual() 420 for (int i = 0; i < ShapeUtil::TupleElementCount(lhs); ++i) { in LayoutsInShapesEqual()
|
D | literal_comparison.cc | 190 const int64 tuple_elements = ShapeUtil::TupleElementCount(shape); in RecursiveElementCount() 718 for (int i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { in EqualHelper() 746 for (int64 i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { in NearHelper() 827 if (ShapeUtil::TupleElementCount(expected) != in EqualShapes() 828 ShapeUtil::TupleElementCount(actual)) { in EqualShapes() 831 ShapeUtil::TupleElementCount(expected), in EqualShapes() 832 ShapeUtil::TupleElementCount(actual)); in EqualShapes()
|
D | shape_util.cc | 335 return shape.IsTuple() && TupleElementCount(shape) == 0; in IsEmptyTuple() 338 /* static */ int64 ShapeUtil::TupleElementCount(const Shape& shape) { in TupleElementCount() function in xla::ShapeUtil 346 CHECK_GT(TupleElementCount(shape), index); in GetTupleElementShape() 362 CHECK_LE(start, TupleElementCount(tuple)); in SliceTuple() 363 CHECK_LE(limit, TupleElementCount(tuple)); in SliceTuple() 863 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ForEachSubshapeHelper() 880 for (int64 i = 0; i < ShapeUtil::TupleElementCount(*shape); ++i) { in ForEachMutableSubshapeHelper()
|
D | shape_tree.h | 399 int64 count = ShapeUtil::TupleElementCount(shape); in CountSubshapes() 411 const int64 size = ShapeUtil::TupleElementCount(shape); in InitChildren() 447 const int64 size = ShapeUtil::TupleElementCount(shape); in InitChildren()
|
D | literal.cc | 112 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in SetPiece() 317 ShapeUtil::TupleElementCount(piece->subshape())) { in CreateFromProto() 320 ShapeUtil::TupleElementCount(piece->subshape()), in CreateFromProto() 341 for (int i = 0; i < ShapeUtil::TupleElementCount(shape()); ++i) { in DecomposeTuple() 1057 for (int i = 0; i < ShapeUtil::TupleElementCount(subshape); ++i) { in TupleToStringHelper() 1435 for (int i = 0; i < ShapeUtil::TupleElementCount(shape()); ++i) { in ConvertToShape() 2132 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CopyPieceSubtree() 2246 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in BuildPieceSubtree() 2275 CHECK_EQ(src_buf_ptrs.size(), ShapeUtil::TupleElementCount(*shape_)); in BorrowingLiteral()
|
D | shape_util.h | 469 static int64 TupleElementCount(const Shape& shape);
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_element_type_converter.cc | 69 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in GetConvertedTupleShape() 90 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ConvertTupleElements()
|
D | hlo_module_dce.cc | 62 ShapeUtil::TupleElementCount(xla_while->shape()); in RunWhileDCE()
|
D | gather_expander_test.cc | 93 ASSERT_EQ(ShapeUtil::TupleElementCount(while_shape), 4); in TEST()
|
D | transfer_manager.cc | 253 for (int64 i = 0; i < ShapeUtil::TupleElementCount(device_subshape); in WriteTupleIndexTablesAsync() 276 i < ShapeUtil::TupleElementCount(device_buffer.on_device_shape()); ++i) { in WriteRootTupleIndexTable()
|
D | generic_transfer_manager.cc | 46 TF_RET_CHECK(elements.size() == ShapeUtil::TupleElementCount(shape)); in WriteSingleTupleIndexTable()
|
D | allocation_tracker.cc | 159 i < ShapeUtil::TupleElementCount(shaped_buffer->on_device_shape()); in DeconstructTuple()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | shape_util.cc | 31 int64 tuple_elements = xla::ShapeUtil::TupleElementCount(shape); in PopulateInfeedLayoutVector() 135 int64 tuple_elements = xla::ShapeUtil::TupleElementCount(input_shape); in GetShapeWithLayout()
|
D | xla_jit_compiled_cpu_function_test.cc | 127 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 1); in TEST()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | cpu_transfer_manager.cc | 114 buffers.reserve(ShapeUtil::TupleElementCount(shape)); in TransferLiteralToInfeed() 121 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in TransferLiteralToInfeed()
|
/external/tensorflow/tensorflow/compiler/xla/python/ |
D | numpy_bridge.cc | 136 int num_elements = ShapeUtil::TupleElementCount(shape); in PyShapeInfoFromXlaShape() 137 dimensions = make_safe(PyTuple_New(ShapeUtil::TupleElementCount(shape))); in PyShapeInfoFromXlaShape() 372 int num_elements = ShapeUtil::TupleElementCount(literal.shape()); in PyObjectFromXlaLiteral()
|
D | xrt.cc | 274 for (int32 i = 0; i < ShapeUtil::TupleElementCount(tuple_shape); ++i) { in DestructureXrtAllocationTuple()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | infeed_thunk.cc | 77 const int64 tuple_element_count = ShapeUtil::TupleElementCount(shape); in ExecuteOnStream()
|
D | gpu_transfer_manager.cc | 134 const int64 tuple_element_count = ShapeUtil::TupleElementCount(shape); in ShapeTreeToLiteral()
|
/external/tensorflow/tensorflow/c/eager/ |
D | c_api_debug.cc | 87 if (xla::ShapeUtil::TupleElementCount(padded_shape) != 2) { in TFE_TensorHandleTensorDebugInfo()
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | client_test.cc | 100 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.shape())); in XLA_TEST_F()
|
D | local_client_execute_test.cc | 209 EXPECT_EQ(3, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F() 237 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F() 315 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
D | tuple_ops.cc | 53 for (int i = 0; i < ShapeUtil::TupleElementCount(select.GetShape()); ++i) { in EmitTupleSelect()
|
D | fused_ir_emitter.cc | 180 for (size_t i = 0; i < ShapeUtil::TupleElementCount(tuple->shape()); ++i) { in HandleTuple()
|
/external/tensorflow/tensorflow/compiler/xrt/kernels/ |
D | xrt_execute_op.cc | 253 xla::ShapeUtil::TupleElementCount(output_tuple->on_device_shape()); in DoWork()
|