/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_module_dce_test.cc | 198 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F() 308 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F() 309 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[1]->shape())); in TEST_F() 375 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F() 376 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[1]->shape())); in TEST_F() 512 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
|
D | transfer_manager.cc | 295 ShapeUtil::TupleElementCount(device_subshape) > 0) { in WriteTupleIndexTablesAsync() 302 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(device_subshape); in WriteTupleIndexTablesAsync() 319 if (ShapeUtil::TupleElementCount(device_buffer.on_device_shape()) == 0) { in WriteRootTupleIndexTable() 328 i < ShapeUtil::TupleElementCount(device_buffer.on_device_shape()); ++i) { in WriteRootTupleIndexTable() 338 if (ShapeUtil::TupleElementCount(buffer_tree.shape()) == 0) { in WriteRootTupleIndexTable() 347 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(buffer_tree.shape()); in WriteRootTupleIndexTable()
|
D | conditional_simplifier_test.cc | 321 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 0); in TEST_F() 367 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F() 413 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F() 476 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
|
D | hlo_element_type_converter.cc | 69 const int64_t n = ShapeUtil::TupleElementCount(shape); in GetConvertedTupleShape() 92 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ConvertTupleElements()
|
D | hlo_module_dce.cc | 66 ShapeUtil::TupleElementCount(xla_while->shape()); in RunWhileDCE()
|
D | gather_expander_test.cc | 102 ASSERT_EQ(ShapeUtil::TupleElementCount(while_shape), 4); in TEST_F()
|
D | generic_transfer_manager.cc | 45 TF_RET_CHECK(elements.size() == ShapeUtil::TupleElementCount(shape)); in WriteSingleTupleIndexTable()
|
D | conditional_simplifier.cc | 90 ShapeUtil::TupleElementCount(param->shape()); in TryRemoveUnusedConditionalOperands() 560 ShapeUtil::TupleElementCount(f->shape()); in TryRemoveConditional()
|
D | allocation_tracker.cc | 148 const auto n = ShapeUtil::TupleElementCount(shaped_buffer->on_device_shape()); in DeconstructTuple()
|
D | hlo_verifier.cc | 543 return ShapeUtil::TupleElementCount(shape) != in CheckBufferOffset() 586 if (ShapeUtil::TupleElementCount(input_buffer_shape) != in CheckInplaceCollectivePermute() 587 ShapeUtil::TupleElementCount(output_buffer_shape)) { in CheckInplaceCollectivePermute() 591 ShapeUtil::TupleElementCount(input_offset_shape) != in CheckInplaceCollectivePermute() 592 ShapeUtil::TupleElementCount(input_buffer_shape)) { in CheckInplaceCollectivePermute() 604 ShapeUtil::TupleElementCount(output_offset_shape) != in CheckInplaceCollectivePermute() 605 ShapeUtil::TupleElementCount(output_buffer_shape)) { in CheckInplaceCollectivePermute()
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
D | tensor_list_utils.cc | 128 *is_nested_list = (xla::ShapeUtil::TupleElementCount(list_shape) > 2); in IsNestedTensorList() 167 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in GetTensorListPushIndex() 180 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in SetTensorListPushIndex() 230 int tuple_size = xla::ShapeUtil::TupleElementCount(element_tensor_list_shape); in GetTensorListShapeFromElementTensorListShape() 275 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in CreateZerosTensorListWithShape() 365 int list_tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in ExecuteTensorListPushBack() 372 int element_tuple_size = xla::ShapeUtil::TupleElementCount(element_shape); in ExecuteTensorListPushBack() 430 int list_tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in ExecuteTensorListPopBack()
|
/external/tensorflow/tensorflow/compiler/xla/ |
D | layout_util.cc | 434 if (ShapeUtil::TupleElementCount(src) != in CopyLayoutInternal() 435 ShapeUtil::TupleElementCount(*dst)) { in CopyLayoutInternal() 439 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(src); ++i) { in CopyLayoutInternal() 468 if (!rhs.IsTuple() || ShapeUtil::TupleElementCount(lhs) != in LayoutsInShapesEqual() 469 ShapeUtil::TupleElementCount(rhs)) { in LayoutsInShapesEqual() 472 for (int i = 0; i < ShapeUtil::TupleElementCount(lhs); ++i) { in LayoutsInShapesEqual()
|
D | literal_comparison.cc | 208 const int64_t tuple_elements = ShapeUtil::TupleElementCount(shape); in RecursiveElementCount() 692 for (int i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { in EqualHelper() 788 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); in NearHelper() 876 if (ShapeUtil::TupleElementCount(expected) != in EqualShapes() 877 ShapeUtil::TupleElementCount(actual)) { in EqualShapes() 880 ShapeUtil::TupleElementCount(expected), in EqualShapes() 881 ShapeUtil::TupleElementCount(actual)); in EqualShapes()
|
D | shape_util.cc | 554 return shape.IsTuple() && TupleElementCount(shape) == 0; in IsEmptyTuple() 557 /* static */ int64_t ShapeUtil::TupleElementCount(const Shape& shape) { in TupleElementCount() function in xla::ShapeUtil 565 CHECK_GT(TupleElementCount(shape), index); in GetTupleElementShape() 581 CHECK_LE(start, TupleElementCount(tuple)); in SliceTuple() 582 CHECK_LE(limit, TupleElementCount(tuple)); in SliceTuple() 1069 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ForEachSubshapeHelper() 1086 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(*shape); ++i) { in ForEachMutableSubshapeHelper()
|
D | literal.cc | 232 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in mutable_shape_do_not_use() 484 ShapeUtil::TupleElementCount(piece->subshape())) { in CreateFromProto() 487 ShapeUtil::TupleElementCount(piece->subshape()), in CreateFromProto() 524 const auto tuple_element_count = ShapeUtil::TupleElementCount(shape()); in DecomposeTuple() 1307 const auto tuple_element_count = ShapeUtil::TupleElementCount(subshape); in TupleToStringHelper() 1309 for (int i = 0; i < ShapeUtil::TupleElementCount(subshape); ++i) { in TupleToStringHelper() 1736 const auto tuple_element_count = ShapeUtil::TupleElementCount(shape()); in ConvertToShape() 2565 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CopyPieceSubtree() 2657 CHECK_EQ(src_buf_ptrs.size(), ShapeUtil::TupleElementCount(*shape_)); in MutableBorrowingLiteral() 2687 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in BuildPieceSubtree() [all …]
|
/external/tensorflow/tensorflow/core/tpu/kernels/ |
D | tpu_reshard_variables_op_util.cc | 106 if (var_list.size() != xla::ShapeUtil::TupleElementCount(input_host_shape)) { in BuildInputBuffers() 110 xla::ShapeUtil::TupleElementCount(input_host_shape)); in BuildInputBuffers() 233 xla::ShapeUtil::TupleElementCount(result_buffers.on_host_shape()); in UpdateOutputVariables()
|
D | tpu_execute_op.cc | 203 if (arg_list.size() != xla::ShapeUtil::TupleElementCount(input_host_shape)) { in BuildComputationInputs() 207 xla::ShapeUtil::TupleElementCount(input_host_shape)); in BuildComputationInputs() 404 xla::ShapeUtil::TupleElementCount(scoped_buffers.on_host_shape()); in AllocateOutputTensors()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | shape_util.cc | 31 int64_t tuple_elements = xla::ShapeUtil::TupleElementCount(shape); in PopulateInfeedLayoutVector() 177 int64_t tuple_elements = xla::ShapeUtil::TupleElementCount(input_shape); in GetShapeWithLayout()
|
D | xla_jit_compiled_cpu_function_test.cc | 227 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 1); in TEST() 278 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 2); in TEST()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | cpu_xfeed.cc | 197 buffers.reserve(ShapeUtil::TupleElementCount(shape)); in TransferLiteralToInfeedOnCpu() 204 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in TransferLiteralToInfeedOnCpu()
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
D | client_test.cc | 99 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.shape())); in XLA_TEST_F()
|
D | local_client_execute_test.cc | 212 EXPECT_EQ(3, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F() 240 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F() 318 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | gpu_fusible.cc | 622 return ShapeUtil::TupleElementCount(root->shape()); in GetOutputSizeOfFusible()
|
/external/tensorflow/tensorflow/compiler/xla/service/spmd/ |
D | spmd_partitioner_util.h | 60 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CreateConstantBase()
|
/external/tensorflow/tensorflow/compiler/xrt/ |
D | xrt_util.cc | 354 xla::ShapeUtil::TupleElementCount(output_tuple->on_device_shape()); in CreateExecuteOutput()
|