Home
last modified time | relevance | path

Searched refs:TupleElementCount (Results 1 – 25 of 45) sorted by relevance

12

/external/tensorflow/tensorflow/compiler/xla/service/
Dhlo_module_dce_test.cc199 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
309 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
310 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[1]->shape())); in TEST_F()
376 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
377 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[1]->shape())); in TEST_F()
513 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
Dtransfer_manager.cc301 ShapeUtil::TupleElementCount(device_subshape) > 0) { in WriteTupleIndexTablesAsync()
308 for (int64 i = 0; i < ShapeUtil::TupleElementCount(device_subshape); in WriteTupleIndexTablesAsync()
325 if (ShapeUtil::TupleElementCount(device_buffer.on_device_shape()) == 0) { in WriteRootTupleIndexTable()
334 i < ShapeUtil::TupleElementCount(device_buffer.on_device_shape()); ++i) { in WriteRootTupleIndexTable()
344 if (ShapeUtil::TupleElementCount(buffer_tree.shape()) == 0) { in WriteRootTupleIndexTable()
353 for (int64 i = 0; i < ShapeUtil::TupleElementCount(buffer_tree.shape()); in WriteRootTupleIndexTable()
Dconditional_simplifier_test.cc322 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 0); in TEST_F()
368 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
414 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
477 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
Dhlo_element_type_converter.cc69 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in GetConvertedTupleShape()
90 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ConvertTupleElements()
Dhlo_module_dce.cc65 ShapeUtil::TupleElementCount(xla_while->shape()); in RunWhileDCE()
Dgather_expander_test.cc102 ASSERT_EQ(ShapeUtil::TupleElementCount(while_shape), 4); in TEST_F()
Dgeneric_transfer_manager.cc45 TF_RET_CHECK(elements.size() == ShapeUtil::TupleElementCount(shape)); in WriteSingleTupleIndexTable()
Dconditional_simplifier.cc185 ShapeUtil::TupleElementCount(f->shape()); in TryRemoveConditional()
224 int64 old_tuple_element_count = ShapeUtil::TupleElementCount(param->shape()); in TryRemoveUnusedConditionalOperands()
Dallocation_tracker.cc148 i < ShapeUtil::TupleElementCount(shaped_buffer->on_device_shape()); in DeconstructTuple()
/external/tensorflow/tensorflow/compiler/xla/
Dlayout_util.cc372 if (ShapeUtil::TupleElementCount(src) != in CopyLayoutInternal()
373 ShapeUtil::TupleElementCount(*dst)) { in CopyLayoutInternal()
377 for (int64 i = 0; i < ShapeUtil::TupleElementCount(src); ++i) { in CopyLayoutInternal()
406 if (!rhs.IsTuple() || ShapeUtil::TupleElementCount(lhs) != in LayoutsInShapesEqual()
407 ShapeUtil::TupleElementCount(rhs)) { in LayoutsInShapesEqual()
410 for (int i = 0; i < ShapeUtil::TupleElementCount(lhs); ++i) { in LayoutsInShapesEqual()
Dliteral_comparison.cc205 const int64 tuple_elements = ShapeUtil::TupleElementCount(shape); in RecursiveElementCount()
685 for (int i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { in EqualHelper()
781 for (int64 i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { in NearHelper()
868 if (ShapeUtil::TupleElementCount(expected) != in EqualShapes()
869 ShapeUtil::TupleElementCount(actual)) { in EqualShapes()
872 ShapeUtil::TupleElementCount(expected), in EqualShapes()
873 ShapeUtil::TupleElementCount(actual)); in EqualShapes()
Dshape_util.cc492 return shape.IsTuple() && TupleElementCount(shape) == 0; in IsEmptyTuple()
495 /* static */ int64 ShapeUtil::TupleElementCount(const Shape& shape) { in TupleElementCount() function in xla::ShapeUtil
503 CHECK_GT(TupleElementCount(shape), index); in GetTupleElementShape()
519 CHECK_LE(start, TupleElementCount(tuple)); in SliceTuple()
520 CHECK_LE(limit, TupleElementCount(tuple)); in SliceTuple()
990 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ForEachSubshapeHelper()
1007 for (int64 i = 0; i < ShapeUtil::TupleElementCount(*shape); ++i) { in ForEachMutableSubshapeHelper()
Dliteral.cc146 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in SetPiece()
378 ShapeUtil::TupleElementCount(piece->subshape())) { in CreateFromProto()
381 ShapeUtil::TupleElementCount(piece->subshape()), in CreateFromProto()
408 for (int i = 0; i < ShapeUtil::TupleElementCount(shape()); ++i) { in DecomposeTuple()
1174 for (int i = 0; i < ShapeUtil::TupleElementCount(subshape); ++i) { in TupleToStringHelper()
1540 for (int i = 0; i < ShapeUtil::TupleElementCount(shape()); ++i) { in ConvertToShape()
2278 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CopyPieceSubtree()
2371 CHECK_EQ(src_buf_ptrs.size(), ShapeUtil::TupleElementCount(*shape_)); in MutableBorrowingLiteral()
2401 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in BuildPieceSubtree()
2430 CHECK_EQ(src_buf_ptrs.size(), ShapeUtil::TupleElementCount(*shape_)); in BorrowingLiteral()
Dshape_tree.h441 int64 count = ShapeUtil::TupleElementCount(shape); in CountSubshapes()
453 const int64 size = ShapeUtil::TupleElementCount(shape); in InitChildren()
489 const int64 size = ShapeUtil::TupleElementCount(shape); in InitChildren()
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Dtensor_list_utils.cc128 *is_nested_list = (xla::ShapeUtil::TupleElementCount(list_shape) > 2); in IsNestedTensorList()
167 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in GetTensorListPushIndex()
180 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in SetTensorListPushIndex()
229 int tuple_size = xla::ShapeUtil::TupleElementCount(element_tensor_list_shape); in GetTensorListShapeFromElementTensorListShape()
273 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in CreateZerosTensorListWithShape()
363 int list_tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in ExecuteTensorListPushBack()
370 int element_tuple_size = xla::ShapeUtil::TupleElementCount(element_shape); in ExecuteTensorListPushBack()
428 int list_tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in ExecuteTensorListPopBack()
/external/tensorflow/tensorflow/compiler/tf2xla/
Dshape_util.cc31 int64 tuple_elements = xla::ShapeUtil::TupleElementCount(shape); in PopulateInfeedLayoutVector()
178 int64 tuple_elements = xla::ShapeUtil::TupleElementCount(input_shape); in GetShapeWithLayout()
Dxla_jit_compiled_cpu_function_test.cc227 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 1); in TEST()
278 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 2); in TEST()
/external/tensorflow/tensorflow/compiler/xla/service/cpu/
Dcpu_transfer_manager.cc114 buffers.reserve(ShapeUtil::TupleElementCount(shape)); in TransferLiteralToInfeed()
121 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in TransferLiteralToInfeed()
/external/tensorflow/tensorflow/compiler/xla/tests/
Dclient_test.cc100 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.shape())); in XLA_TEST_F()
Dlocal_client_execute_test.cc212 EXPECT_EQ(3, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
240 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
318 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
/external/tensorflow/tensorflow/core/tpu/kernels/
Dtpu_execute_op.cc204 if (arg_list.size() != xla::ShapeUtil::TupleElementCount(input_host_shape)) { in BuildComputationInputs()
208 xla::ShapeUtil::TupleElementCount(input_host_shape)); in BuildComputationInputs()
405 xla::ShapeUtil::TupleElementCount(scoped_buffers.on_host_shape()); in AllocateOutputTensors()
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/
Dfused_ir_emitter.cc134 for (size_t i = 0; i < ShapeUtil::TupleElementCount(tuple->shape()); ++i) { in HandleTuple()
/external/tensorflow/tensorflow/compiler/xla/service/spmd/
Dspmd_partitioner_util.cc58 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CreateConstant()
74 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CreateZero()
95 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CreateOne()
129 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in EvenlyPartitions()
151 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in MakePartitionedShape()
171 for (int64 i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in MakeNonPaddedShapeForGivenPartition()
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dhorizontal_loop_fusion.cc51 return ShapeUtil::TupleElementCount(root->shape()); in GetOutputSizeOfFusion()
/external/tensorflow/tensorflow/compiler/xrt/
Dxrt_util.cc351 xla::ShapeUtil::TupleElementCount(output_tuple->on_device_shape()); in CreateExecuteOutput()

12