Home
last modified time | relevance | path

Searched refs:TupleElementCount (Results 1 – 25 of 47) sorted by relevance

12

/external/tensorflow/tensorflow/compiler/xla/service/
Dhlo_module_dce_test.cc198 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
308 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
309 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[1]->shape())); in TEST_F()
375 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
376 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[1]->shape())); in TEST_F()
512 EXPECT_EQ(1, ShapeUtil::TupleElementCount(while_loops[0]->shape())); in TEST_F()
Dtransfer_manager.cc295 ShapeUtil::TupleElementCount(device_subshape) > 0) { in WriteTupleIndexTablesAsync()
302 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(device_subshape); in WriteTupleIndexTablesAsync()
319 if (ShapeUtil::TupleElementCount(device_buffer.on_device_shape()) == 0) { in WriteRootTupleIndexTable()
328 i < ShapeUtil::TupleElementCount(device_buffer.on_device_shape()); ++i) { in WriteRootTupleIndexTable()
338 if (ShapeUtil::TupleElementCount(buffer_tree.shape()) == 0) { in WriteRootTupleIndexTable()
347 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(buffer_tree.shape()); in WriteRootTupleIndexTable()
Dconditional_simplifier_test.cc321 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 0); in TEST_F()
367 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
413 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
476 EXPECT_EQ(ShapeUtil::TupleElementCount(conditional->shape()), 1); in TEST_F()
Dhlo_element_type_converter.cc69 const int64_t n = ShapeUtil::TupleElementCount(shape); in GetConvertedTupleShape()
92 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ConvertTupleElements()
Dhlo_module_dce.cc66 ShapeUtil::TupleElementCount(xla_while->shape()); in RunWhileDCE()
Dgather_expander_test.cc102 ASSERT_EQ(ShapeUtil::TupleElementCount(while_shape), 4); in TEST_F()
Dgeneric_transfer_manager.cc45 TF_RET_CHECK(elements.size() == ShapeUtil::TupleElementCount(shape)); in WriteSingleTupleIndexTable()
Dconditional_simplifier.cc90 ShapeUtil::TupleElementCount(param->shape()); in TryRemoveUnusedConditionalOperands()
560 ShapeUtil::TupleElementCount(f->shape()); in TryRemoveConditional()
Dallocation_tracker.cc148 const auto n = ShapeUtil::TupleElementCount(shaped_buffer->on_device_shape()); in DeconstructTuple()
Dhlo_verifier.cc543 return ShapeUtil::TupleElementCount(shape) != in CheckBufferOffset()
586 if (ShapeUtil::TupleElementCount(input_buffer_shape) != in CheckInplaceCollectivePermute()
587 ShapeUtil::TupleElementCount(output_buffer_shape)) { in CheckInplaceCollectivePermute()
591 ShapeUtil::TupleElementCount(input_offset_shape) != in CheckInplaceCollectivePermute()
592 ShapeUtil::TupleElementCount(input_buffer_shape)) { in CheckInplaceCollectivePermute()
604 ShapeUtil::TupleElementCount(output_offset_shape) != in CheckInplaceCollectivePermute()
605 ShapeUtil::TupleElementCount(output_buffer_shape)) { in CheckInplaceCollectivePermute()
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Dtensor_list_utils.cc128 *is_nested_list = (xla::ShapeUtil::TupleElementCount(list_shape) > 2); in IsNestedTensorList()
167 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in GetTensorListPushIndex()
180 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in SetTensorListPushIndex()
230 int tuple_size = xla::ShapeUtil::TupleElementCount(element_tensor_list_shape); in GetTensorListShapeFromElementTensorListShape()
275 int tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in CreateZerosTensorListWithShape()
365 int list_tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in ExecuteTensorListPushBack()
372 int element_tuple_size = xla::ShapeUtil::TupleElementCount(element_shape); in ExecuteTensorListPushBack()
430 int list_tuple_size = xla::ShapeUtil::TupleElementCount(list_shape); in ExecuteTensorListPopBack()
/external/tensorflow/tensorflow/compiler/xla/
Dlayout_util.cc434 if (ShapeUtil::TupleElementCount(src) != in CopyLayoutInternal()
435 ShapeUtil::TupleElementCount(*dst)) { in CopyLayoutInternal()
439 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(src); ++i) { in CopyLayoutInternal()
468 if (!rhs.IsTuple() || ShapeUtil::TupleElementCount(lhs) != in LayoutsInShapesEqual()
469 ShapeUtil::TupleElementCount(rhs)) { in LayoutsInShapesEqual()
472 for (int i = 0; i < ShapeUtil::TupleElementCount(lhs); ++i) { in LayoutsInShapesEqual()
Dliteral_comparison.cc208 const int64_t tuple_elements = ShapeUtil::TupleElementCount(shape); in RecursiveElementCount()
692 for (int i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); ++i) { in EqualHelper()
788 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(expected.shape()); in NearHelper()
876 if (ShapeUtil::TupleElementCount(expected) != in EqualShapes()
877 ShapeUtil::TupleElementCount(actual)) { in EqualShapes()
880 ShapeUtil::TupleElementCount(expected), in EqualShapes()
881 ShapeUtil::TupleElementCount(actual)); in EqualShapes()
Dshape_util.cc554 return shape.IsTuple() && TupleElementCount(shape) == 0; in IsEmptyTuple()
557 /* static */ int64_t ShapeUtil::TupleElementCount(const Shape& shape) { in TupleElementCount() function in xla::ShapeUtil
565 CHECK_GT(TupleElementCount(shape), index); in GetTupleElementShape()
581 CHECK_LE(start, TupleElementCount(tuple)); in SliceTuple()
582 CHECK_LE(limit, TupleElementCount(tuple)); in SliceTuple()
1069 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in ForEachSubshapeHelper()
1086 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(*shape); ++i) { in ForEachMutableSubshapeHelper()
Dliteral.cc232 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in mutable_shape_do_not_use()
484 ShapeUtil::TupleElementCount(piece->subshape())) { in CreateFromProto()
487 ShapeUtil::TupleElementCount(piece->subshape()), in CreateFromProto()
524 const auto tuple_element_count = ShapeUtil::TupleElementCount(shape()); in DecomposeTuple()
1307 const auto tuple_element_count = ShapeUtil::TupleElementCount(subshape); in TupleToStringHelper()
1309 for (int i = 0; i < ShapeUtil::TupleElementCount(subshape); ++i) { in TupleToStringHelper()
1736 const auto tuple_element_count = ShapeUtil::TupleElementCount(shape()); in ConvertToShape()
2565 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CopyPieceSubtree()
2657 CHECK_EQ(src_buf_ptrs.size(), ShapeUtil::TupleElementCount(*shape_)); in MutableBorrowingLiteral()
2687 for (int i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in BuildPieceSubtree()
[all …]
/external/tensorflow/tensorflow/core/tpu/kernels/
Dtpu_reshard_variables_op_util.cc106 if (var_list.size() != xla::ShapeUtil::TupleElementCount(input_host_shape)) { in BuildInputBuffers()
110 xla::ShapeUtil::TupleElementCount(input_host_shape)); in BuildInputBuffers()
233 xla::ShapeUtil::TupleElementCount(result_buffers.on_host_shape()); in UpdateOutputVariables()
Dtpu_execute_op.cc203 if (arg_list.size() != xla::ShapeUtil::TupleElementCount(input_host_shape)) { in BuildComputationInputs()
207 xla::ShapeUtil::TupleElementCount(input_host_shape)); in BuildComputationInputs()
404 xla::ShapeUtil::TupleElementCount(scoped_buffers.on_host_shape()); in AllocateOutputTensors()
/external/tensorflow/tensorflow/compiler/tf2xla/
Dshape_util.cc31 int64_t tuple_elements = xla::ShapeUtil::TupleElementCount(shape); in PopulateInfeedLayoutVector()
177 int64_t tuple_elements = xla::ShapeUtil::TupleElementCount(input_shape); in GetShapeWithLayout()
Dxla_jit_compiled_cpu_function_test.cc227 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 1); in TEST()
278 ASSERT_EQ(ShapeUtil::TupleElementCount(result), 2); in TEST()
/external/tensorflow/tensorflow/compiler/xla/service/cpu/
Dcpu_xfeed.cc197 buffers.reserve(ShapeUtil::TupleElementCount(shape)); in TransferLiteralToInfeedOnCpu()
204 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in TransferLiteralToInfeedOnCpu()
/external/tensorflow/tensorflow/compiler/xla/tests/
Dclient_test.cc99 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.shape())); in XLA_TEST_F()
Dlocal_client_execute_test.cc212 EXPECT_EQ(3, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
240 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
318 EXPECT_EQ(2, ShapeUtil::TupleElementCount(result.on_host_shape())); in XLA_TEST_F()
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dgpu_fusible.cc622 return ShapeUtil::TupleElementCount(root->shape()); in GetOutputSizeOfFusible()
/external/tensorflow/tensorflow/compiler/xla/service/spmd/
Dspmd_partitioner_util.h60 for (int64_t i = 0; i < ShapeUtil::TupleElementCount(shape); ++i) { in CreateConstantBase()
/external/tensorflow/tensorflow/compiler/xrt/
Dxrt_util.cc354 xla::ShapeUtil::TupleElementCount(output_tuple->on_device_shape()); in CreateExecuteOutput()

12