/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | fusion_merger.cc | 73 bytes += ShapeUtil::ByteSizeOf(user->shape()); in CalculateBytesReadByFusionParameter() 77 bytes += ShapeUtil::ByteSizeOf(instruction->shape()); in CalculateBytesReadByFusionParameter() 103 bytes += ShapeUtil::ByteSizeOf(operand->shape()); in CalculateFlopsToBytesRatio() 106 bytes += ShapeUtil::ByteSizeOf(fusion->fused_expression_root()->shape()); in CalculateFlopsToBytesRatio() 124 bytes_written += ShapeUtil::ByteSizeOf(operand->shape()); in GetCurrentBytesTransferred() 128 ShapeUtil::ByteSizeOf(fusion->fused_expression_root()->shape()); in GetCurrentBytesTransferred()
|
D | ir_emitter_nested.cc | 58 llvm_ir::ByteSizeOf(param_shape, module_->getDataLayout()); in EmitBasePointersForNestedComputation() 67 int64 root_size = llvm_ir::ByteSizeOf( in EmitBasePointersForNestedComputation()
|
D | cudnn_convolution_algorithm_picker.cc | 198 ShapeUtil::ByteSizeOf(input_shape)); in PickBestAlgorithm() 201 ShapeUtil::ByteSizeOf(filter_shape)); in PickBestAlgorithm() 204 ShapeUtil::ByteSizeOf(output_shape)); in PickBestAlgorithm()
|
D | gpu_compiler.h | 72 return ShapeUtil::ByteSizeOf(shape, pointer_size); in ShapeSizeBytesFunction()
|
D | hlo_schedule.cc | 203 return ShapeUtil::ByteSizeOf(buffer.shape(), pointer_size); in Build()
|
D | fusion_merger_test.cc | 128 ShapeUtil::ByteSizeOf(data_shape_) * 3; in BuildComputation1()
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | hlo_rematerialization_test.cc | 134 static int64 ByteSizeOf(const Shape& shape) { in ByteSizeOf() function in xla::__anonc0d2dd690111::HloRematerializationTest 135 return ShapeUtil::ByteSizeOf(shape, sizeof(void*)); in ByteSizeOf() 163 ByteSizeOf, in TEST_F() 196 ByteSizeOf, in TEST_F() 237 ByteSizeOf, in TEST_F() 273 ByteSizeOf, in TEST_F() 315 ByteSizeOf, in TEST_F() 386 ByteSizeOf, in TEST_F() 387 /*memory_limit_bytes=*/4 * ByteSizeOf(vec1024_shape_), in TEST_F() 481 ByteSizeOf, in TEST_F() [all …]
|
D | hlo_scheduling_test.cc | 80 return ShapeUtil::ByteSizeOf(buffer.shape(), /*pointer_size=*/8); in TEST_F()
|
D | hlo_execution_profile_test.cc | 71 return ShapeUtil::ByteSizeOf(shape, pointer_size); in TEST_F()
|
D | generic_transfer_manager.cc | 174 return ShapeUtil::ByteSizeOf(shape, pointer_size_); in GetByteSizeRequirement()
|
D | buffer_assignment_test.cc | 1511 CreateMemoryMinimizingSequence(*module, ByteSizeOf).ConsumeValueOrDie(); in RunBufferAssignment() 1514 ByteSizeOf, in RunBufferAssignment() 1519 static int64 ByteSizeOf(const LogicalBuffer& buffer) { in ByteSizeOf() function in xla::__anon203ae3150111::WhileBufferAssignmentTest 1520 return ShapeUtil::ByteSizeOf(buffer.shape(), sizeof(void*)); in ByteSizeOf() 1866 CreateMemoryMinimizingSequence(*module, ByteSizeOf).ConsumeValueOrDie(); in TEST_F() 1884 ByteSizeOf, [](LogicalBuffer::Color) { return 1; }) in TEST_F()
|
D | hlo_ordering_test.cc | 70 return ShapeUtil::ByteSizeOf(buffer.shape()); in TEST_F()
|
/external/tensorflow/tensorflow/compiler/xla/ |
D | shape_util_test.cc | 276 EXPECT_EQ(4, ShapeUtil::ByteSizeOf(ShapeUtil::MakeShape(F32, {}))); in TEST() 277 EXPECT_EQ(800, ShapeUtil::ByteSizeOf(ShapeUtil::MakeShape(F32, {10, 20}))); in TEST() 280 EXPECT_EQ(8, ShapeUtil::ByteSizeOf(ShapeUtil::MakeShape(F64, {}))); in TEST() 281 EXPECT_EQ(1600, ShapeUtil::ByteSizeOf(ShapeUtil::MakeShape(F64, {10, 20}))); in TEST() 284 EXPECT_EQ(8, ShapeUtil::ByteSizeOf(ShapeUtil::MakeShape(C64, {}))); in TEST() 285 EXPECT_EQ(1600, ShapeUtil::ByteSizeOf(ShapeUtil::MakeShape(C64, {10, 20}))); in TEST() 291 EXPECT_EQ(800, ShapeUtil::ByteSizeOf(shape)); in TEST() 295 EXPECT_EQ(15 * 21 * 4, ShapeUtil::ByteSizeOf(shape)); in TEST()
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | external_constant_pool.cc | 35 int64 literal_size = ShapeUtil::ByteSizeOf(literal.shape()); in Insert()
|
D | parallel_cpu_executable.h | 83 return ShapeUtil::ByteSizeOf(shape, sizeof(void*)); in ShapeSizeBytes()
|
D | ir_emitter.cc | 176 ByteSizeOf(literal.shape()) >= kMaxInternalConstantSizeInBytes) { in HandleConstant() 255 int64 IrEmitter::ByteSizeOf(const Shape& shape) const { in ByteSizeOf() function in xla::cpu::IrEmitter 256 return llvm_ir::ByteSizeOf(shape, module_->getDataLayout()); in ByteSizeOf() 265 int64 buffer_size = ByteSizeOf(shape); in MinimumAlignmentForShape() 290 AttachDereferenceableMetadataForLoad(load, ByteSizeOf(shape)); in AttachDereferenceableMetadataForLoad() 380 int64 length = ByteSizeOf(shape); in EmitXfeedTransfer() 1948 ShapeUtil::ByteSizeOf(logical_element_shape) * memcpy_elements; in HandleSlice() 2829 attr_builder.addDereferenceableAttr(ByteSizeOf(target_shape)); in EmitTargetAddressForOp() 2906 int64 source_size = ByteSizeOf(source.shape()); in EmitMemcpy()
|
D | cpu_executable.cc | 360 return ShapeUtil::ByteSizeOf(shape, sizeof(void*)); in ShapeSizeBytes()
|
D | ir_emitter.h | 521 int64 ByteSizeOf(const Shape& shape) const;
|
/external/tensorflow/tensorflow/compiler/xla/client/lib/ |
D | testing.cc | 39 return ShapeUtil::ByteSizeOf(shape); in DataSizeOfShape()
|
/external/tensorflow/tensorflow/compiler/xla/service/interpreter/ |
D | executable.cc | 123 return ShapeUtil::ByteSizeOf(shape, sizeof(void*)); in ShapeSizeBytes()
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
D | tuple_ops.cc | 99 ByteSizeOf(target_shape, src_buffer->getModule()->getDataLayout())); in EmitGetTupleElement()
|
D | llvm_util.h | 257 int64 ByteSizeOf(const Shape& shape, const llvm::DataLayout& data_layout);
|
D | llvm_util.cc | 606 int64 ByteSizeOf(const Shape& shape, const llvm::DataLayout& data_layout) { in ByteSizeOf() function 608 return ShapeUtil::ByteSizeOf(shape, pointer_size); in ByteSizeOf()
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
D | xla_jit_compiled_cpu_function.cc | 47 arg_sizes.push_back(xla::ShapeUtil::ByteSizeOf(arg_shape, kPointerSize)); in ComputeArgSizes()
|
/external/tensorflow/tensorflow/compiler/aot/ |
D | codegen.cc | 105 arg_sizes->push_back(xla::ShapeUtil::ByteSizeOf( in ComputeArgSizes()
|