/third_party/mindspore/mindspore/lite/test/ |
D | runtest.sh | 14 cp ${BUILD_DIR}/test/lite-test ./ 28 ./lite-test --gtest_filter="*MindDataTestTensorDE*" 29 ./lite-test --gtest_filter="*MindDataTestEager*" 37 ./lite-test --gtest_filter=TestFcFp32* 38 ./lite-test --gtest_filter=TestConv1x1Fp32* 39 ./lite-test --gtest_filter=TestDeConvolutionFp32* 43 ./lite-test --gtest_filter=TestDeconvInt8.* 46 ./lite-test --gtest_filter="TestConvolutionOpenCL.simple_test*" 47 ./lite-test --gtest_filter="TestArithmeticSelfOpenCLCI.ArithmeticSelfRound*" 48 ./lite-test --gtest_filter="TestConcatOpenCLCI.ConcatFp32_2inputforCI*" [all …]
|
/third_party/mindspore/mindspore/lite/test/ut/src/runtime/kernel/arm/int8/ |
D | slice_int8_tests.cc | 32 lite::Tensor in_tensor(kNumberTypeInt8, {1, 3, 2, 3}); in TEST_F() 33 lite::Tensor out_tensor(kNumberTypeInt8, {1, 2, 2, 3}); in TEST_F() 39 lite::Tensor begin_tensor(kNumberTypeInt32, {4}); in TEST_F() 42 lite::Tensor size_tensor(kNumberTypeInt32, {4}); in TEST_F() 46 const lite::LiteQuantParam quant_in0 = {0.00784314f, 0}; // -1.0--1.0 -> 0--255 in TEST_F() 47 const lite::LiteQuantParam quant_out = {0.00784314f, 0}; in TEST_F() 51 std::vector<lite::Tensor *> inputs = {&in_tensor, &begin_tensor, &size_tensor}; in TEST_F() 52 std::vector<lite::Tensor *> outputs = {&out_tensor}; in TEST_F() 58 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() 61 auto ctx = std::make_shared<lite::InnerContext>(); in TEST_F() [all …]
|
D | arithmetic_self_int8_tests.cc | 41 lite::LiteQuantParam input_quant_arg; in TEST_F() 44 lite::LiteQuantParam output_quant_arg; in TEST_F() 49 lite::Tensor *input_tensor1 = new lite::Tensor; in TEST_F() 54 std::vector<lite::Tensor *> inputs_tensor(1); in TEST_F() 57 lite::Tensor *output0_tensor = new lite::Tensor; in TEST_F() 62 std::vector<lite::Tensor *> outputs_tensor(1); in TEST_F() 67 lite::InnerContext *ctx = new lite::InnerContext; in TEST_F() 69 ASSERT_EQ(lite::RET_OK, ctx->Init()); in TEST_F() 71 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() 104 lite::LiteQuantParam input_quant_arg; in TEST_F() [all …]
|
D | crop_int8_tests.cc | 42 lite::LiteQuantParam input_quant_arg; in TEST_F() 45 lite::LiteQuantParam output_quant_arg; in TEST_F() 49 lite::Tensor *input_tensor1 = new lite::Tensor; in TEST_F() 56 std::vector<lite::Tensor *> inputs_tensor(1); in TEST_F() 59 std::vector<lite::Tensor *> outputs_tensor(1); in TEST_F() 60 lite::Tensor *output0_tensor = new lite::Tensor; in TEST_F() 69 lite::InnerContext *ctx = new lite::InnerContext; in TEST_F() 71 ASSERT_EQ(lite::RET_OK, ctx->Init()); in TEST_F() 76 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() 109 lite::LiteQuantParam input_quant_arg; in TEST_F() [all …]
|
D | mul_int8_tests.cc | 45 lite::LiteQuantParam input_quant_arg; in TEST_F() 48 lite::LiteQuantParam output_quant_arg; in TEST_F() 52 lite::Tensor *input_tensor1 = new lite::Tensor; in TEST_F() 59 lite::Tensor *input_tensor2 = new lite::Tensor; in TEST_F() 65 std::vector<lite::Tensor *> inputs_tensor(2); in TEST_F() 69 std::vector<lite::Tensor *> outputs_tensor(1); in TEST_F() 70 lite::Tensor *output0_tensor = new lite::Tensor; in TEST_F() 79 lite::InnerContext *ctx = new lite::InnerContext; in TEST_F() 81 ASSERT_EQ(lite::RET_OK, ctx->Init()); in TEST_F() 83 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() [all …]
|
D | concat_int8_tests.cc | 45 lite::LiteQuantParam input_quant_arg; in TEST_F() 48 lite::LiteQuantParam output_quant_arg; in TEST_F() 52 lite::Tensor *input_tensor1 = new lite::Tensor; in TEST_F() 59 lite::Tensor *input_tensor2 = new lite::Tensor; in TEST_F() 65 std::vector<lite::Tensor *> inputs_tensor(2); in TEST_F() 69 std::vector<lite::Tensor *> outputs_tensor(1); in TEST_F() 70 lite::Tensor *output0_tensor = new lite::Tensor; in TEST_F() 80 lite::InnerContext *ctx = new lite::InnerContext; in TEST_F() 82 ASSERT_EQ(lite::RET_OK, ctx->Init()); in TEST_F() 84 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() [all …]
|
/third_party/mindspore/mindspore/lite/tools/optimizer/format/ |
D | delete_redundant_transpose.cc | 30 MS_CHECK_TRUE_RET(node != nullptr, lite::RET_NULL_PTR); in DeleteNot4DTranspose() 38 lite::ReturnCode::GetSingleReturnCode()->UpdateReturnCode(lite::RET_NULL_PTR); in DeleteNot4DTranspose() 39 return lite::RET_NULL_PTR; in DeleteNot4DTranspose() 41 if (DeleteNot4DTranspose(sub_func_graph) != lite::RET_OK) { in DeleteNot4DTranspose() 43 return lite::RET_ERROR; in DeleteNot4DTranspose() 47 lite::ReturnCode::GetSingleReturnCode()->UpdateReturnCode(lite::RET_NULL_PTR); in DeleteNot4DTranspose() 48 return lite::RET_NULL_PTR; in DeleteNot4DTranspose() 50 if (DeleteNot4DTranspose(sub_func_graph) != lite::RET_OK) { in DeleteNot4DTranspose() 52 return lite::RET_ERROR; in DeleteNot4DTranspose() 61 if (FetchShapeFromAbstract(abstract, &shape) != lite::RET_OK) { in DeleteNot4DTranspose() [all …]
|
D | to_format_base.cc | 26 using mindspore::lite::NHWC_SHAPE; 39 if (status != lite::RET_OK && status != lite::RET_INFER_INVALID) { in GenNewInput() 41 return lite::RET_ERROR; in GenNewInput() 62 return lite::RET_ERROR; in GenNewInput() 65 return lite::RET_OK; in GenNewInput() 73 return lite::RET_ERROR; in ModifyCNode() 90 if (FetchShapeFromAbstract(abstract, &shape) != lite::RET_OK) { in ModifyCNode() 92 return lite::RET_ERROR; in ModifyCNode() 106 return lite::RET_OK; in ModifyCNode() 117 return lite::RET_ERROR; in InsertPreTransNode() [all …]
|
/third_party/mindspore/mindspore/lite/test/ut/src/runtime/ |
D | runtime_pass_tests.cc | 23 namespace lite { namespace 33 void Nc4hw4PassConstruct(std::vector<kernel::LiteKernel *> *kernels, std::vector<lite::Tensor *> *t… in Nc4hw4PassConstruct() 34 lite::InnerContext *ctx) { in Nc4hw4PassConstruct() 35 auto reg = lite::KernelRegistry::GetInstance(); in Nc4hw4PassConstruct() 37 lite::Tensor *conv_in_tensor = new lite::Tensor(kNumberTypeFloat32, {1, 1, 1, 1}, NHWC); in Nc4hw4PassConstruct() 39 lite::Tensor *conv_weight = new lite::Tensor(kNumberTypeFloat32, {1, 1, 1, 1}, NHWC); in Nc4hw4PassConstruct() 41 lite::Tensor *conv_out_tensor = new lite::Tensor(kNumberTypeFloat32, {1, 1, 1, 1}, NHWC); in Nc4hw4PassConstruct() 43 std::vector<lite::Tensor *> conv_in = {conv_in_tensor, conv_weight}; in Nc4hw4PassConstruct() 44 std::vector<lite::Tensor *> conv_out = {conv_out_tensor}; in Nc4hw4PassConstruct() 54 lite::Tensor *trans_param_tensor = new lite::Tensor(kNumberTypeFloat32, {1, 1, 1, 1}, NHWC); in Nc4hw4PassConstruct() [all …]
|
/third_party/mindspore/mindspore/lite/tools/converter/acl/ |
D | acl_pass.cc | 78 if (ConvertGraphToOm(func_graph, &om_data) != lite::RET_OK) { in BuildGraph() 80 return lite::RET_ERROR; in BuildGraph() 85 return lite::RET_ERROR; in BuildGraph() 87 return lite::RET_OK; in BuildGraph() 94 lite::GetAllFuncGraph(func_graph, &all_func_graphs); in RunPrimitiveMapper() 105 return lite::RET_ERROR; in RunPrimitiveMapper() 108 auto mapper = lite::PrimitiveMapperRegister::GetInstance().GetPrimitiveMapper(name); in RunPrimitiveMapper() 115 if (status != lite::RET_OK) { in RunPrimitiveMapper() 117 return lite::RET_ERROR; in RunPrimitiveMapper() 121 return lite::RET_OK; in RunPrimitiveMapper() [all …]
|
/third_party/mindspore/mindspore/lite/test/ut/src/runtime/kernel/arm/fp32/ |
D | sparse_to_dense_fp32_tests.cc | 44 lite::Tensor *input_tensor1 = new lite::Tensor; in TEST_F() 49 lite::Tensor *input_tensor2 = new lite::Tensor; in TEST_F() 54 lite::Tensor *input_tensor3 = new lite::Tensor; in TEST_F() 59 lite::Tensor *input_tensor4 = new lite::Tensor; in TEST_F() 64 std::vector<lite::Tensor *> inputs_tensor(4); in TEST_F() 74 lite::Tensor *output0_tensor = new lite::Tensor; in TEST_F() 78 std::vector<lite::Tensor *> outputs_tensor(1); in TEST_F() 83 lite::InnerContext *ctx = new lite::InnerContext; in TEST_F() 85 ASSERT_EQ(lite::RET_OK, ctx->Init()); in TEST_F() 88 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() [all …]
|
/third_party/mindspore/mindspore/lite/tools/optimizer/graph/ |
D | node_infershape.cc | 35 void FreeTensors(std::vector<lite::Tensor *> *tensors) { in FreeTensors() 46 void RectifyFormat(const std::vector<lite::Tensor *> &inputs, FmkType fmk_type) { in RectifyFormat() 59 tensor::TensorPtr NewTensorInfo(lite::Tensor *tensor) { in NewTensorInfo() 76 auto prim_t = lite::GetPrimitiveT(cnode->input(0)); in JudgeOpSupportInfer() 81 …lite::PopulateRegistry::GetInstance()->GetParameterCreator(static_cast<int>(prim_t->value.type), l… in JudgeOpSupportInfer() 94 return lite::RET_ERROR; in InferShape() 97 std::vector<lite::Tensor *> inputs; in InferShape() 98 std::vector<lite::Tensor *> outputs; in InferShape() 99 if (GetCNodeInputTensors(cnode, &inputs) != lite::RET_OK) { in InferShape() 102 return lite::RET_ERROR; in InferShape() [all …]
|
D | decrease_transpose_algo.cc | 60 lite::RemoveIfDepend(cur_node); in FindAreaSurroundedByTranspose() 84 return lite::RET_ERROR; in FindAreaSurroundedByTranspose() 99 return lite::RET_OK; in FindAreaSurroundedByTranspose() 107 …if (!CheckPrimitiveType(cnode, prim::kPrimTranspose) || GetTransposePerm(cnode, &perm) != lite::RE… in SetTransType() 153 return lite::RET_OK; in ConvertTensorToNCOrNH() 155 lite::DataInfo data_info; in ConvertTensorToNCOrNH() 159 MS_CHECK_TRUE_MSG(input_node != nullptr, lite::RET_ERROR, "input_node is nullptr"); in ConvertTensorToNCOrNH() 161 return lite::RET_OK; in ConvertTensorToNCOrNH() 163 status = lite::FetchDataFromParameterNode(cnode, index, fmk_type, train_flag, &data_info); in ConvertTensorToNCOrNH() 165 status = lite::FetchDataFromValueNode(cnode, index, fmk_type, train_flag, &data_info); in ConvertTensorToNCOrNH() [all …]
|
D | infershape_pass.cc | 28 lite::RemoveIfDepend(cnode); in GetCNodeCertainInputFormat() 29 lite::RemoveIfMakeTuple(cnode); in GetCNodeCertainInputFormat() 34 return lite::RET_ERROR; in GetCNodeCertainInputFormat() 38 return lite::RET_NO_CHANGE; in GetCNodeCertainInputFormat() 48 MS_CHECK_TRUE_MSG(primitive != nullptr, lite::RET_NULL_PTR, "GetValueNode Failed"); in GetCNodeCertainInputFormat() 51 return lite::RET_ERROR; in GetCNodeCertainInputFormat() 54 MS_CHECK_TRUE_MSG(format_attr != nullptr, lite::RET_NULL_PTR, "GetAttr Failed"); in GetCNodeCertainInputFormat() 58 if (GetTransposePerm(real_cnode, &perm) != lite::RET_OK) { in GetCNodeCertainInputFormat() 60 return lite::RET_ERROR; in GetCNodeCertainInputFormat() 71 return lite::RET_OK; in GetCNodeCertainInputFormat() [all …]
|
D | redundant_op_remove_pass.cc | 64 return lite::RET_NO_CHANGE; in ProcessInputIsMonad() 73 return lite::RET_OK; in ProcessInputIsMonad() 101 return lite::RET_NO_CHANGE; in ProcessDependencyWithTwoNodes() 108 MS_CHECK_TRUE_MSG(depend_prim != nullptr, lite::RET_NULL_PTR, "NewCNode Failed"); in ProcessDependencyWithTwoNodes() 109 MS_CHECK_TRUE_MSG(depend_node != nullptr, lite::RET_NULL_PTR, "NewCNode Failed"); in ProcessDependencyWithTwoNodes() 112 return lite::RET_OK; in ProcessDependencyWithTwoNodes() 117 if (ProcessDependencyWithTwoNodes(func_graph, cnode, true) == lite::RET_OK) { in ProcessInputHaveDependency() 118 return lite::RET_OK; in ProcessInputHaveDependency() 120 if (ProcessDependencyWithTwoNodes(func_graph, cnode, false) == lite::RET_OK) { in ProcessInputHaveDependency() 121 return lite::RET_OK; in ProcessInputHaveDependency() [all …]
|
/third_party/mindspore/mindspore/lite/tools/optimizer/fusion/ |
D | conv_transform_fusion.cc | 128 if (GenTransParam(transform_node, kernel_nums, trans_scale, trans_bias) != lite::RET_OK) { in Process() 134 …if (GenNewConvTensor(func_graph, conv_node, kernel_nums, trans_scale, trans_bias) != lite::RET_OK)… in Process() 151 lite::ReturnCode::GetSingleReturnCode()->UpdateReturnCode(lite::RET_NULL_PTR); in GenTransParam() 152 return lite::RET_NULL_PTR; in GenTransParam() 156 lite::ReturnCode::GetSingleReturnCode()->UpdateReturnCode(lite::RET_NULL_PTR); in GenTransParam() 157 return lite::RET_NULL_PTR; in GenTransParam() 161 lite::ReturnCode::GetSingleReturnCode()->UpdateReturnCode(lite::RET_MEMORY_FAILED); in GenTransParam() 162 return lite::RET_ERROR; in GenTransParam() 166 lite::ReturnCode::GetSingleReturnCode()->UpdateReturnCode(lite::RET_MEMORY_FAILED); in GenTransParam() 167 return lite::RET_ERROR; in GenTransParam() [all …]
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/base/ |
D | group_convolution_creator.cc | 20 void CopyTensorQuantParam(lite::Tensor *dst, lite::Tensor *src) { in CopyTensorQuantParam() 36 void FreeCurrentConv(ConvParameter *conv_param, std::vector<lite::Tensor *> *new_inputs, in FreeCurrentConv() 37 std::vector<lite::Tensor *> *new_outputs) { in FreeCurrentConv() 55 static inline lite::Tensor *TensorMalloc(lite::Tensor *tensor) { in TensorMalloc() 56 if (tensor->MallocData() != lite::RET_OK) { in TensorMalloc() 64 lite::Tensor *CreateConstTensor(lite::Tensor *tensor, const std::vector<int> &shape, const int inde… in CreateConstTensor() 66 …new (std::nothrow) lite::Tensor(tensor->data_type(), shape, mindspore::NHWC, lite::Tensor::Categor… in CreateConstTensor() 72 if (ret != lite::RET_OK) { in CreateConstTensor() 83 lite::Tensor *CreateVarTensor(const TensorInfo &tensor_info, bool inferred) { in CreateVarTensor() 84 auto tensor = new (std::nothrow) lite::Tensor(); in CreateVarTensor() [all …]
|
/third_party/mindspore/mindspore/lite/test/ut/src/runtime/kernel/arm/common/ |
D | strided_slice_tests.cc | 29 void InitStridedSliceParam(StridedSliceParameter *param, const lite::Tensor *in_tensor, in InitStridedSliceParam() 30 const lite::Tensor *begin_tensor, const lite::Tensor *end_tensor, in InitStridedSliceParam() 31 const lite::Tensor *stride_tensor) { in InitStridedSliceParam() 54 lite::Tensor in_tensor(kNumberTypeFloat32, {1, 2, 4}); in TEST_F() 55 lite::Tensor out_tensor(kNumberTypeFloat32, {1, 1, 2}); in TEST_F() 60 lite::Tensor begins_tensor(kNumberTypeInt32, {3}); in TEST_F() 63 lite::Tensor ends_tensor(kNumberTypeInt32, {3}); in TEST_F() 66 lite::Tensor strides_tensor(kNumberTypeInt32, {3}); in TEST_F() 69 std::vector<lite::Tensor *> inputs = {&in_tensor, &begins_tensor, &ends_tensor, &strides_tensor}; in TEST_F() 70 std::vector<lite::Tensor *> outputs = {&out_tensor}; in TEST_F() [all …]
|
/third_party/mindspore/mindspore/lite/src/train/ |
D | train_populate_parameter.cc | 34 using mindspore::lite::Registry; 516 …lite::Registry ApplyMomentumParameterRegistry(schema::PrimitiveType_ApplyMomentum, PopulateApplyMo… in PopulateTrainParameters() 517 lite::SCHEMA_CUR); in PopulateTrainParameters() 518 …lite::Registry BiasGradParameterRegistry(schema::PrimitiveType_BiasAddGrad, PopulateBiasGradParame… in PopulateTrainParameters() 519 lite::SCHEMA_CUR); in PopulateTrainParameters() 520 …lite::Registry SoftmaxCrossEntropyParameterRegistry(schema::PrimitiveType_SoftmaxCrossEntropyWithL… in PopulateTrainParameters() 521 … PopulateSoftmaxCrossEntropyParameter, lite::SCHEMA_CUR); in PopulateTrainParameters() 522 …lite::Registry SparseSoftmaxCrossEntropyParameterRegistry(schema::PrimitiveType_SparseSoftmaxCross… in PopulateTrainParameters() 524 lite::SCHEMA_CUR); in PopulateTrainParameters() 525 …lite::Registry ActivationParameterRegistry(schema::PrimitiveType_ActivationGrad, PopulateActivatio… in PopulateTrainParameters() [all …]
|
/third_party/mindspore/mindspore/lite/src/runtime/gpu/opencl/ |
D | opencl_runtime_wrapper.cc | 35 lite::opencl::OpenCLRuntimeInnerWrapper ocl_runtime_wrap; in LoadSource() 36 lite::opencl::OpenCLRuntime *ocl_runtime = ocl_runtime_wrap.GetInstance(); in LoadSource() 48 lite::opencl::OpenCLRuntimeInnerWrapper ocl_runtime_wrap; in BuildKernel() 49 lite::opencl::OpenCLRuntime *ocl_runtime = ocl_runtime_wrap.GetInstance(); in BuildKernel() 59 lite::opencl::OpenCLRuntimeInnerWrapper ocl_runtime_wrap; in SetKernelArg() 60 lite::opencl::OpenCLRuntime *ocl_runtime = ocl_runtime_wrap.GetInstance(); in SetKernelArg() 70 lite::opencl::OpenCLRuntimeInnerWrapper ocl_runtime_wrap; in RunKernel() 71 lite::opencl::OpenCLRuntime *ocl_runtime = ocl_runtime_wrap.GetInstance(); in RunKernel() 80 lite::opencl::OpenCLRuntimeInnerWrapper ocl_runtime_wrap; in SyncCommandQueue() 81 lite::opencl::OpenCLRuntime *ocl_runtime = ocl_runtime_wrap.GetInstance(); in SyncCommandQueue() [all …]
|
/third_party/mindspore/mindspore/lite/test/ut/src/runtime/kernel/arm/fp32_grad/ |
D | arithmetic_grad_fp32_tests.cc | 32 std::vector<lite::Tensor *> inputs, in PopulateArithmeticParameter() 33 std::vector<lite::Tensor *> outputs) { in PopulateArithmeticParameter() 56 std::vector<lite::Tensor *> GenerateTensorsForTest(const char *test, int test_id) { in GenerateTensorsForTest() 58 std::vector<lite::Tensor *> ret_vector; in GenerateTensorsForTest() 83 auto dy_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(test, &input_size)); in GenerateTensorsForTest() 88 lite::Tensor *dy_tensor = new lite::Tensor(TypeId::kNumberTypeFloat32, large_dim); in GenerateTensorsForTest() 96 auto x1_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(dx1_file, &input_size)); in GenerateTensorsForTest() 103 lite::Tensor *x1_tensor = new lite::Tensor(TypeId::kNumberTypeFloat32, large_dim); in GenerateTensorsForTest() 113 auto x2_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(dx2_file, &input_size)); in GenerateTensorsForTest() 122 lite::Tensor *x2_tensor = new lite::Tensor(TypeId::kNumberTypeFloat32, small_dim); in GenerateTensorsForTest() [all …]
|
D | convolution_grad_fp32_tests.cc | 85 auto dy_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(dy_path.c_str(), &dy_size)); in TEST_F() 88 lite::Tensor dy_tensor(TypeId::kNumberTypeFloat32, dim_dy); in TEST_F() 99 …auto input_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(input_path.c_str(), &input_s… in TEST_F() 102 lite::Tensor x_tensor(TypeId::kNumberTypeFloat32, dim_x); in TEST_F() 108 lite::Tensor dw_tensor(TypeId::kNumberTypeFloat32, dim_dw); in TEST_F() 110 std::vector<lite::Tensor *> inputs = {&dy_tensor, &x_tensor}; in TEST_F() 111 std::vector<lite::Tensor *> outputs = {&dw_tensor}; in TEST_F() 113 lite::InnerContext context; in TEST_F() 115 ASSERT_EQ(lite::RET_OK, context.Init()); in TEST_F() 118 auto creator = lite::KernelRegistry::GetInstance()->GetCreator(desc); in TEST_F() [all …]
|
D | pooling_grad_fp32_tests.cc | 76 …auto input_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(input_path.c_str(), &input_s… in TEST_F() 89 auto time_start = mindspore::lite::GetTimeUs(); in TEST_F() 94 auto time_end = mindspore::lite::GetTimeUs(); in TEST_F() 131 …auto input_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(input_path.c_str(), &input_s… in TEST_F() 134 lite::Tensor dy_tensor(TypeId::kNumberTypeFloat32, dim_dy); in TEST_F() 138 …auto input1_data = reinterpret_cast<float *>(mindspore::lite::ReadFile(input1_path.c_str(), &input… in TEST_F() 141 lite::Tensor x_tensor(TypeId::kNumberTypeFloat32, dim_x); in TEST_F() 144 std::vector<lite::Tensor *> inputs = {&x_tensor, &x_tensor, &dy_tensor}; in TEST_F() 149 lite::Tensor dx_tensor(TypeId::kNumberTypeFloat32, dim_dx); in TEST_F() 151 std::vector<lite::Tensor *> outputs = {&dx_tensor}; in TEST_F() [all …]
|
/third_party/mindspore/mindspore/lite/tools/providers/NNIE/Hi3516D/ |
D | compile_nnie.sh | 10 tar -xf mindspore-lite-${version}-linux-x64.tar.gz 13 cd ${open_source_ms_path}/output/mindspore-lite-${version}-linux-x64 || exit 1 14 rm -rf ${nnie_code_path}/mindspore/mindspore/lite/tools/converter/nnie/third_party/ms_lite/ 15 …mkdir -p ${nnie_code_path}/mindspore/mindspore/lite/tools/converter/nnie/third_party/ms_lite/ || e… 16 …cp -r ./tools/ ${nnie_code_path}/mindspore/mindspore/lite/tools/converter/nnie/third_party/ms_lite… 27 …hi3516d_release_path=${open_source_ms_path}/output/mindspore-lite-${version}-linux-x64/tools/conve… 31 …cp ${nnie_code_path}/mindspore/mindspore/lite/tools/providers/NNIE/Hi3516D/opencv-4.2.0/lib/* ${hi… 32 …cp ${nnie_code_path}/mindspore/mindspore/lite/tools/providers/NNIE/Hi3516D/protobuf-3.9.0/lib/* ${… 33 …cp ${nnie_code_path}/mindspore/mindspore/lite/tools/providers/NNIE/Hi3516D/libnnie_mapper.so ${hi3… 34 …cp ${nnie_code_path}/mindspore/mindspore/lite/build/tools/converter/nnie/libmslite_nnie_converter.… [all …]
|
/third_party/mindspore/mindspore/lite/src/runtime/kernel/ascend310/src/ |
D | model_process.cc | 80 return lite::RET_ERROR; in ConstructTensorDesc() 83 return lite::RET_OK; in ConstructTensorDesc() 103 return lite::RET_ERROR; in PreInitModelResource() 106 if (ret != lite::RET_OK) { in PreInitModelResource() 111 if (ret != lite::RET_OK) { in PreInitModelResource() 115 return lite::RET_OK; in PreInitModelResource() 129 return lite::RET_ERROR; in InitInputsBuffer() 140 return lite::RET_ERROR; in InitInputsBuffer() 153 return lite::RET_OK; in InitInputsBuffer() 159 return lite::RET_ERROR; in CreateDataBuffer() [all …]
|