1 /** 2 * Copyright 2020 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #ifndef MINDSPORE_LITE_SRC_MINDRT_EXECUTOR_H_ 18 #define MINDSPORE_LITE_SRC_MINDRT_EXECUTOR_H_ 19 20 #include <memory> 21 #include <vector> 22 #include <unordered_map> 23 #include "src/runtime/inner_allocator.h" 24 #include "src/lite_kernel.h" 25 #include "src/lite_mindrt.h" 26 #include "src/executor.h" 27 #include "include/lite_session.h" 28 29 namespace mindspore::lite { 30 class MindrtExecutor : public Executor { 31 public: MindrtExecutor(std::unordered_map<Tensor *,Tensor * > * output_map)32 explicit MindrtExecutor(std::unordered_map<Tensor *, Tensor *> *output_map) : output_tensor_map_(output_map) {} ~MindrtExecutor()33 virtual ~MindrtExecutor() { MindrtTerminate(op_actors_); } 34 35 int Prepare(const std::vector<kernel::LiteKernel *> &kernels, const std::vector<Tensor *> &inputs, 36 const std::vector<Tensor *> &outputs, const lite::InnerContext *ctx) override; 37 38 int Run(const std::vector<Tensor *> &in_tensors, const std::vector<Tensor *> &out_tensors, 39 const std::vector<kernel::LiteKernel *> &kernels, const KernelCallBack &before = nullptr, 40 const KernelCallBack &after = nullptr) override; 41 42 int Resize(const std::vector<mindspore::tensor::MSTensor *> &inputs, 43 const std::vector<std::vector<int>> &dims) override; 44 45 private: 46 void TransferGraphOutput(); 47 void FreeOutputTensor(); 48 49 protected: 50 int PrepareInputData(const std::vector<kernel::LiteKernel *> &kernels, const std::vector<Tensor *> &inputs); 51 int PrepareOutputData(const std::vector<kernel::LiteKernel *> &kernels, const std::vector<Tensor *> &outputs); 52 std::vector<std::shared_ptr<LiteOpActor>> op_actors_; 53 std::vector<OpDataPtr<Tensor>> input_data_; 54 std::vector<OpDataPtr<Tensor>> output_data_; 55 std::unordered_map<Tensor *, Tensor *> *output_tensor_map_; 56 }; 57 58 } // namespace mindspore::lite 59 #endif 60