1 /** 2 * Copyright 2022 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef MINDSPORE_LITE_SRC_EXTENDRT_CXX_API_MODEL_POOL_MODEL_PARALLEL_RUNNER_IMPL_H_ 17 #define MINDSPORE_LITE_SRC_EXTENDRT_CXX_API_MODEL_POOL_MODEL_PARALLEL_RUNNER_IMPL_H_ 18 #include <vector> 19 #include <memory> 20 #include <utility> 21 #include <map> 22 #include <string> 23 #include "src/extendrt/cxx_api/model_pool/model_pool.h" 24 #include "include/api/context.h" 25 namespace mindspore { 26 class ModelParallelRunnerImpl { 27 public: 28 ModelParallelRunnerImpl() = default; 29 ~ModelParallelRunnerImpl(); 30 31 Status Init(const std::string &model_path, const std::shared_ptr<RunnerConfig> &runner_config = nullptr); 32 33 Status Init(const void *model_data, size_t data_size, const std::shared_ptr<RunnerConfig> &runner_config = nullptr); 34 35 std::vector<MSTensor> GetInputs(); 36 37 std::vector<MSTensor> GetOutputs(); 38 39 Status Predict(const std::vector<MSTensor> &inputs, std::vector<MSTensor> *outputs, 40 const MSKernelCallBack &before = nullptr, const MSKernelCallBack &after = nullptr); 41 42 private: 43 ModelPool *model_pool_ = nullptr; 44 std::shared_mutex model_parallel_runner_impl_mutex_; 45 }; 46 } // namespace mindspore 47 #endif // MINDSPORE_LITE_SRC_EXTENDRT_CXX_API_MODEL_POOL_MODEL_PARALLEL_RUNNER_IMPL_H_ 48