1 /** 2 * Copyright 2020-2022 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #ifndef MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_ 18 #define MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_ 19 #include <vector> 20 #include <string> 21 #include "include/graph/op/all_ops.h" 22 #include "include/graph/op/nn_defs.h" 23 #include "src/litert/delegate/npu/op/npu_op.h" 24 25 namespace mindspore::lite { 26 class ScaleNPUOp : public NPUOp { 27 public: ScaleNPUOp(const schema::Primitive * primitive,const std::vector<mindspore::MSTensor> & in_tensors,const std::vector<mindspore::MSTensor> & out_tensors,std::string name)28 ScaleNPUOp(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors, 29 const std::vector<mindspore::MSTensor> &out_tensors, std::string name) 30 : NPUOp(primitive, in_tensors, out_tensors, name) {} 31 32 ~ScaleNPUOp() override; 33 34 int IsSupport(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors, 35 const std::vector<mindspore::MSTensor> &out_tensors) override; 36 37 int Init(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors, 38 const std::vector<mindspore::MSTensor> &out_tensors) override; 39 40 int SetNPUInputs(const std::vector<mindspore::MSTensor> &in_tensors, 41 const std::vector<mindspore::MSTensor> &out_tensors, 42 const std::vector<ge::Operator *> &npu_inputs) override; 43 44 ge::Operator *GetNPUOp() override; 45 GetAxis()46 int GetAxis() { return axis_; } 47 48 private: 49 int SetActivation(); 50 51 int ConvertScaleToMul(const std::vector<ge::Operator *> &npu_inputs, 52 const std::vector<mindspore::MSTensor> &in_tensors); 53 54 int Adopt4DScale(const std::vector<ge::Operator *> &npu_inputs, const std::vector<mindspore::MSTensor> &in_tensors); 55 56 ge::Operator *ChangeDims(const ge::Operator *input, std::vector<int64_t> dst_shape, std::string name, 57 bool need_expand_4d = false); 58 59 int axis_ = 0; 60 bool use_mul_ = false; 61 bool need_expand_ = false; 62 schema::ActivationType act_type_ = schema::ActivationType_NO_ACTIVATION; 63 hiai::op::Reshape *out_reshape_ = nullptr; 64 hiai::op::Scale *scale_ = nullptr; 65 hiai::op::Mul *mul_ = nullptr; 66 hiai::op::Activation *act_ = nullptr; 67 std::vector<ge::Operator *> scale_ops_ = {}; 68 }; 69 } // namespace mindspore::lite 70 #endif // MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_ 71