• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /**
2  * Copyright 2020-2022 Huawei Technologies Co., Ltd
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  * http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #ifndef MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_
18 #define MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_
19 #include <vector>
20 #include <string>
21 #include "include/graph/op/all_ops.h"
22 #include "include/graph/op/nn_defs.h"
23 #include "src/litert/delegate/npu/op/npu_op.h"
24 
25 namespace mindspore::lite {
26 class ScaleNPUOp : public NPUOp {
27  public:
ScaleNPUOp(const schema::Primitive * primitive,const std::vector<mindspore::MSTensor> & in_tensors,const std::vector<mindspore::MSTensor> & out_tensors,std::string name)28   ScaleNPUOp(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors,
29              const std::vector<mindspore::MSTensor> &out_tensors, std::string name)
30       : NPUOp(primitive, in_tensors, out_tensors, name) {}
31 
32   ~ScaleNPUOp() override;
33 
34   int IsSupport(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors,
35                 const std::vector<mindspore::MSTensor> &out_tensors) override;
36 
37   int Init(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors,
38            const std::vector<mindspore::MSTensor> &out_tensors) override;
39 
40   int SetNPUInputs(const std::vector<mindspore::MSTensor> &in_tensors,
41                    const std::vector<mindspore::MSTensor> &out_tensors,
42                    const std::vector<ge::Operator *> &npu_inputs) override;
43 
44   ge::Operator *GetNPUOp() override;
45 
GetAxis()46   int GetAxis() { return axis_; }
47 
48  private:
49   int SetActivation();
50 
51   int ConvertScaleToMul(const std::vector<ge::Operator *> &npu_inputs,
52                         const std::vector<mindspore::MSTensor> &in_tensors);
53 
54   int Adopt4DScale(const std::vector<ge::Operator *> &npu_inputs, const std::vector<mindspore::MSTensor> &in_tensors);
55 
56   ge::Operator *ChangeDims(const ge::Operator *input, std::vector<int64_t> dst_shape, std::string name,
57                            bool need_expand_4d = false);
58 
59   int axis_ = 0;
60   bool use_mul_ = false;
61   bool need_expand_ = false;
62   schema::ActivationType act_type_ = schema::ActivationType_NO_ACTIVATION;
63   hiai::op::Reshape *out_reshape_ = nullptr;
64   hiai::op::Scale *scale_ = nullptr;
65   hiai::op::Mul *mul_ = nullptr;
66   hiai::op::Activation *act_ = nullptr;
67   std::vector<ge::Operator *> scale_ops_ = {};
68 };
69 }  // namespace mindspore::lite
70 #endif  // MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_
71