• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /**
2  * Copyright 2020-2021 Huawei Technologies Co., Ltd
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  * http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 #ifndef MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_
18 #define MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_
19 #include <vector>
20 #include <string>
21 #include "include/graph/op/all_ops.h"
22 #include "include/graph/op/nn_defs.h"
23 #include "src/delegate/npu/op/npu_op.h"
24 
25 namespace mindspore {
26 class ScaleNPUOp : public NPUOp {
27  public:
ScaleNPUOp(const schema::Primitive * primitive,const std::vector<mindspore::MSTensor> & in_tensors,const std::vector<mindspore::MSTensor> & out_tensors,std::string name)28   ScaleNPUOp(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors,
29              const std::vector<mindspore::MSTensor> &out_tensors, std::string name)
30       : NPUOp(primitive, in_tensors, out_tensors, name) {}
31 
32   ~ScaleNPUOp() override;
33 
34   int IsSupport(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors,
35                 const std::vector<mindspore::MSTensor> &out_tensors) override;
36 
37   int Init(const schema::Primitive *primitive, const std::vector<mindspore::MSTensor> &in_tensors,
38            const std::vector<mindspore::MSTensor> &out_tensors) override;
39 
40   int SetNPUInputs(const std::vector<mindspore::MSTensor> &in_tensors,
41                    const std::vector<mindspore::MSTensor> &out_tensors,
42                    const std::vector<ge::Operator *> &npu_inputs) override;
43 
44   ge::Operator *GetNPUOp() override;
45 
GetAxis()46   int GetAxis() { return axis_; }
47 
48  private:
49   int SetActivation(const ge::Operator *input);
50 
51   int ConvertScaleToMul(const std::vector<ge::Operator *> &npu_inputs, ge::Operator *cur_op,
52                         const std::vector<mindspore::MSTensor> &in_tensors);
53 
54   int axis_ = 0;
55   bool use_mul_ = false;
56   schema::ActivationType act_type_ = schema::ActivationType_NO_ACTIVATION;
57   ge::Operator *op_ = nullptr;
58   hiai::op::Reshape *reshape_ = nullptr;
59   hiai::op::Const *scale_ = nullptr;
60   hiai::op::Const *bias_ = nullptr;
61   hiai::op::Const *shape_ = nullptr;
62   hiai::op::Activation *act_ = nullptr;
63 };
64 }  // namespace mindspore
65 #endif  // MINDSPORE_LITE_SRC_RUNTIME_DELEGATE_NPU_OP_SCALE_NPU_H_
66