1 /** 2 * Copyright 2020-2021 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #ifndef MINDSPORE_LITE_TOOLS_OPTIMIZER_FUSION_CONV_TRANSFORM_FUSION_H_ 18 #define MINDSPORE_LITE_TOOLS_OPTIMIZER_FUSION_CONV_TRANSFORM_FUSION_H_ 19 20 #include <string> 21 #include "tools/optimizer/common/pattern_process_pass_extends.h" 22 #include "include/registry/converter_context.h" 23 24 using mindspore::converter::FmkType; 25 namespace mindspore::opt { 26 class ConvTransformFusion : public LitePatternProcessPass { 27 public: 28 explicit ConvTransformFusion(bool multigraph = true, const std::string &name = "ConvTransformFusion") LitePatternProcessPass(name,multigraph)29 : LitePatternProcessPass(name, multigraph) {} 30 ~ConvTransformFusion() override = default; 31 const AnfNodePtr Process(const FuncGraphPtr &, const AnfNodePtr &, const EquivPtr &) const override; 32 33 protected: 34 virtual int InitTransParam(const CNodePtr &, int, float *, float *) const = 0; 35 36 private: 37 int GenTransParam(const CNodePtr &, int, float *, float *) const; 38 int GenNewConvTensor(const FuncGraphPtr &, const CNodePtr &, int, const float *, const float *) const; 39 int CalNewWeightTensor(const CNodePtr &, const tensor::TensorPtr &, int, const float *) const; 40 int CalNewBiasTensor(float *, int, bool, const float *, const float *) const; 41 bool CheckCanFused(const FuncGraphPtr &func_graph, const CNodePtr &conv_node) const; 42 bool AdjustActivationType(const CNodePtr &conv_node, const CNodePtr &transform_node) const; 43 44 protected: 45 FmkType fmk_type_ = converter::kFmkTypeTf; 46 bool nchw_format_ = false; 47 }; 48 } // namespace mindspore::opt 49 #endif // MINDSPORE_LITE_TOOLS_OPTIMIZER_FUSION_CONV_TRANSFORM_FUSION_H_ 50