1 /** 2 * Copyright 2020-2021 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #ifndef MINDSPORE_LITE_TOOLS_OPTIMIZER_FUSION_BATCHNORM_TO_SCALE_FUSION_H_ 18 #define MINDSPORE_LITE_TOOLS_OPTIMIZER_FUSION_BATCHNORM_TO_SCALE_FUSION_H_ 19 20 #include <vector> 21 #include "tools/optimizer/common/pattern_process_pass_extends.h" 22 23 namespace mindspore::opt { 24 class BatchNormToScaleFusion : public Pass { 25 public: BatchNormToScaleFusion()26 BatchNormToScaleFusion() : Pass("BatchNormToScaleFusion") {} 27 ~BatchNormToScaleFusion() override = default; 28 bool Run(const FuncGraphPtr &func_graph) override; 29 30 private: 31 bool CheckBNCanFused(const AnfNodePtr &node); 32 33 private: 34 std::vector<int64_t> input_shape_; 35 }; 36 37 int CalculateScaleAndBiasFromBN(const CNodePtr &bn_node, int kernel_num, float *trans_scale, float *trans_bias); 38 } // namespace mindspore::opt 39 #endif // MINDSPORE_LITE_TOOLS_OPTIMIZER_FUSION_BATCHNORM_TO_SCALE_FUSION_H_ 40