1 /** 2 * Copyright 2020 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ADD_LAYERNORM_FUSION_H_ 17 #define MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ADD_LAYERNORM_FUSION_H_ 18 19 #include <memory> 20 #include "include/backend/optimizer/optimizer.h" 21 22 namespace mindspore { 23 namespace opt { 24 class AddLayernormFusion : public PatternProcessPass { 25 public: 26 explicit AddLayernormFusion(bool multigraph = true) : PatternProcessPass("add_layer_norm_fusion", multigraph) { 27 x1_ = std::make_shared<Var>(); 28 x2_ = std::make_shared<Var>(); 29 gamma_ = std::make_shared<Var>(); 30 beta_ = std::make_shared<Var>(); 31 begin_norm_axis_ = std::make_shared<Var>(); 32 begin_params_axis_ = std::make_shared<Var>(); 33 eps_ = std::make_shared<Var>(); 34 } 35 ~AddLayernormFusion() override = default; 36 const BaseRef DefinePattern() const override; 37 const AnfNodePtr Process(const FuncGraphPtr &, const AnfNodePtr &, const EquivPtr &) const override; 38 39 private: 40 VarPtr x1_; 41 VarPtr x2_; 42 VarPtr gamma_; 43 VarPtr beta_; 44 VarPtr begin_norm_axis_; 45 VarPtr begin_params_axis_; 46 VarPtr eps_; 47 }; 48 } // namespace opt 49 } // namespace mindspore 50 #endif // MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ADD_LAYERNORM_FUSION_H_ 51