1 /** 2 * Copyright 2020 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ASCEND_IR_FISSION_LAYER_NORM_GRAD_SPLIT_H_ 17 #define MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ASCEND_IR_FISSION_LAYER_NORM_GRAD_SPLIT_H_ 18 19 #include <vector> 20 #include <memory> 21 #include "backend/optimizer/common/optimizer.h" 22 #include "backend/optimizer/common/helper.h" 23 #include "backend/optimizer/ascend/ascend_helper.h" 24 25 namespace mindspore { 26 namespace opt { 27 class LayerNormGradSplit : public PatternProcessPass { 28 public: 29 explicit LayerNormGradSplit(bool multigraph = true) : PatternProcessPass("layer_norm_grad_split", multigraph) {} 30 ~LayerNormGradSplit() override = default; 31 const BaseRef DefinePattern() const override; 32 const AnfNodePtr Process(const FuncGraphPtr &, const AnfNodePtr &, const EquivPtr &) const override; 33 34 private: 35 void CreateOutputsOfLayerNormXBackpropV2(const FuncGraphPtr &graph, const CNodePtr &layer_norm_grad, 36 std::vector<AnfNodePtr> *layer_norm_grad_outputs, bool is_dynamic) const; 37 void CreateOutputsOfLayerNormBetaGammaBackpropV2(const FuncGraphPtr &graph, const CNodePtr &layer_norm_grad, 38 const AnfNodePtr &res_for_gamma, 39 std::vector<AnfNodePtr> *layer_norm_beta_gamma_outputs, 40 bool is_dynamic) const; 41 }; 42 } // namespace opt 43 } // namespace mindspore 44 #endif // MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ASCEND_IR_FISSION_LAYER_NORM_GRAD_SPLIT_H_ 45