1 /** 2 * Copyright 2020 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ASCEND_IR_FUSION_CONFUSION_SOFTMAX_GRAD_RULE_H_ 17 #define MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ASCEND_IR_FUSION_CONFUSION_SOFTMAX_GRAD_RULE_H_ 18 19 #include <memory> 20 #include "backend/optimizer/common/optimizer.h" 21 22 namespace mindspore { 23 namespace opt { 24 class ConfusionSoftmaxGradRule : public PatternProcessPass { 25 public: 26 explicit ConfusionSoftmaxGradRule(bool multigraph = true) 27 : PatternProcessPass("confusion_softmax_grad_rule", multigraph) { 28 input0_ = std::make_shared<Var>(); 29 input1_ = std::make_shared<Var>(); 30 reduce_sum_ = std::make_shared<Var>(std::make_shared<Primitive>(prim::kPrimReduceSum->name())); 31 } 32 ~ConfusionSoftmaxGradRule() override = default; 33 const BaseRef DefinePattern() const override; 34 const AnfNodePtr Process(const FuncGraphPtr &, const AnfNodePtr &, const EquivPtr &) const override; 35 36 private: 37 VarPtr input0_; 38 VarPtr input1_; 39 VarPtr reduce_sum_; 40 }; 41 } // namespace opt 42 } // namespace mindspore 43 #endif // MINDSPORE_CCSRC_BACKEND_OPTIMIZER_ASCEND_IR_FUSION_CONFUSION_SOFTMAX_GRAD_RULE_H_ 44