1 /** 2 * Copyright 2022 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef MINDSPORE_CCSRC_BACKEND_OPTIMIZER_PASS_INSERT_TENSOR_MOVE_FOR_COMMUNICATION_H_ 17 #define MINDSPORE_CCSRC_BACKEND_OPTIMIZER_PASS_INSERT_TENSOR_MOVE_FOR_COMMUNICATION_H_ 18 19 #include "include/backend/optimizer/optimizer.h" 20 21 namespace mindspore { 22 namespace opt { 23 // If the input Tensor of the graph is connected to the AllReduce operator, 24 // and the input Tensor of the graph already has a device address, 25 // we need to copy the data in the device address to the contiguous memory of AllReduce. 26 class BACKEND_EXPORT InsertTensorMoveForCommunication : public Pass { 27 public: InsertTensorMoveForCommunication()28 InsertTensorMoveForCommunication() : Pass("insert_tensor_move_for_communication") {} 29 ~InsertTensorMoveForCommunication() override = default; 30 bool Run(const FuncGraphPtr &graph) override; 31 }; 32 } // namespace opt 33 } // namespace mindspore 34 35 #endif // MINDSPORE_CCSRC_BACKEND_OPTIMIZER_PASS_INSERT_TENSOR_MOVE_FOR_COMMUNICATION_H_ 36