/third_party/mindspore/mindspore/ccsrc/frontend/parallel/tensor_layout/ |
D | tensor_info.h | 35 TensorInfo(const TensorLayout &tensor_layout, Shape shape, Shape slice_shape) in TensorInfo() 37 explicit TensorInfo(const TensorLayout &tensor_layout) : tensor_layout_(tensor_layout) { in TensorInfo() 44 TensorLayout tensor_layout() const { return tensor_layout_; } in tensor_layout() function
|
D | tensor_layout.h | 73 bool IsSameTensorShape(const TensorLayout &tensor_layout) const { in IsSameTensorShape() 77 bool IsSameDeviceArrangement(const TensorLayout &tensor_layout) const { in IsSameDeviceArrangement() 81 …bool IsSameTensorMap(const TensorLayout &tensor_layout) const { return (tensor_map_ == tensor_layo… in IsSameTensorMap()
|
D | redistribution_operator_infer.cc | 26 Status RedistributionOperatorInfer::Init(const TensorLayout &tensor_layout, const Map &out_tensor_m… in Init()
|
/third_party/mindspore/tests/ut/cpp/parallel/tensor_layout/ |
D | tensor_layout_test.cc | 53 TensorLayout tensor_layout; in ReshapeExpandDeviceArrangementTestFunction() local 173 TensorLayout tensor_layout; in ExpandTensorShapeTestFunction() local 226 TensorLayout tensor_layout; in TEST_F() local 244 TensorLayout tensor_layout; in TEST_F() local 267 TensorLayout tensor_layout; in RemoveElementEqualToOneInDeviceArrangementTestFunction() local 324 TensorLayout tensor_layout; in TEST_F() local 342 TensorLayout tensor_layout; in TEST_F() local 363 TensorLayout tensor_layout; in TEST_F() local 384 TensorLayout tensor_layout; in TEST_F() local
|
D | construct_operator_test.cc | 93 TensorLayout tensor_layout; in TEST_F() local
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/graph_util/ |
D | get_parallel_info.cc | 38 auto tensor_layout = para->user_data<parallel::TensorLayout>(); in GetParameterLayout() local
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/ |
D | parameter_manager.cc | 378 auto tensor_layout = param_ptr->user_data<parallel::TensorLayout>(); in IsFullySplitParameter() local 507 auto tensor_layout = cloned_from_parameter->user_data<TensorLayout>(); in SetClonedTensorShapeForOptimizer() local 550 auto tensor_layout = param->user_data<TensorLayout>(); in HandleAdaFactorOpt() local
|
D | step_parallel.cc | 619 TensorLayout tensor_layout = tensor_info.tensor_layout(); in SplitTensor() local 669 TensorLayout tensor_layout = tensor_info.tensor_layout(); in SplitTensorList() local 1609 static std::string GetOptShardGroup(const AnfNodePtr ¶meter, TensorLayout *const tensor_layout, in GetOptShardGroup() 1656 auto tensor_layout = parameter_ptr->user_data<TensorLayout>(); in SetSharedParameterFlag() local 1680 TensorLayout tensor_layout = tensorinfo_in.tensor_layout(); in SetParallelShape() local
|
/third_party/mindspore/mindspore/parallel/ |
D | _utils.py | 300 def _remove_repeated_slices(tensor_layout): argument
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/strategy_checkpoint/ |
D | parallel_strategy_checkpoint.cc | 163 TensorLayoutPtr tensor_layout = node_tensor_info.second; in Save() local
|
/third_party/mindspore/mindspore/train/ |
D | serialization.py | 1507 def _convert_to_layout(param_name, tensor_layout): argument
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/ops_info/ |
D | operator_info.cc | 459 Operator CreateGetTensorSliceOp(const TensorLayout &tensor_layout) { in CreateGetTensorSliceOp() 547 Status OperatorInfo::CreateGroupForOptShard(TensorLayout *const tensor_layout, std::vector<Group> *… in CreateGroupForOptShard()
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/pipeline_transformer/ |
D | pipeline_transformer.cc | 647 auto tensor_layout = tensor_info.tensor_layout(); in InsertReceive() local
|