/third_party/mindspore/mindspore/ops/_op_impl/aicpu/ |
D | stack_push_pop.py | 20 .fusion_type("OPAQUE") \ 25 .fusion_type("OPAQUE") \ 43 .fusion_type("OPAQUE") \ 61 .fusion_type("OPAQUE") \
|
D | init_data_set_queue.py | 20 .fusion_type("OPAQUE") \
|
D | end_of_sequence.py | 20 .fusion_type("OPAQUE") \
|
D | standard_normal.py | 20 .fusion_type("OPAQUE") \
|
D | uniform_real.py | 20 .fusion_type("OPAQUE") \
|
D | standard_laplace.py | 20 .fusion_type("OPAQUE") \
|
D | dropout_genmask.py | 20 .fusion_type("OPAQUE") \
|
D | poisson.py | 20 .fusion_type("OPAQUE") \
|
D | resize_bilinear.py | 20 .fusion_type("OPAQUE") \
|
/third_party/mindspore/config/ |
D | op_info.config | 1 …, {"index": 1, "name": "idx", "param_type": "required"}], "attr": [], "fusion_type": "OPAQUE", "dt… 2 …type": "required"}], "attr": [{"name": "ngram_size", "type": "int"}], "fusion_type": "OPAQUE", "dt… 3 …: [], "outputs": [], "attr": [{"name": "queue_name", "type": "str"}], "fusion_type": "OPAQUE", "dt… 4 …"index": 0, "name": "output", "param_type": "required"}], "attr": [], "fusion_type": "OPAQUE", "dt… 5 …pe": "required"}], "attr": [{"name": "pad_dim_size", "type": "int"}], "fusion_type": "OPAQUE", "dt… 6 …"index": 0, "name": "output", "param_type": "required"}], "attr": [], "fusion_type": "OPAQUE", "dt… 7 …"param_type": "required"}], "attr": [{"name": "dim", "type": "int"}], "fusion_type": "OPAQUE", "dt… 8 …"index": 0, "name": "output", "param_type": "required"}], "attr": [], "fusion_type": "OPAQUE", "dt… 9 …": [{"index": 0, "name": "y", "param_type": "required"}], "attr": [], "fusion_type": "OPAQUE", "dt… 10 …type": "required"}], "attr": [{"name": "normalize", "type": "bool"}], "fusion_type": "OPAQUE", "dt… [all …]
|
/third_party/mindspore/mindspore/nn/wrap/ |
D | grad_reducer.py | 32 fusion_type = 2 ** 10 36 fusion = fusion + (fusion_type,) 40 fusion_type += 1 52 def _init_allreduce_operators_by_parameters(parameters, split_indices, group, fusion_type=1): argument 74 if split_indices and fusion_type == 1: 366 …def __init__(self, parameters, mean=True, degree=None, fusion_type=1, group=GlobalComm.WORLD_COMM_… argument 386 fusion_type) 389 self.allreduce = AllReduce().add_prim_attr('fusion', fusion_type)
|
/third_party/mindspore/mindspore/_extends/graph_kernel/model/ |
D | graph_parallel.py | 22 def __init__(self, fusion_type, bottleneck, gain, block_assign, type_info): argument 23 self.fusion_type = fusion_type 258 fusion_type = "block_fusion" 266 fusion_type = "block_pipeline_fusion" 269 return fusion_type, type_info 289 fusion_type, type_info = ScheduleAnalyzer.fusion_consult(blocks, op_sizes, tuple(exclude_gid)) 290 return ParalGain(fusion_type, max_weight, sum_weight - max_weight, blocks, type_info)
|
/third_party/mindspore/tests/st/graph_kernel/model/ |
D | test_graph_parallel.py | 38 …print("fusion = {}, bottleneck = {}, gain = {}".format(gain.fusion_type, gain.bottleneck, gain.gai… 39 return gain.fusion_type == "block_fusion" and gain.gain > 0
|
/third_party/mindspore/mindspore/ccsrc/backend/optimizer/graph_kernel/ |
D | parallel_cost_model.cc | 82 FusionInfoPtr ParallelCostModel::ProcessFusionInfo(const py::object &fusion_type, const py::object … in ProcessFusionInfo() argument 83 if (!py::isinstance<py::str>(fusion_type)) { in ProcessFusionInfo() 87 std::string fusion_type_name = py::cast<std::string>(fusion_type); in ProcessFusionInfo()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/oplib/ |
D | opinfo.h | 102 fusion_type_ = opinfo.fusion_type(); in OpInfo() 129 std::string fusion_type() const { return fusion_type_; } in fusion_type() function 147 void set_fusion_type(const std::string &fusion_type) { fusion_type_ = fusion_type; } in set_fusion_type() argument
|
/third_party/mindspore/mindspore/ccsrc/backend/optimizer/ascend/format_type/ |
D | dynamic_rnn_grad_reformat.cc | 56 matmul_new_builder->SetFusionType(matmul_kernel_build_info->fusion_type()); in Process() 70 split_new_builder->SetFusionType(split_kernel_build_info->fusion_type()); in Process()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/ |
D | kernel_build_info.h | 89 FusionType fusion_type() const { return fusion_type_; } in fusion_type() function 131 SetFusionType(kernel_build_info->fusion_type()); in KernelBuildInfoBuilder() 167 void SetFusionType(FusionType fusion_type);
|
/third_party/mindspore/mindspore/core/ir/ |
D | param_info.h | 81 void set_comm_fusion(int32_t fusion_type) { fusion_type_ = fusion_type; } in set_comm_fusion() argument
|
/third_party/mindspore/mindspore/ops/_op_impl/tbe/ |
D | atomic_addr_clean.py | 20 .fusion_type("ELEMWISE") \
|
D | npu_alloc_float_status.py | 20 .fusion_type("OPAQUE") \
|
/third_party/mindspore/mindspore/ops/_op_impl/akg/gpu/ |
D | logical_not.py | 19 .fusion_type("OPAQUE") \
|
D | logical_or.py | 19 .fusion_type("OPAQUE") \
|
D | logical_and.py | 19 .fusion_type("OPAQUE") \
|
D | mean.py | 19 .fusion_type("OPAQUE") \
|
/third_party/mindspore/mindspore/ops/_op_impl/akg/ascend/ |
D | batchmatmul.py | 20 .fusion_type("OPAQUE") \
|