/third_party/mindspore/mindspore/ccsrc/backend/optimizer/graph_kernel/ |
D | reorder_ops.cc | 104 auto node_input = AnfAlgo::GetInputNode(new_node, 0); in SetNodeInfo() local 105 MS_EXCEPTION_IF_NULL(node_input); in SetNodeInfo() 106 MS_EXCEPTION_IF_NULL(node_input->abstract()); in SetNodeInfo() 108 node_input->abstract()->BuildShape()); in SetNodeInfo() 217 auto node_input = AnfAlgo::GetInputNode(node, 0); in ReorderTypeInsensitiveCastDown() local 218 auto type_insens_node = node_input->cast<CNodePtr>(); in ReorderTypeInsensitiveCastDown() 274 auto node_input = AnfAlgo::GetInputNode(node, index); in ReorderCastUpTypeInsensitive() local 275 auto cast_node = node_input->cast<CNodePtr>(); in ReorderCastUpTypeInsensitive()
|
/third_party/mindspore/mindspore/ccsrc/frontend/optimizer/irpass/ |
D | special_op_eliminate.h | 294 std::vector<AnfNodePtr> node_input = in operator() local 296 auto prim_anf_node = node_input[0]->cast<ValueNodePtr>(); in operator() 306 CNodePtr new_node = func_graph->NewCNode(node_input); in operator() 333 std::vector<AnfNodePtr> node_input = in operator() local 335 auto prim_anf_node = node_input[0]->cast<ValueNodePtr>(); in operator() 345 CNodePtr new_node = func_graph->NewCNode(node_input); in operator()
|
/third_party/mindspore/mindspore/ccsrc/backend/session/ |
D | ascend_auto_monad.cc | 894 auto node_input = node->input(i); in FindInputNode() local 896 if (HasAbstractMonad(node_input)) { in FindInputNode() 899 MS_EXCEPTION_IF_NULL(node_input); in FindInputNode() 900 MS_LOG(DEBUG) << "check node input[" << i << "]: " << node_input->DebugString(); in FindInputNode() 901 if (node_input->isa<Parameter>()) { in FindInputNode() 902 MS_LOG(DEBUG) << "node_input:" << node_input->DebugString() << " is a param"; in FindInputNode() 903 CNodePtr stack_pop = InsertStackPop(node_input, stack_pushs); in FindInputNode() 909 … [node_input](auto before_node) { return before_node == node_input; }); in FindInputNode() 919 CNodePtr InsertStackPop(const AnfNodePtr &node_input, std::vector<CNodePtr> *stack_pushs) { in InsertStackPop() argument 920 MS_EXCEPTION_IF_NULL(node_input); in InsertStackPop() [all …]
|
D | session_basic.cc | 749 CNodePtr SessionBasic::CreateSwitchInput(const CNodePtr &cnode, const AnfNodePtr &node_input, Kerne… in CreateSwitchInput() argument 750 MS_EXCEPTION_IF_NULL(node_input); in CreateSwitchInput() 754 if (AnfAlgo::CheckPrimitiveType(node_input, prim::kPrimPartial)) { in CreateSwitchInput() 755 auto backend_node = graph->GetBackendAnfByFrontAnf(node_input); in CreateSwitchInput() 757 } else if (node_input->isa<ValueNode>() && IsValueNode<FuncGraph>(node_input)) { in CreateSwitchInput() 758 partial_inputs.emplace_back(graph->GetBackendAnfByFrontAnf(node_input)); in CreateSwitchInput() 770 partial_inputs.emplace_back(graph->GetBackendAnfByFrontAnf(node_input)); in CreateSwitchInput() 1007 auto node_input = cnode->input(index); in CreateCNodeInputs() local 1008 auto switch_input = CreateSwitchInput(cnode, node_input, graph); in CreateCNodeInputs()
|
D | session_basic.h | 172 …CNodePtr CreateSwitchInput(const CNodePtr &cnode, const AnfNodePtr &node_input, KernelGraph *graph…
|
/third_party/mindspore/mindspore/ccsrc/frontend/parallel/ |
D | step_parallel.cc | 205 std::vector<AnfNodePtr> node_input; in InsertNode() local 207 node_input = CreateMirrorInput(root, op, pre_node, instance_name, param_name); in InsertNode() 209 node_input = CreateInput(op, pre_node, instance_name); in InsertNode() 211 CNodePtr new_node = func_graph->NewCNode(node_input); in InsertNode() 216 auto new_node_value = node_input[0]->cast<ValueNodePtr>(); in InsertNode() 225 node_input[0]->set_scope(scope); in InsertNode() 239 std::vector<AnfNodePtr> node_input; in ReplaceNode() local 241 node_input = CreateMirrorInput(root, op, pre_node, instance_name, param_name); in ReplaceNode() 243 node_input = CreateInput(op, pre_node, instance_name); in ReplaceNode() 245 CNodePtr new_node = func_graph->NewCNode(node_input); in ReplaceNode() [all …]
|
/third_party/mindspore/mindspore/ccsrc/backend/optimizer/ascend/mindir/ |
D | dropout_unify_mindir.cc | 125 CNodePtr CreateDynamicShapeCNode(const FuncGraphPtr &func_graph, const AnfNodePtr &node_input, in CreateDynamicShapeCNode() argument 129 …dePtr> dynamic_shape_inputs{NewValueNode(std::make_shared<Primitive>("DynamicShape")), node_input}; in CreateDynamicShapeCNode()
|
/third_party/mindspore/mindspore/ccsrc/pipeline/jit/static_analysis/ |
D | program_specialize.cc | 334 auto node_input = old_inputs[i]; in ProcessNode() local 335 AnfNodeConfigPtr iconf = MakeConfig(node_input); in ProcessNode() 345 MS_LOG(DEBUG) << "Build possible value node for node: " << node_input->DebugString() in ProcessNode()
|
/third_party/chromium/patch/ |
D | 0001-cve.patch | 150701 - Node* node_input) const { 150702 - if (CanCover(user, node) && CanCover(node, node_input)) { 150705 - // If {node_input} is pure, the effect levels do not matter. 150706 - if (node_input->op()->HasProperty(Operator::kPure)) return true; 150707 - // Otherwise, {user} and {node_input} must have the same effect level. 150708 - return GetEffectLevel(user) == GetEffectLevel(node_input); 150774 - bool CanCoverTransitively(Node* user, Node* node, Node* node_input) const;
|