/external/pytorch/torch/csrc/jit/passes/ |
D | variadic_ops.cpp | 55 void recordSchema(Node* op_node) { in recordSchema() argument 56 const auto& schema = op_node->schema(); in recordSchema() 64 const std::vector<size_t>& getListIndices(Node* op_node) const { in getListIndices() 65 const auto& schema = op_node->schema(); in getListIndices() 83 bool allListInputsAreValid(Node* op_node) { in allListInputsAreValid() argument 84 const size_t num_inputs = op_node->inputs().size(); in allListInputsAreValid() 85 for (const auto list_idx : getListIndices(op_node)) { in allListInputsAreValid() 87 const auto list = op_node->input(list_idx)->node(); in allListInputsAreValid() 92 !alias_db_.couldMoveBeforeTopologically(list, op_node)) { in allListInputsAreValid() 120 void deleteOpNodeAndLists(Node* op_node) { in deleteOpNodeAndLists() argument [all …]
|
/external/pytorch/torch/ao/quantization/fx/ |
D | _equalize.py | 347 op_node = None 350 op_node = user 353 assert op_node is not None 354 if op_node.op == "call_module": 362 assert equalization_node_name_to_qconfig.get(op_node.name, None) is not None 364 op_node.name, None 368 return op_node, weight_eq_obs 370 elif op_node.op == "call_function": 371 weight_node = maybe_get_weight_eq_obs_node(op_node, modules) 375 return op_node, weight_eq_obs [all …]
|
D | _lower_to_native_backend.py | 482 op_node = next(iter(prepack_node.users)) 483 module_path, _ = node_name_to_scope[op_node.name]
|
/external/tensorflow/tensorflow/lite/delegates/hexagon/builders/ |
D | op_builder.h | 295 const OpNode* op_node = builders_[i]->Build(); in Build() local 297 graph_id_, op_node->node_id, op_node->op_type, op_node->padding_type, in Build() 298 op_node->inputs.data(), op_node->inputs.size(), in Build() 299 op_node->outputs.data(), op_node->outputs.size()); in Build() 301 printf("Error adding node: id:%d, op_type:%d\n", op_node->node_id, in Build() 302 op_node->op_type); in Build()
|
/external/executorch/backends/cadence/aot/quantizer/ |
D | fusion_pass.py | 215 op_node: fx.Node, 221 stride = [1, 1] if len(op_node.args) < 4 else get_conv_args(op_node.args[3], 1) 222 padding = [0, 0] if len(op_node.args) < 5 else get_conv_args(op_node.args[4], 0) 223 dilation = [1, 1] if len(op_node.args) < 6 else get_conv_args(op_node.args[5], 1) 224 groups = 1 if len(op_node.args) < 7 else op_node.args[6] 378 op_node = anchors.output[0][0] 380 assert len(op_node.users) == 1 381 quant_node = list(op_node.users.keys())[0] 383 with graph_module.graph.inserting_after(op_node): 397 op_node,
|
/external/tensorflow/tensorflow/lite/delegates/gpu/common/transformations/ |
D | merge_padding_with.cc | 56 Node* op_node = sequence.back(); in ApplyToNodesSequence() local 70 Attr* node_attr = absl::any_cast<Attr>(&op_node->operation.attributes); in ApplyToNodesSequence() 71 absl::Status status = RemovePrecedingNode(graph, pad_node, op_node); in ApplyToNodesSequence()
|
D | remove_noop.cc | 54 Node* op_node = sequence.back(); in ApplyToNodesSequence() local 55 if (!remove_predicate_(graph, op_node)) { in ApplyToNodesSequence() 58 absl::Status status = RemoveFollowingNode(graph, op_node, prev_op_node); in ApplyToNodesSequence()
|
/external/tensorflow/tensorflow/security/fuzzing/op_fuzzing/ |
D | fuzz_session.h | 39 auto op_node = \ 41 (void)tensorflow::ops::opName(scope.WithOpName("output"), op_node); \
|
/external/tensorflow/tensorflow/core/kernels/fuzzing/ |
D | fuzz_session.h | 40 auto op_node = \ 42 (void)tensorflow::ops::opName(scope.WithOpName("output"), op_node); \
|
/external/mesa3d/src/amd/compiler/ |
D | aco_lower_to_cssa.cpp | 337 merge_node& op_node = ctx.merge_node_table[copy.op.tempId()]; in try_coalesce_copy() local 338 if (op_node.defined_at == -1u) { in try_coalesce_copy() 344 op_node.defined_at = block_idx; in try_coalesce_copy() 345 op_node.value = copy.op; in try_coalesce_copy() 349 if (op_node.index == -1u) { in try_coalesce_copy() 356 if (op_node.index == ctx.merge_node_table[copy.def.tempId()].index) in try_coalesce_copy() 360 return try_merge_merge_set(ctx, copy.def.getTemp(), ctx.merge_sets[op_node.index]); in try_coalesce_copy()
|
/external/pytorch/torch/csrc/jit/passes/onnx/ |
D | unpack_quantized_weights.cpp | 726 auto op_node = match_vmap.at(vmap.at("r"))->node(); in insertPermutesHelper() local 731 permute_node_before->insertBefore(op_node); in insertPermutesHelper() 732 op_node->removeInput(0); in insertPermutesHelper() 733 op_node->insertInput(0, permute_node_before->output()); in insertPermutesHelper() 737 {op_node->outputs()[0]}); in insertPermutesHelper() 738 permute_node_after->insertAfter(op_node); in insertPermutesHelper() 739 auto v = op_node->outputs().at(0); in insertPermutesHelper()
|
/external/tensorflow/tensorflow/dtensor/cc/ |
D | dtensor_device_util.cc | 794 tensorflow::Node* op_node = graph->AddNode(op_node_def, &status); in PrepareGraphForMlir() local 798 graph->AddEdge(graph_op_inputs[i].node, 0, op_node, i); in PrepareGraphForMlir() 800 TF_RETURN_IF_ERROR(shape_refiner.AddNode(op_node)); in PrepareGraphForMlir() 803 output_layouts->reserve(op_node->num_outputs()); in PrepareGraphForMlir() 804 global_output_shapes->reserve(op_node->num_outputs()); in PrepareGraphForMlir() 805 for (int output_index = 0; output_index < op_node->num_outputs(); in PrepareGraphForMlir() 810 tensorflow::DataType output_type = op_node->output_type(output_index); in PrepareGraphForMlir() 818 graph->AddEdge(op_node, output_index, ret_node, 0); in PrepareGraphForMlir() 822 shape_refiner.GetContext(op_node); in PrepareGraphForMlir()
|
/external/executorch/backends/qualcomm/builders/ |
D | node_visitor.py | 92 def get_tensor(self, input_node, op_node, idx=None): argument 109 if len(tensor.shape) != 0 and QCOM_AXIS_ORDER in op_node.meta: 110 tensor = tensor.permute(dims=op_node.meta[QCOM_AXIS_ORDER]).contiguous()
|
/external/tensorflow/tensorflow/core/ops/ |
D | math_grad_test.cc | 43 Status Unary(const FDH::Node& op_node, const Tensor& x, const DataType dst, in Unary() argument 53 op_node, in Unary() 98 const FDH::Node op_node = {{"y"}, op, {"x"}, {{"T", x.dtype()}}}; in Unary() local 99 return Unary(op_node, x, x.dtype(), y); in Unary() 111 const FDH::Node op_node = { in SymCastGrad() local 113 TF_CHECK_OK(Unary(op_node, x, dst, &ret)); in SymCastGrad()
|
/external/tensorflow/tensorflow/core/grappler/optimizers/ |
D | loop_optimizer.cc | 554 Status EvaluateBoolOpForConstantOperands(const NodeDef& op_node, in EvaluateBoolOpForConstantOperands() argument 560 VLOG(4) << "Evaluate bool op: op_node=" << op_node.name() in EvaluateBoolOpForConstantOperands() 576 EvaluateNode(op_node, inputs, cpu_device, resource_mgr, &outputs)); in EvaluateBoolOpForConstantOperands()
|
/external/pytorch/test/export/ |
D | test_export.py | 5891 op_node = [ 5894 self.assertEqual(op_node.target._name, "aten::sub.Tensor") 5910 op_node = [ 5913 self.assertEqual(op_node.target._name, "aten::add.Tensor")
|