/external/tensorflow/tensorflow/python/tools/ |
D | strip_unused_lib.py | 32 def strip_unused(input_graph_def, input_node_names, output_node_names, argument 88 output_node_names) 93 output_binary, input_node_names, output_node_names, argument 101 if not output_node_names: 115 output_node_names.split(","),
|
D | freeze_graph_test.py | 83 output_node_names = "output_node" 94 output_node_names, 195 output_node_names = "output_node" 203 output_node_names, restore_op_name, filename_tensor_name, 235 output_node_names = "output_node" 247 input_binary, checkpoint_path, output_node_names, 326 output_node_names = "save/restore_all" 334 output_node_names=output_node_names,
|
D | freeze_graph.py | 80 output_node_names, argument 131 if not output_node_names: 224 output_node_names.replace(" ", "").split(","), 231 output_node_names.replace(" ", "").split(","), 290 output_node_names, argument 350 output_node_names, 373 flags.input_checkpoint, flags.output_node_names,
|
D | strip_unused_test.py | 62 output_node_names = "output_node" 69 output_node_names, 129 output_node_names = "output_node" 135 output_node_names,
|
D | optimize_for_inference_lib.py | 93 def optimize_for_inference(input_graph_def, input_node_names, output_node_names, argument 114 optimized_graph_def, input_node_names, output_node_names, 117 optimized_graph_def, output_node_names) 121 output_node_names) 419 def fuse_resize_and_conv(input_graph_def, output_node_names): argument 451 for output_name in output_node_names:
|
D | strip_unused.py | 60 FLAGS.output_node_names,
|
D | saved_model_aot_compile.py | 327 output_node_names=[
|
/external/tensorflow/tensorflow/core/kernels/hexagon/ |
D | graph_transferer.h | 63 const std::vector<string>& output_node_names, 73 const std::vector<string>& output_node_names, const bool is_text_proto, 79 void SortParams(const std::vector<string>& output_node_names); 122 const std::vector<string>& output_node_names); 173 const std::vector<string>& output_node_names);
|
D | graph_transferer_test.cc | 307 const std::vector<string> output_node_names = {NAME_A_PLUS_B}; in TEST_F() local 309 inputs, output_node_names, false); in TEST_F() 318 const std::vector<string> output_node_names = {"softmax"}; in TEST_F() local 320 input_node_info_list, output_node_names, in TEST_F() 344 const std::vector<string> output_node_names = {"softmax"}; in TEST_F() local 346 input_node_info_list, output_node_names, in TEST_F() 379 std::vector<string> output_node_names = {}; in TEST() local 393 *ops_definitions, filename, input_node_info_list, output_node_names, in TEST() 442 std::vector<string> output_node_names = {}; in TEST() local 459 *ops_definitions, filename, input_node_info_list, output_node_names, in TEST() [all …]
|
D | hexagon_graph_execution_test.cc | 306 std::vector<string> output_node_names; in RunFusedGraph() local 307 output_node_names.emplace_back(REMOTE_FUSED_GRAPH_EXECUTE_NODE_NAME); in RunFusedGraph() 311 TF_ASSERT_OK(session->Run(run_options, input_tensors, output_node_names, {}, in RunFusedGraph() 413 std::vector<string> output_node_names = {"softmax"}; in TEST() local 421 *ops_definitions, MODEL_FILENAME, inputs, output_node_names, in TEST() 446 std::vector<string> output_node_names = {"softmax"}; in TEST() local 455 output_node_names, in TEST() 480 std::vector<string> output_node_names = {"softmax"}; in TEST() local 488 *ops_definitions, MODEL_FILENAME, inputs, output_node_names, in TEST() 549 std::vector<string> output_node_names = {"softmax"}; in TEST() local [all …]
|
D | graph_transferer.cc | 95 const std::vector<string>& output_node_names, in LoadGraphFromProto() argument 142 output_node_names); in LoadGraphFromProto() 149 SortParams(output_node_names); in LoadGraphFromProto() 162 for (const string& output_node_name : output_node_names) { in LoadGraphFromProto() 202 const std::vector<string>& output_node_names, const bool is_text_proto, in LoadGraphFromProtoFile() argument 236 output_node_names, in LoadGraphFromProtoFile() 240 void GraphTransferer::SortParams(const std::vector<string>& output_node_names) { in SortParams() argument 254 for (const string& output_node_name : output_node_names) { in SortParams() 425 const std::vector<string>& output_node_names) { in RegisterNode() argument 843 const std::vector<string>& output_node_names) { in RegisterNodeIfAllInputsAreCached() argument [all …]
|
/external/tensorflow/tensorflow/python/framework/ |
D | graph_util_impl.py | 250 output_node_names, argument 279 output_node_names=output_node_names,
|
D | convert_to_constants.py | 854 output_node_names, argument 857 graph_def = graph_util.extract_sub_graph(graph_def, output_node_names) 1125 output_node_names, argument 1163 output_node_names=output_node_names,
|
D | convert_to_constants_test.py | 582 output_node_names=["output_node"])) 623 output_node_names=["output_node", initializer_name], 688 output_node_names=["output_node"])) 729 output_node_names=["output_node"]))
|
/external/tensorflow/tensorflow/core/kernels/ |
D | remote_fused_graph_execute_utils.cc | 204 const std::vector<string>& output_node_names, const bool initialize_by_zero, in DryRunInference() argument 258 status = session->Run(run_options, input_tensors, output_node_names, {}, in DryRunInference() 275 std::vector<string> output_node_names; in DryRunInferenceForAllNode() local 288 output_node_names.emplace_back(strings::StrCat(node->name(), ":", i)); in DryRunInferenceForAllNode() 292 status = DryRunInference(graph_def, input_node_info_list, output_node_names, in DryRunInferenceForAllNode() 299 CHECK_EQ(output_node_names.size(), output_tensors.size()) in DryRunInferenceForAllNode() 300 << output_node_names.size() << ", " << output_tensors.size(); in DryRunInferenceForAllNode() 309 for (int i = 0; static_cast<size_t>(i) < output_node_names.size(); ++i) { in DryRunInferenceForAllNode() 310 const string& name = output_node_names.at(i); in DryRunInferenceForAllNode() 316 const Tensor& tensor = output_tensors.at(output_node_names.size() + i); in DryRunInferenceForAllNode() [all …]
|
D | remote_fused_graph_execute_utils.h | 108 const std::vector<string>& output_node_names, 127 const std::vector<string>& output_node_names,
|
/external/tensorflow/tensorflow/tools/api/golden/v1/ |
D | tensorflow.graph_util.pbtxt | 5 …argspec: "args=[\'sess\', \'input_graph_def\', \'output_node_names\', \'variable_names_whitelist\'…
|
/external/tensorflow/tensorflow/python/compiler/tensorrt/ |
D | trt_convert.py | 549 output_node_names = _gather_names(input_signature_def.inputs).union( 557 output_node_names.add(op.name.split(":")[0]) 563 list(output_node_names))
|
/external/tensorflow/tensorflow/lite/python/ |
D | util.py | 336 output_node_names = [tensor.name.split(":")[0] for tensor in output_tensors] 338 output_node_names)
|
/external/tensorflow/tensorflow/python/compiler/tensorrt/model_tests/ |
D | model_handler.py | 125 output_node_names = [ 131 sess, meta_graph.graph_def, output_node_names))
|
/external/tensorflow/tensorflow/python/compiler/tensorrt/test/ |
D | quantization_mnist_test.py | 141 sess, sess.graph_def, output_node_names=[OUTPUT_NODE_NAME])
|
/external/tensorflow/tensorflow/lite/g3doc/r1/convert/ |
D | python_api.md | 170 --output_node_names=name1,name2.....`
|
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/translate/ |
D | import_model.cc | 2331 absl::flat_hash_set<absl::string_view> output_node_names; in InferMainFunctionType() local 2333 output_node_names.reserve(specs.outputs.size()); in InferMainFunctionType() 2338 output_node_names.insert(remapped_it->second); in InferMainFunctionType() 2341 output_node_names.insert(tensor.node()); in InferMainFunctionType() 2358 if (output_node_names.contains(n->name())) { in InferMainFunctionType()
|
/external/tensorflow/tensorflow/compiler/aot/ |
D | tfcompile.bzl | 144 " --output_node_names=$$(<$(location " + out_nodes_file +
|
/external/tensorflow/tensorflow/lite/micro/examples/person_detection/ |
D | training_a_model.md | 323 --output_node_names=MobilenetV1/Predictions/Reshape_1
|