/external/tensorflow/tensorflow/lite/toco/tflite/ |
D | import.cc | 30 void LoadTensorsTable(const ::tflite::Model& input_model, in LoadTensorsTable() argument 33 auto tensors = (*input_model.subgraphs())[0]->tensors(); in LoadTensorsTable() 40 void LoadOperatorsTable(const ::tflite::Model& input_model, in LoadOperatorsTable() argument 42 auto opcodes = input_model.operator_codes(); in LoadOperatorsTable() 55 void ImportTensors(const ::tflite::Model& input_model, Model* model) { in ImportTensors() argument 56 auto tensors = (*input_model.subgraphs())[0]->tensors(); in ImportTensors() 57 auto* buffers = input_model.buffers(); in ImportTensors() 101 const ::tflite::Model& input_model, in ImportOperators() 106 auto ops = (*input_model.subgraphs())[0]->operators(); in ImportOperators() 169 const ::tflite::Model& input_model, in ImportIOTensors() [all …]
|
D | import.h | 39 void LoadTensorsTable(const ::tflite::Model &input_model, 41 void LoadOperatorsTable(const ::tflite::Model &input_model,
|
D | export.cc | 655 const ::tflite::Model* input_model = ::tflite::GetModel(buffer); in Export() local 665 if (::tflite::optimize::QuantizeWeights(&q_builder, input_model, in Export()
|
/external/tensorflow/tensorflow/lite/tools/ |
D | gen_op_registration_main.cc | 38 void ParseFlagAndInit(int* argc, char** argv, string* input_model, in ParseFlagAndInit() argument 42 Flag(kInputModelFlag, input_model, "path to the tflite model"), in ParseFlagAndInit() 135 void AddOpsFromModel(const string& input_model, in AddOpsFromModel() argument 138 std::ifstream fin(input_model); in AddOpsFromModel() 150 string input_model; in main() local 155 ParseFlagAndInit(&argc, argv, &input_model, &output_registration, in main() 160 if (!input_model.empty()) { in main() 161 AddOpsFromModel(input_model, &builtin_ops, &custom_ops); in main()
|
/external/tensorflow/tensorflow/lite/tools/optimize/ |
D | quantize_model.h | 38 ModelT* input_model, ErrorReporter* error_reporter); 45 ModelT* input_model, const TensorType& input_type, 54 ModelT* input_model, const TensorType& input_type, 63 ModelT* input_model, const TensorType& input_type,
|
D | quantize_weights.h | 40 flatbuffers::FlatBufferBuilder* builder, const Model* input_model, 46 const Model* input_model, 60 const Model* input_model, 71 const Model* input_model,
|
D | quantize_weights.cc | 342 const Model* input_model, in QuantizeWeightsInt8() argument 347 model.reset(input_model->UnPack()); in QuantizeWeightsInt8() 468 const Model* input_model) { in QuantizeWeightsFloat16() argument 470 model.reset(input_model->UnPack()); in QuantizeWeightsFloat16() 550 const Model* input_model, in QuantizeWeights() argument 556 return QuantizeWeightsInt8(builder, input_model, use_hybrid_evaluation, in QuantizeWeights() 562 const Model* input_model, in QuantizeWeights() argument 565 return QuantizeWeightsInt8(builder, input_model, true, in QuantizeWeights() 570 const Model* input_model, BufferType quant_type) { in QuantizeWeights() argument 576 return QuantizeWeightsInt8(builder, input_model, true, in QuantizeWeights() [all …]
|
D | quantization_wrapper_utils.cc | 57 auto input_model = FlatBufferModel::BuildFromFile(path.c_str()); in LoadModel() local 58 if (!input_model) { in LoadModel() 61 auto readonly_model = input_model->GetModel(); in LoadModel()
|
/external/tensorflow/tensorflow/compiler/mlir/lite/sparsity/ |
D | sparsify_model.cc | 35 TfLiteStatus SparsifyModel(const tflite::ModelT& input_model, in SparsifyModel() argument 45 tflite::Model::Pack(input_builder, &input_model); in SparsifyModel()
|
D | sparsify_model.h | 29 TfLiteStatus SparsifyModel(const tflite::ModelT& input_model,
|
/external/tensorflow/tensorflow/compiler/mlir/lite/quantization/lite/ |
D | quantize_model.cc | 39 const tflite::ModelT& input_model, const tflite::TensorType& input_type, in QuantizeModel() argument 58 tflite::Model::Pack(input_builder, &input_model); in QuantizeModel()
|
D | quantize_model.h | 32 const tflite::ModelT& input_model, const tflite::TensorType& input_type,
|
/external/tensorflow/tensorflow/compiler/mlir/lite/ |
D | tf_to_tfl_flatbuffer.cc | 136 const ::tflite::Model* input_model = ::tflite::GetModel(buffer); in ConvertTFExecutorToTFLOrFlatbuffer() local 146 if (::tflite::optimize::QuantizeWeights(&q_builder, input_model, in ConvertTFExecutorToTFLOrFlatbuffer()
|
/external/tensorflow/tensorflow/lite/tools/optimize/g3doc/ |
D | quantize_weights.md | 64 ::tflite::Model* input_model = ...; 66 TfLiteStatus status = ::tflite::optimize::QuantizeWeights(&builder, input_model);
|