Home
last modified time | relevance | path

Searched defs:inference_type (Results 1 – 20 of 20) sorted by relevance

/external/tensorflow/tensorflow/compiler/mlir/lite/experimental/tac/common/
Dtargets.h63 inline std::string GetInferenceString(InferenceType inference_type) { in GetInferenceString()
98 auto inference_type = op->getAttrOfType<StringAttr>(kInferenceType); in GetInferenceTypeAnnotation() local
108 InferenceType inference_type; member
134 auto inference_type = GetInferenceTypeAnnotation(op); in GetInferenceDeviceTypeForOp() local
/external/tensorflow/tensorflow/compiler/mlir/lite/experimental/tac/hardwares/
Dgpu_hardware.cc86 InferenceType inference_type = GetInferenceType(op); in IsOpSupported() local
108 InferenceType inference_type = GetInferenceType(op); in IsOpSupported() local
130 InferenceType inference_type = GetInferenceType(op); in IsOpSupported() local
154 InferenceType inference_type = GetInferenceType(op); in IsOpSupported() local
Dcpu_hardware.cc40 inline float InferenceTypeEfficiency(InferenceType inference_type) { in InferenceTypeEfficiency()
/external/tensorflow/tensorflow/compiler/mlir/lite/quantization/lite/
Dquantize_weights.cc80 const tflite::TensorType& inference_type, const StringSet& denylisted_ops, in QuantizeWeights()
192 tflite::TensorType inference_type; in QuantizeWeights() local
216 const tflite::TensorType inference_type = tflite::TensorType_INT8; in QuantizeWeights() local
Dquantize_model.cc53 const tflite::TensorType& inference_type, in QuantizeModel()
/external/tensorflow/tensorflow/compiler/mlir/lite/quantization/
Dquantization_config.cc93 absl::string_view inference_type, in ParseInputNodeQuantSpecs()
134 tensorflow::DataType inference_type, QuantizationSpecs* quant_specs) { in GetInputNodeQuantSpecs()
Dquantization_config.h99 tensorflow::DataType inference_type = tensorflow::DT_FLOAT; member
Dquantization_utils.h391 tensorflow::DataType inference_type = in matchAndRewrite() local
/external/tensorflow/tensorflow/compiler/mlir/tensorflow/translate/
Dtf_mlir_translate_cl.cc58 opt<std::string> inference_type( variable
/external/tensorflow/tensorflow/lite/python/
Dwrap_toco.py37 fully_quantize, inference_type, argument
Dconvert.py202 inference_type=_types_pb2.QUANTIZED_INT8, argument
471 def build_conversion_flags(inference_type=dtypes.float32, argument
Dlite.py761 inference_type=None, argument
/external/tensorflow/tensorflow/python/lite/
Dtoco_python_api_wrapper.cc58 py::object node_blocklist) { in PYBIND11_MODULE()
/external/tensorflow/tensorflow/lite/toco/python/
Dtoco_python_api.cc242 tflite::TensorType FromTocoDataTypeToTflitToTensorType(int inference_type) { in FromTocoDataTypeToTflitToTensorType()
292 bool fully_quantize, int inference_type, in MlirQuantizeModel()
/external/tensorflow/tensorflow/compiler/mlir/lite/experimental/tac/transforms/
Dtarget_annotation.cc85 const InferenceType inference_type = GetInferenceType(op); in SetTargetAnnotation() local
Dget_alternative_subgraph.cc68 InferenceType inference_type, ArrayRef<std::string> devices) { in GetAllAlternativeInferenceDeviceType()
/external/tensorflow/tensorflow/lite/toco/
Dtoco_flags.proto90 optional IODataType inference_type = 4; field
Dtoco_tooling.cc248 const IODataType inference_type = toco_flags.inference_type(); in TransformWithStatus() local
Dargs.h172 Arg<std::string> inference_type; member
/external/tensorflow/tensorflow/compiler/mlir/lite/python/
Dtf_tfl_flatbuffer_helpers.cc216 tensorflow::DataType inference_type = in PopulateQuantizationSpecs() local