/external/tensorflow/tensorflow/compiler/mlir/lite/tests/ |
D | prepare-tf-fake-quant.mlir | 62 // CHECK: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CONSTANT]]) {qtype = tensor<8x!quant.uniform<u8:f32,… 63 // CHECK: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) 78 // CHECK: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CONSTANT]]) {qtype = tensor<8x!quant.uniform<u8:f32,… 79 // CHECK: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) 92 // CHECK: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CONSTANT]]) {qtype = tensor<8x!quant.uniform<u8:f32,… 93 // CHECK: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) 110 // CHECK: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CONSTANT]]) {qtype = tensor<8x!quant.uniform<u8:f32,… 111 // CHECK: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) 168 // CHECK: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CONSTANT0]]) {qtype = tensor<16x3x3x3x!quant.uniform… 169 // CHECK: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) [all …]
|
D | prepare-quantize-signed.mlir | 195 // CHECK: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CST]]) {qtype = tensor<1x32x42x128x!quant.uniform<i8… 196 // CHECK: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) : (tensor<1x32x42x128x!quant.uniform… 200 // PerTensor: %[[QUANTIZE:.*]] = "tfl.quantize"(%[[CST]]) {qtype = tensor<1x32x42x128x!quant.unifor… 201 // PerTensor: %[[DEQUANTIZE:.*]] = "tfl.dequantize"(%[[QUANTIZE]]) : (tensor<1x32x42x128x!quant.uni…
|
/external/tensorflow/tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/ |
D | tf_to_quant.mlir | 55 // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT]]) : (tensor<8xf32>) -> tensor<8x!quant.… 56 // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]]) 83 // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT0]]) : (tensor<3x3x3x16xf32>) -> tensor<3… 84 // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]]) 102 // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT0]]) : (tensor<3x3x3x16xf32>) -> tensor<3… 105 // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]]) 123 // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT0]]) : (tensor<3x3x3x16xf32>) -> tensor<3… 124 // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]]) 142 // CHECK: %[[QUANTIZE:.*]] = "quantfork.qcast"(%[[CONSTANT0]]) : (tensor<3x3x3x16xf32>) -> tensor<3… 145 // CHECK: %[[DEQUANTIZE:.*]] = "quantfork.dcast"(%[[QUANTIZE]])
|
/external/armnn/src/armnnTfLiteParser/test/ |
D | Quantize.cpp | 21 "operator_codes": [ { "builtin_code": "QUANTIZE" } ], in QuantizeFixture()
|
/external/tensorflow/tensorflow/lite/python/ |
D | util.py | 572 if builtin_code == schema_fb.BuiltinOperator.QUANTIZE: 834 if builtin_code == schema_fb.BuiltinOperator.QUANTIZE: 840 quant_op.builtinCode = schema_fb.BuiltinOperator.QUANTIZE 841 quant_op.deprecatedBuiltinCode = schema_fb.BuiltinOperator.QUANTIZE
|
D | convert_phase.py | 120 QUANTIZE = SubComponentItem("QUANTIZE", Component.OPTIMIZE_TFLITE_MODEL) variable in SubComponent
|
D | convert.py | 198 @convert_phase(Component.OPTIMIZE_TFLITE_MODEL, SubComponent.QUANTIZE)
|
/external/armnn/docs/ |
D | 05_03_delegate.dox | 144 - QUANTIZE
|
D | 05_01_parsers.dox | 164 - QUANTIZE
|
/external/ComputeLibrary/src/core/CL/cl_kernels/common/ |
D | reduction_operation.cl | 170 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, 1); 256 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, VEC_SIZE); 351 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, VEC_SIZE); 452 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, VEC_SIZE);
|
D | generate_proposals_quantized.cl | 85 …(QUANTIZE(shifted_anchor, OFFSET, SCALE, DATA_TYPE, NUM_ROI_FIELDS), 0, (__global DATA_TYPE *)rois…
|
D | bounding_box_transform_quantized.cl | 108 …vstore4(QUANTIZE(pred_box, OFFSET_PRED_BOXES, SCALE_PRED_BOXES, DATA_TYPE, 4), 0, (__global DATA_T…
|
D | roi_pooling_layer.cl | 111 return QUANTIZE(temp, OFFSET_OUT, SCALE_OUT, DATA_TYPE, 1);
|
D | roi_align_layer_quantized.cl | 99 return QUANTIZE(res_f32, OFFSET_OUT, SCALE_OUT, DATA_TYPE, 1);
|
/external/android-nn-driver/ |
D | NnapiSupport.txt | 65 QUANTIZE (FLOAT32 (input only), QUANT8_ASYMM and QUANT8_ASYMM_SIGNED (output on…
|
/external/tensorflow/tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/ |
D | quantization.mlir | 9 // CHECK-NEXT: builtin_code: QUANTIZE
|
/external/ComputeLibrary/src/core/CL/cl_kernels/ |
D | helpers_asymm.h | 389 #define QUANTIZE(input, offset, scale, type, size) QUANTIZE_STR(input, offset, scale, type, size) macro
|
/external/android-nn-driver/1.2/ |
D | HalPolicy.cpp | 138 case V1_2::OperationType::QUANTIZE: in ConvertOperation()
|
/external/android-nn-driver/1.3/ |
D | HalPolicy.cpp | 122 case V1_3::OperationType::QUANTIZE: in ConvertOperation()
|
/external/tensorflow/tensorflow/lite/g3doc/performance/ |
D | quantization_spec.md | 535 QUANTIZE (Requantization)
|
/external/ComputeLibrary/cl_kernels/common/ |
D | reduction_operation.clembed | 2753 #define QUANTIZE(input, offset, scale, type, size) QUANTIZE_STR(input, offset, scale, type, size) 3044 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, 1); 3113 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, VEC_SIZE); 3187 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, VEC_SIZE); 3262 res = QUANTIZE(res_f, OFFSET, SCALE, DATA_TYPE_PROMOTED, VEC_SIZE);
|
D | bounding_box_transform_quantized.clembed | 1509 #define QUANTIZE(input, offset, scale, type, size) QUANTIZE_STR(input, offset, scale, type, size) 1740 …vstore4(QUANTIZE(pred_box, OFFSET_PRED_BOXES, SCALE_PRED_BOXES, DATA_TYPE, 4), 0, (__global DATA_T…
|
D | roi_align_layer_quantized.clembed | 1509 #define QUANTIZE(input, offset, scale, type, size) QUANTIZE_STR(input, offset, scale, type, size) 1754 return QUANTIZE(res_f32, OFFSET_OUT, SCALE_OUT, DATA_TYPE, 1);
|
/external/tensorflow/tensorflow/lite/schema/ |
D | schema_v3a.fbs | 341 QUANTIZE = 114,
|
/external/tensorflow/tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/ |
D | test_schema.fbs | 338 QUANTIZE = 114,
|