/external/tensorflow/tensorflow/core/kernels/ |
D | quantized_activation_ops_test.cc | 51 Tensor input_quantized = in TEST_F() local 56 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 57 input_quantized.flat<quint8>()); in TEST_F() 82 Tensor input_quantized = in TEST_F() local 87 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 88 input_quantized.flat<quint8>()); in TEST_F()
|
D | quantized_pooling_ops_test.cc | 61 Tensor input_quantized = in TEST_F() local 70 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 71 input_quantized.flat<quint8>()); in TEST_F() 106 Tensor input_quantized = in TEST_F() local 115 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 116 input_quantized.flat<quint8>()); in TEST_F()
|
D | quantized_bias_add_op_test.cc | 58 Tensor input_quantized = in TEST_F() local 73 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 74 input_quantized.flat<quint8>()); in TEST_F() 118 Tensor input_quantized = in TEST_F() local 155 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 156 input_quantized.flat<quint8>()); in TEST_F()
|
D | quantized_batch_norm_op_test.cc | 70 Tensor input_quantized = in TEST_F() local 97 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 98 input_quantized.flat<quint8>()); in TEST_F() 167 Tensor input_quantized = in TEST_F() local 194 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 195 input_quantized.flat<quint8>()); in TEST_F()
|
/external/tensorflow/tensorflow/core/kernels/mkl/ |
D | mkl_quantized_pooling_ops_test.cc | 60 Tensor input_quantized = in TEST_F() local 77 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 78 input_quantized.flat<quint8>()); in TEST_F() 117 Tensor input_quantized = in TEST_F() local 133 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F() 134 input_quantized.flat<quint8>()); in TEST_F()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | basic_rnn.cc | 117 TfLiteTensor* input_quantized; in Prepare() local 119 &input_quantized)); in Prepare() 120 input_quantized->type = input_weights->type; in Prepare() 121 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 122 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 124 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare() 310 TfLiteTensor* input_quantized; in Eval() local 312 GetTemporarySafe(context, node, 0, &input_quantized)); in Eval() 328 input_quantized, hidden_state_quantized, in Eval()
|
D | svdf.cc | 173 TfLiteTensor* input_quantized; in Prepare() local 175 &input_quantized)); in Prepare() 176 input_quantized->type = weights_feature->type; in Prepare() 177 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 178 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 180 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare() 331 TfLiteTensor* input_quantized; in Eval() local 333 &input_quantized)); in Eval() 379 GetTensorData<int8_t>(input_quantized), GetTensorData<float>(state), in Eval()
|
D | unidirectional_sequence_rnn.cc | 124 TfLiteTensor* input_quantized; in Prepare() local 126 &input_quantized)); in Prepare() 127 input_quantized->type = input_weights->type; in Prepare() 128 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 129 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 131 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare() 385 TfLiteTensor* input_quantized; in Eval() local 387 GetTemporarySafe(context, node, 0, &input_quantized)); in Eval() 403 input_quantized, hidden_state_quantized, in Eval()
|
D | fully_connected.cc | 349 TfLiteTensor* input_quantized; in PrepareImpl() local 351 &input_quantized)); in PrepareImpl() 352 input_quantized->type = filter->type; in PrepareImpl() 353 input_quantized->allocation_type = kTfLiteArenaRw; in PrepareImpl() 356 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in PrepareImpl() 515 TfLiteTensor* input_quantized, TfLiteTensor* scaling_factors, in EvalHybridDense() argument 553 int8_t* quant_data = GetTensorData<int8_t>(input_quantized); in EvalHybridDense() 583 int thread_end, TfLiteTensor* input_quantized, in EvalSparseHybridImpl() argument 634 GetTensorData<int8_t>(input_quantized) + thread_start * input_depth; in EvalSparseHybridImpl() 673 TfLiteTensor* input_quantized, TfLiteTensor* scaling_factors, in SparseHybridFullyConnectedTask() [all …]
|
D | bidirectional_sequence_rnn.cc | 202 TfLiteTensor* input_quantized; in Prepare() local 204 &input_quantized)); in Prepare() 205 input_quantized->type = fw_input_weights->type; in Prepare() 206 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 207 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 209 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare() 525 TfLiteTensor* scaling_factors, TfLiteTensor* input_quantized, in EvalHybrid() argument 572 int8_t* quantized_input_ptr = GetTensorData<int8_t>(input_quantized); in EvalHybrid() 792 TfLiteTensor* input_quantized; in Eval() local 795 GetTemporarySafe(context, node, kInputQuantized, &input_quantized)); in Eval() [all …]
|
D | depthwise_conv.cc | 236 TfLiteTensor* input_quantized; in Prepare() local 239 &input_quantized)); in Prepare() 240 input_quantized->type = kTfLiteInt8; in Prepare() 241 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 242 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 244 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare() 465 TfLiteTensor* input_quantized; in EvalHybridPerChannel() local 468 &input_quantized)); in EvalHybridPerChannel() 469 int8_t* quantized_input_ptr_batch = input_quantized->data.int8; in EvalHybridPerChannel()
|
D | batch_matmul.cc | 222 TfLiteTensor* input_quantized; in InitializeTemporaries() local 224 &input_quantized)); in InitializeTemporaries() 225 input_quantized->type = op_context->rhs->type; in InitializeTemporaries() 226 input_quantized->allocation_type = kTfLiteArenaRw; in InitializeTemporaries() 230 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in InitializeTemporaries() 450 TfLiteTensor* input_quantized, in EvalHybrid() argument 478 int8_t* quant_data = GetTensorData<int8_t>(input_quantized); in EvalHybrid() 625 TfLiteTensor* input_quantized; in EvalQuantized() local 627 &input_quantized)); in EvalQuantized() 641 context, node, data, lhs_shape, lhs, rhs_shape, rhs, input_quantized, in EvalQuantized()
|
D | bidirectional_sequence_lstm.cc | 669 TfLiteTensor* input_quantized; in Prepare() local 671 &input_quantized)); in Prepare() 672 input_quantized->type = fw_input_to_output_weights->type; in Prepare() 673 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 674 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 676 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare() 1224 TfLiteTensor* input_quantized; in Eval() local 1227 GetTemporarySafe(context, node, kInputQuantized, &input_quantized)); in Eval() 1297 prod_scaling_factors, recovered_cell_weights, input_quantized, in Eval() 1338 prod_scaling_factors, recovered_cell_weights, input_quantized, in Eval()
|
D | conv.cc | 551 TfLiteTensor* input_quantized; in Prepare() local 554 &input_quantized)); in Prepare() 555 input_quantized->type = kTfLiteInt8; in Prepare() 556 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 557 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 559 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
|
D | unidirectional_sequence_lstm.cc | 992 TfLiteTensor* input_quantized; in Prepare() local 994 &input_quantized)); in Prepare() 995 input_quantized->type = input_to_output_weights->type; in Prepare() 996 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 997 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 999 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
|
D | lstm_eval.h | 164 TfLiteTensor* input_quantized, TfLiteTensor* aux_input_quantized,
|
D | lstm.cc | 1442 TfLiteTensor* input_quantized; in Prepare() local 1444 &input_quantized)); in Prepare() 1445 input_quantized->type = input_to_output_weights->type; in Prepare() 1446 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare() 1447 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare() 1449 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
|
D | lstm_eval.cc | 1970 TfLiteTensor* input_quantized, TfLiteTensor* aux_input_quantized, in EvalHybrid() 2100 GetTensorData<int8_t>(input_quantized), in EvalHybrid() 2199 GetTensorData<int8_t>(input_quantized), in EvalHybrid()
|
/external/pytorch/test/quantization/bc/ |
D | test_backward_compatibility.py | 78 input_quantized=True, argument 100 if input_quantized: 130 input_quantized=True, argument 302 input_quantized=False, 311 input_quantized=False, 533 input_quantized=False,
|
/external/tensorflow/tensorflow/lite/tools/versioning/ |
D | op_signature.h | 64 bool input_quantized; member
|
D | op_signature.cc | 215 op_sig.ext_options.abs.input_quantized = true; in GetOpSignature()
|
D | op_version.cc | 419 return op_sig.ext_options.abs.input_quantized ? 3 : 4; in GetBuiltinOperatorVersion()
|
D | op_version_test.cc | 1017 fake_op_sig.ext_options.abs.input_quantized = true; in TEST()
|