Home
last modified time | relevance | path

Searched refs:input_quantized (Results 1 – 23 of 23) sorted by relevance

/external/tensorflow/tensorflow/core/kernels/
Dquantized_activation_ops_test.cc51 Tensor input_quantized = in TEST_F() local
56 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
57 input_quantized.flat<quint8>()); in TEST_F()
82 Tensor input_quantized = in TEST_F() local
87 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
88 input_quantized.flat<quint8>()); in TEST_F()
Dquantized_pooling_ops_test.cc61 Tensor input_quantized = in TEST_F() local
70 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
71 input_quantized.flat<quint8>()); in TEST_F()
106 Tensor input_quantized = in TEST_F() local
115 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
116 input_quantized.flat<quint8>()); in TEST_F()
Dquantized_bias_add_op_test.cc58 Tensor input_quantized = in TEST_F() local
73 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
74 input_quantized.flat<quint8>()); in TEST_F()
118 Tensor input_quantized = in TEST_F() local
155 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
156 input_quantized.flat<quint8>()); in TEST_F()
Dquantized_batch_norm_op_test.cc70 Tensor input_quantized = in TEST_F() local
97 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
98 input_quantized.flat<quint8>()); in TEST_F()
167 Tensor input_quantized = in TEST_F() local
194 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
195 input_quantized.flat<quint8>()); in TEST_F()
/external/tensorflow/tensorflow/core/kernels/mkl/
Dmkl_quantized_pooling_ops_test.cc60 Tensor input_quantized = in TEST_F() local
77 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
78 input_quantized.flat<quint8>()); in TEST_F()
117 Tensor input_quantized = in TEST_F() local
133 AddInputFromArray<quint8>(input_quantized.shape(), in TEST_F()
134 input_quantized.flat<quint8>()); in TEST_F()
/external/tensorflow/tensorflow/lite/kernels/
Dbasic_rnn.cc117 TfLiteTensor* input_quantized; in Prepare() local
119 &input_quantized)); in Prepare()
120 input_quantized->type = input_weights->type; in Prepare()
121 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
122 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
124 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
310 TfLiteTensor* input_quantized; in Eval() local
312 GetTemporarySafe(context, node, 0, &input_quantized)); in Eval()
328 input_quantized, hidden_state_quantized, in Eval()
Dsvdf.cc173 TfLiteTensor* input_quantized; in Prepare() local
175 &input_quantized)); in Prepare()
176 input_quantized->type = weights_feature->type; in Prepare()
177 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
178 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
180 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
331 TfLiteTensor* input_quantized; in Eval() local
333 &input_quantized)); in Eval()
379 GetTensorData<int8_t>(input_quantized), GetTensorData<float>(state), in Eval()
Dunidirectional_sequence_rnn.cc124 TfLiteTensor* input_quantized; in Prepare() local
126 &input_quantized)); in Prepare()
127 input_quantized->type = input_weights->type; in Prepare()
128 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
129 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
131 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
385 TfLiteTensor* input_quantized; in Eval() local
387 GetTemporarySafe(context, node, 0, &input_quantized)); in Eval()
403 input_quantized, hidden_state_quantized, in Eval()
Dfully_connected.cc349 TfLiteTensor* input_quantized; in PrepareImpl() local
351 &input_quantized)); in PrepareImpl()
352 input_quantized->type = filter->type; in PrepareImpl()
353 input_quantized->allocation_type = kTfLiteArenaRw; in PrepareImpl()
356 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in PrepareImpl()
515 TfLiteTensor* input_quantized, TfLiteTensor* scaling_factors, in EvalHybridDense() argument
553 int8_t* quant_data = GetTensorData<int8_t>(input_quantized); in EvalHybridDense()
583 int thread_end, TfLiteTensor* input_quantized, in EvalSparseHybridImpl() argument
634 GetTensorData<int8_t>(input_quantized) + thread_start * input_depth; in EvalSparseHybridImpl()
673 TfLiteTensor* input_quantized, TfLiteTensor* scaling_factors, in SparseHybridFullyConnectedTask()
[all …]
Dbidirectional_sequence_rnn.cc202 TfLiteTensor* input_quantized; in Prepare() local
204 &input_quantized)); in Prepare()
205 input_quantized->type = fw_input_weights->type; in Prepare()
206 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
207 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
209 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
525 TfLiteTensor* scaling_factors, TfLiteTensor* input_quantized, in EvalHybrid() argument
572 int8_t* quantized_input_ptr = GetTensorData<int8_t>(input_quantized); in EvalHybrid()
792 TfLiteTensor* input_quantized; in Eval() local
795 GetTemporarySafe(context, node, kInputQuantized, &input_quantized)); in Eval()
[all …]
Ddepthwise_conv.cc236 TfLiteTensor* input_quantized; in Prepare() local
239 &input_quantized)); in Prepare()
240 input_quantized->type = kTfLiteInt8; in Prepare()
241 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
242 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
244 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
465 TfLiteTensor* input_quantized; in EvalHybridPerChannel() local
468 &input_quantized)); in EvalHybridPerChannel()
469 int8_t* quantized_input_ptr_batch = input_quantized->data.int8; in EvalHybridPerChannel()
Dbatch_matmul.cc222 TfLiteTensor* input_quantized; in InitializeTemporaries() local
224 &input_quantized)); in InitializeTemporaries()
225 input_quantized->type = op_context->rhs->type; in InitializeTemporaries()
226 input_quantized->allocation_type = kTfLiteArenaRw; in InitializeTemporaries()
230 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in InitializeTemporaries()
450 TfLiteTensor* input_quantized, in EvalHybrid() argument
478 int8_t* quant_data = GetTensorData<int8_t>(input_quantized); in EvalHybrid()
625 TfLiteTensor* input_quantized; in EvalQuantized() local
627 &input_quantized)); in EvalQuantized()
641 context, node, data, lhs_shape, lhs, rhs_shape, rhs, input_quantized, in EvalQuantized()
Dbidirectional_sequence_lstm.cc669 TfLiteTensor* input_quantized; in Prepare() local
671 &input_quantized)); in Prepare()
672 input_quantized->type = fw_input_to_output_weights->type; in Prepare()
673 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
674 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
676 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
1224 TfLiteTensor* input_quantized; in Eval() local
1227 GetTemporarySafe(context, node, kInputQuantized, &input_quantized)); in Eval()
1297 prod_scaling_factors, recovered_cell_weights, input_quantized, in Eval()
1338 prod_scaling_factors, recovered_cell_weights, input_quantized, in Eval()
Dconv.cc551 TfLiteTensor* input_quantized; in Prepare() local
554 &input_quantized)); in Prepare()
555 input_quantized->type = kTfLiteInt8; in Prepare()
556 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
557 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
559 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
Dunidirectional_sequence_lstm.cc992 TfLiteTensor* input_quantized; in Prepare() local
994 &input_quantized)); in Prepare()
995 input_quantized->type = input_to_output_weights->type; in Prepare()
996 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
997 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
999 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
Dlstm_eval.h164 TfLiteTensor* input_quantized, TfLiteTensor* aux_input_quantized,
Dlstm.cc1442 TfLiteTensor* input_quantized; in Prepare() local
1444 &input_quantized)); in Prepare()
1445 input_quantized->type = input_to_output_weights->type; in Prepare()
1446 input_quantized->allocation_type = kTfLiteArenaRw; in Prepare()
1447 if (!TfLiteIntArrayEqual(input_quantized->dims, input->dims)) { in Prepare()
1449 TF_LITE_ENSURE_OK(context, context->ResizeTensor(context, input_quantized, in Prepare()
Dlstm_eval.cc1970 TfLiteTensor* input_quantized, TfLiteTensor* aux_input_quantized, in EvalHybrid()
2100 GetTensorData<int8_t>(input_quantized), in EvalHybrid()
2199 GetTensorData<int8_t>(input_quantized), in EvalHybrid()
/external/pytorch/test/quantization/bc/
Dtest_backward_compatibility.py78 input_quantized=True, argument
100 if input_quantized:
130 input_quantized=True, argument
302 input_quantized=False,
311 input_quantized=False,
533 input_quantized=False,
/external/tensorflow/tensorflow/lite/tools/versioning/
Dop_signature.h64 bool input_quantized; member
Dop_signature.cc215 op_sig.ext_options.abs.input_quantized = true; in GetOpSignature()
Dop_version.cc419 return op_sig.ext_options.abs.input_quantized ? 3 : 4; in GetBuiltinOperatorVersion()
Dop_version_test.cc1017 fake_op_sig.ext_options.abs.input_quantized = true; in TEST()