| /external/tensorflow/tensorflow/lite/delegates/xnnpack/ |
| D | depth_to_space_tester.cc | 63 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 73 static_cast<int32_t>(delegate_output_data[index])) in Test() 111 float* delegate_output_data = in Test() local 121 ASSERT_EQ(default_output_data[index], delegate_output_data[index]) in Test()
|
| D | unary_elementwise_tester.cc | 98 float* delegate_output_data = in Test() local 113 ASSERT_EQ(default_output_data[i], delegate_output_data[i]); in Test() 119 default_output_data[i], delegate_output_data[i], in Test()
|
| D | quantized_leaky_relu_tester.cc | 58 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 62 static_cast<int32_t>(delegate_output_data[i])), in Test() 65 << ", delegate " << static_cast<int32_t>(delegate_output_data[i]) in Test()
|
| D | quantized_unary_elementwise_tester.cc | 58 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 62 static_cast<int32_t>(delegate_output_data[i])), in Test() 65 << ", delegate " << static_cast<int32_t>(delegate_output_data[i]) in Test()
|
| D | quantized_reduce_tester.cc | 59 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 64 static_cast<int32_t>(delegate_output_data[i])), in Test() 67 << ", delegate " << static_cast<int32_t>(delegate_output_data[i]) in Test()
|
| D | quantize_tester.cc | 81 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in InvokeAndCheckOutput() local 85 static_cast<int32_t>(delegate_output_data[i])), in InvokeAndCheckOutput() 88 << ", delegate " << static_cast<int32_t>(delegate_output_data[i]) in InvokeAndCheckOutput()
|
| D | leaky_relu_tester.cc | 90 float* delegate_output_data = in Test() local 94 ASSERT_EQ(default_output_data[i], delegate_output_data[i]); in Test()
|
| D | softmax_tester.cc | 90 float* delegate_output_data = in Test() local 94 ASSERT_NEAR(default_output_data[i], delegate_output_data[i], in Test()
|
| D | quantized_conv_2d_tester.cc | 62 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 72 static_cast<int32_t>(delegate_output_data[index])), in Test() 76 << static_cast<int32_t>(delegate_output_data[index]) << ", batch " in Test()
|
| D | quantized_depthwise_conv_2d_tester.cc | 62 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 72 static_cast<int32_t>(delegate_output_data[index])), in Test() 76 << static_cast<int32_t>(delegate_output_data[index]) << ", batch " in Test()
|
| D | reduce_tester.cc | 91 float* delegate_output_data = in Test() local 97 default_output_data[i], delegate_output_data[i], in Test()
|
| D | reshape_tester.cc | 92 float* delegate_output_data = in Test() local 96 ASSERT_EQ(delegate_output_data[i], default_output_data[i]); in Test()
|
| D | dequantize_tester.cc | 59 float* delegate_output_data = in Test() local 63 ASSERT_EQ(default_output_data[i], delegate_output_data[i]) in Test()
|
| D | quantized_binary_elementwise_tester.cc | 103 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 107 static_cast<int32_t>(delegate_output_data[i])), in Test() 110 << ", delegate " << static_cast<int32_t>(delegate_output_data[i]) in Test()
|
| D | transpose_tester.cc | 52 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 55 ASSERT_EQ(default_output_data[i], delegate_output_data[i]); in Test()
|
| D | resize_bilinear_tester.cc | 94 float* delegate_output_data = in Test() local 103 ASSERT_NEAR(default_output_data[index], delegate_output_data[index], in Test()
|
| D | pad_tester.cc | 110 float* delegate_output_data = in Test() local 114 ASSERT_EQ(default_output_data[i], delegate_output_data[i]); in Test()
|
| D | quantized_pad_tester.cc | 76 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 79 ASSERT_EQ(default_output_data[i], delegate_output_data[i]); in Test()
|
| D | quantized_resize_bilinear_tester.cc | 63 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 72 static_cast<int32_t>(delegate_output_data[index])), in Test()
|
| D | quantized_fully_connected_tester.cc | 74 T* delegate_output_data = delegate_interpreter->typed_output_tensor<T>(0); in Test() local 78 static_cast<int32_t>(delegate_output_data[i])), in Test()
|
| D | fully_connected_tester.cc | 110 float* delegate_output_data = in Test() local 114 ASSERT_NEAR(default_output_data[i], delegate_output_data[i], in Test()
|
| D | depthwise_conv_2d_tester.cc | 99 float* delegate_output_data = in Test() local 109 ASSERT_NEAR(default_output_data[index], delegate_output_data[index], in Test()
|
| D | conv_2d_tester.cc | 99 float* delegate_output_data = in Test() local 109 ASSERT_NEAR(default_output_data[index], delegate_output_data[index], in Test()
|