/external/XNNPACK/tools/ |
D | generate-vbinary-test.py | 58 activation_type = match.group(4) 59 if activation_type is None: 60 activation_type = "LINEAR" 62 activation_type = activation_type.upper() 71 return op_type, activation_type, requantization_type, batch_tile, arch, isa 264 def generate_test_cases(ukernel, op_type, init_fn, activation_type, argument 300 "ACTIVATION_TYPE": activation_type, 355 op_type, activation_type, requantization_type, batch_tile, arch, isa = \ 361 test_case = generate_test_cases(name, op_type, init_fn, activation_type,
|
/external/tensorflow/tensorflow/lite/delegates/nnapi/ |
D | nnapi_delegate_c_api_test.cc | 71 ActivationFunctionType activation_type) { in FloatAddOpModel() argument 72 Init(input1, input2, output, activation_type); in FloatAddOpModel() 78 ActivationFunctionType activation_type) in FloatAddOpModel() argument 80 Init(input1, input2, output, activation_type); in FloatAddOpModel() 96 const TensorData& output, ActivationFunctionType activation_type) { in Init() argument 101 CreateAddOptions(builder_, activation_type).Union()); in Init()
|
D | nnapi_delegate_nnapi_failure_handling_test.cc | 58 ActivationFunctionType activation_type, in AddSubOpsAcceleratedModel() argument 68 Init(input1, input2, input3, output, activation_type, in AddSubOpsAcceleratedModel() 90 ActivationFunctionType activation_type, in Init() argument 98 CreateAddOptions(builder_, activation_type).Union(), in Init() 101 CreateSubOptions(builder_, activation_type).Union(), in Init()
|
D | nnapi_delegate_errno_test.cc | 57 ActivationFunctionType activation_type, in FloatAddOpModel() argument 60 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in FloatAddOpModel() 76 const TensorData& output, ActivationFunctionType activation_type, in Init() argument 82 CreateAddOptions(builder_, activation_type).Union()); in Init()
|
D | nnapi_delegate_device_selection_test.cc | 47 const TensorData& output, ActivationFunctionType activation_type, in Init() argument 57 CreateAddOptions(builder_, activation_type).Union()); in Init() 377 ActivationFunctionType activation_type, in AddSubOpsAcceleratedModel() argument 383 Init(input1, input2, input3, output, activation_type, in AddSubOpsAcceleratedModel() 403 ActivationFunctionType activation_type, in Init() argument 411 CreateAddOptions(builder_, activation_type).Union(), in Init() 414 CreateSubOptions(builder_, activation_type).Union(), in Init() 564 ActivationFunctionType activation_type, in HardSwishAddOpsAcceleratedModel() argument 570 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in HardSwishAddOpsAcceleratedModel() 586 const TensorData& output, ActivationFunctionType activation_type, in Init() argument [all …]
|
D | nnapi_delegate_test.cc | 169 ActivationFunctionType activation_type, in FloatAddOpModel() argument 171 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in FloatAddOpModel() 177 ActivationFunctionType activation_type, in FloatAddOpModel() argument 180 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in FloatAddOpModel() 196 const TensorData& output, ActivationFunctionType activation_type, in Init() argument 202 CreateAddOptions(builder_, activation_type).Union()); in Init() 536 ActivationFunctionType activation_type) { in FloatMulOpModel() argument 541 CreateMulOptions(builder_, activation_type).Union()); in FloatMulOpModel() 1589 ActivationFunctionType activation_type) { in L2NormOpModel() argument 1593 CreateL2NormOptions(builder_, activation_type).Union()); in L2NormOpModel() [all …]
|
/external/tensorflow/tensorflow/lite/tools/delegates/compatibility/nnapi/ |
D | nnapi_compatibility_lib_test.cc | 36 const TensorData& output, ActivationFunctionType activation_type, in AddOpModel() argument 42 CreateAddOptions(builder_, activation_type).Union()); in AddOpModel()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_ops_test.cc | 578 const Tensor& bias_data, const string& activation_type, in RunConv2DWithBiasAndActivation() argument 594 if (activation_type == "Relu") { in RunConv2DWithBiasAndActivation() 596 } else if (activation_type == "Relu6") { in RunConv2DWithBiasAndActivation() 598 } else if (activation_type == "Elu") { in RunConv2DWithBiasAndActivation() 600 } else if (activation_type == "LeakyRelu") { in RunConv2DWithBiasAndActivation() 642 const string& activation_type, const std::string& padding, in RunConv2DWithBatchNormAndActivation() argument 665 if (activation_type == "Relu") { in RunConv2DWithBatchNormAndActivation() 667 } else if (activation_type == "Relu6") { in RunConv2DWithBatchNormAndActivation() 669 } else if (activation_type == "Elu") { in RunConv2DWithBatchNormAndActivation() 671 } else if (activation_type == "LeakyRelu") { in RunConv2DWithBatchNormAndActivation()
|
D | matmul_op_test.cc | 121 bool transpose_a, bool transpose_b, const string& activation_type, in RunMatMulWithBiasAndActivation() argument 135 if (activation_type == "Relu") { in RunMatMulWithBiasAndActivation() 137 } else if (activation_type == "Relu6") { in RunMatMulWithBiasAndActivation() 139 } else if (activation_type == "Elu") { in RunMatMulWithBiasAndActivation() 141 } else if (activation_type == "LeakyRelu") { in RunMatMulWithBiasAndActivation()
|
D | conv_ops_benchmark_test.cc | 146 int out_depth, const string& activation_type, in Conv2DWithBiasAndActivation() argument 157 TF_CHECK_OK(NodeBuilder(graph->NewName("activation"), activation_type) in Conv2DWithBiasAndActivation() 206 int out_depth, const string& activation_type, in Conv2DWithBatchNormAndActivation() argument 217 TF_CHECK_OK(NodeBuilder(graph->NewName("activation"), activation_type) in Conv2DWithBatchNormAndActivation()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | add_test.cc | 35 ActivationFunctionType activation_type) { in BaseAddOpModel() argument 40 CreateAddOptions(builder_, activation_type).Union()); in BaseAddOpModel() 46 ActivationFunctionType activation_type) { in BaseAddOpModel() argument 51 CreateAddOptions(builder_, activation_type).Union()); in BaseAddOpModel() 84 ActivationFunctionType activation_type) in QuantizedAddOpModel() argument 88 activation_type) {} in QuantizedAddOpModel()
|
D | l2norm_test.cc | 35 const ActivationFunctionType activation_type) { in L2NormOpModel() argument 50 CreateL2NormOptions(builder_, activation_type).Union()); in L2NormOpModel()
|
D | pooling3d.cc | 348 #define TF_LITE_AVERAGE_POOL_3D(type, activation_type) \ in AverageEval() argument 350 AveragePool3D<type, activation_type>( \ in AverageEval() 388 #define TF_LITE_MAX_POOL_3D(type, activation_type) \ in MaxEval() argument 390 MaxPool3D<type, activation_type>( \ in MaxEval()
|
D | sub_test.cc | 35 ActivationFunctionType activation_type) { in BaseSubOpModel() argument 40 CreateSubOptions(builder_, activation_type).Union()); in BaseSubOpModel() 79 ActivationFunctionType activation_type) in QuantizedSubOpModel() argument 83 activation_type) {} in QuantizedSubOpModel()
|
D | div_test.cc | 34 ActivationFunctionType activation_type) { in BaseDivOpModel() argument 39 CreateDivOptions(builder_, activation_type).Union()); in BaseDivOpModel()
|
D | mul_test.cc | 36 ActivationFunctionType activation_type) { in BaseMulOpModel() argument 41 CreateMulOptions(builder_, activation_type).Union()); in BaseMulOpModel()
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | types.h | 614 FusedActivationFunctionType activation_type; member
|
/external/tensorflow/tensorflow/lite/tools/optimize/ |
D | quantize_model_test.cc | 72 TensorType GetBiasTensorType(TensorType& activation_type) { in GetBiasTensorType() argument 73 return activation_type == TensorType_INT16 ? TensorType_INT64 in GetBiasTensorType()
|
/external/tensorflow/tensorflow/compiler/xla/stream_executor/cuda/ |
D | cuda_dnn.cc | 3642 dnn::DataType activation_type = GetConvActivationType(input_type); in GetCudnnFusedOperationGraph() local 3723 CreateCudnnTensor(output_dims, output_strides, 'A', activation_type, in GetCudnnFusedOperationGraph() 3728 CreateCudnnTensor(output_dims, output_strides, 'B', activation_type, in GetCudnnFusedOperationGraph() 3743 cudnnDataType_t cudnn_activation_type = ToCudnnDataType(activation_type); in GetCudnnFusedOperationGraph()
|