/external/tensorflow/tensorflow/lite/delegates/nnapi/ |
D | nnapi_delegate_errno_test.cc | 56 ActivationFunctionType activation_type, in FloatAddOpModel() argument 59 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in FloatAddOpModel() 75 const TensorData& output, ActivationFunctionType activation_type, in Init() argument 81 CreateAddOptions(builder_, activation_type).Union()); in Init()
|
D | nnapi_delegate_device_selection_test.cc | 59 const TensorData& output, ActivationFunctionType activation_type, in Init() argument 66 CreateAddOptions(builder_, activation_type).Union()); in Init() 373 ActivationFunctionType activation_type, in AddSubOpsAcceleratedModel() argument 381 Init(input1, input2, input3, output, activation_type, in AddSubOpsAcceleratedModel() 401 ActivationFunctionType activation_type, in Init() argument 409 CreateAddOptions(builder_, activation_type).Union(), in Init() 412 CreateSubOptions(builder_, activation_type).Union(), in Init()
|
D | nnapi_delegate_test.cc | 137 ActivationFunctionType activation_type, in FloatAddOpModel() argument 139 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in FloatAddOpModel() 145 ActivationFunctionType activation_type, in FloatAddOpModel() argument 148 Init(input1, input2, output, activation_type, allow_fp32_relax_to_fp16); in FloatAddOpModel() 164 const TensorData& output, ActivationFunctionType activation_type, in Init() argument 170 CreateAddOptions(builder_, activation_type).Union()); in Init() 398 ActivationFunctionType activation_type) { in FloatMulOpModel() argument 403 CreateMulOptions(builder_, activation_type).Union()); in FloatMulOpModel() 1312 ActivationFunctionType activation_type) { in L2NormOpModel() argument 1316 CreateL2NormOptions(builder_, activation_type).Union()); in L2NormOpModel() [all …]
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | l2norm_test.cc | 30 const ActivationFunctionType activation_type) { in L2NormOpModel() argument 45 CreateL2NormOptions(builder_, activation_type).Union()); in L2NormOpModel()
|
D | div_test.cc | 30 ActivationFunctionType activation_type) { in BaseDivOpModel() argument 35 CreateDivOptions(builder_, activation_type).Union()); in BaseDivOpModel()
|
D | mul_test.cc | 30 ActivationFunctionType activation_type) { in BaseMulOpModel() argument 35 CreateMulOptions(builder_, activation_type).Union()); in BaseMulOpModel()
|
D | sub_test.cc | 30 ActivationFunctionType activation_type) { in BaseSubOpModel() argument 35 CreateSubOptions(builder_, activation_type).Union()); in BaseSubOpModel()
|
D | add_test.cc | 30 ActivationFunctionType activation_type) { in BaseAddOpModel() argument 35 CreateAddOptions(builder_, activation_type).Union()); in BaseAddOpModel()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | matmul_op_test.cc | 119 bool transpose_a, bool transpose_b, const string& activation_type, in RunMatMulWithBiasAndActivation() argument 133 if (activation_type == "Relu") { in RunMatMulWithBiasAndActivation() 135 } else if (activation_type == "Relu6") { in RunMatMulWithBiasAndActivation() 137 } else if (activation_type == "Elu") { in RunMatMulWithBiasAndActivation()
|
D | conv_ops_test.cc | 633 const Tensor& bias_data, const string& activation_type, in RunConv2DWithBiasAndActivation() argument 649 if (activation_type == "Relu") { in RunConv2DWithBiasAndActivation() 651 } else if (activation_type == "Relu6") { in RunConv2DWithBiasAndActivation() 653 } else if (activation_type == "Elu") { in RunConv2DWithBiasAndActivation() 695 const string& activation_type, const std::string& padding, in RunConv2DWithBatchNormAndActivation() argument 718 if (activation_type == "Relu") { in RunConv2DWithBatchNormAndActivation() 720 } else if (activation_type == "Relu6") { in RunConv2DWithBatchNormAndActivation() 722 } else if (activation_type == "Elu") { in RunConv2DWithBatchNormAndActivation()
|
D | conv_ops_benchmark_test.cc | 135 int out_depth, const string& activation_type, in Conv2DWithBiasAndActivation() argument 146 TF_CHECK_OK(NodeBuilder(graph->NewName("activation"), activation_type) in Conv2DWithBiasAndActivation() 195 int out_depth, const string& activation_type, in Conv2DWithBatchNormAndActivation() argument 206 TF_CHECK_OK(NodeBuilder(graph->NewName("activation"), activation_type) in Conv2DWithBatchNormAndActivation()
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | types.h | 731 FusedActivationFunctionType activation_type; member
|
/external/tensorflow/tensorflow/compiler/tf2tensorrt/convert/ |
D | convert_nodes.cc | 1594 auto activation_type = static_cast<const nvinfer1::IActivationLayer*>(layer) in IsClipOrRelu() local 1597 return activation_type == nvinfer1::ActivationType::kRELU || in IsClipOrRelu() 1598 activation_type == nvinfer1::ActivationType::kCLIP; in IsClipOrRelu() 1600 return activation_type == nvinfer1::ActivationType::kRELU; in IsClipOrRelu()
|