/external/tensorflow/tensorflow/lite/toco/graph_transformations/ |
D | fuse_activation_functions.cc | 37 ac_op->type != OperatorType::kRelu) { in Run() 96 } else if (ac_op->type == OperatorType::kRelu) { in Run() 97 op->fused_activation_function = FusedActivationFunctionType::kRelu; in Run()
|
D | identify_prelu.cc | 58 if (relu_input_op == nullptr || relu_input_op->type != OperatorType::kRelu || in Run() 86 FusedActivationFunctionType::kRelu) { in Run() 94 relu_neg_input_op->type != OperatorType::kRelu || in Run()
|
D | remove_trivial_quantized_activation_func.cc | 39 case OperatorType::kRelu: in IsTrivialUnfusedActivationFunc() 69 case FusedActivationFunctionType::kRelu: in IsTrivialFusedActivationFunc()
|
D | propagate_activation_function_into_constants.cc | 36 ac_op->type != OperatorType::kRelu) { in Run() 93 case OperatorType::kRelu: { in Run()
|
D | resolve_constant_unary.cc | 136 case OperatorType::kRelu: in Run() 317 unary_op->type == OperatorType::kRelu) { in Run() 322 case OperatorType::kRelu: { in Run()
|
D | unfuse_activation_functions.cc | 42 case FusedActivationFunctionType::kRelu: in Run()
|
D | reorder_elementwise_unary.cc | 37 case OperatorType::kRelu: in IsElementwiseOperator()
|
D | propagate_fake_quant_num_bits.cc | 113 case OperatorType::kRelu: in DoesOpBlockBackwardPropagation()
|
D | quantize.cc | 63 type == OperatorType::kArgMax || type == OperatorType::kRelu || in SupportsQuantization() 365 op.type == OperatorType::kRelu || op.type == OperatorType::kRelu1 || in ChooseQuantizationForOperatorOutput()
|
D | hardcode_min_max.cc | 410 case OperatorType::kRelu: in Run()
|
/external/tensorflow/tensorflow/lite/toco/tflite/ |
D | types.cc | 215 case FusedActivationFunctionType::kRelu: in Serialize() 232 return FusedActivationFunctionType::kRelu; in Deserialize()
|
D | types_test.cc | 202 {FusedActivationFunctionType::kRelu, in TEST()
|
D | operator_test.cc | 117 CheckSimpleOperator<ReluOperator>("RELU", OperatorType::kRelu); in TEST_F() 436 op.fused_activation_function = FusedActivationFunctionType::kRelu; in TEST_F()
|
/external/tensorflow/tensorflow/core/kernels/neon/ |
D | depthwiseconv_float.h | 560 Ac == FusedActivationFunctionType::kRelu || 670 if (Ac == FusedActivationFunctionType::kRelu) { 693 if (Ac == FusedActivationFunctionType::kRelu) { 708 if (Ac == FusedActivationFunctionType::kRelu) {
|
D | types.h | 23 enum class FusedActivationFunctionType { kNone, kRelu6, kRelu1, kRelu }; enumerator
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | legacy_optimized_ops.h | 416 Ac == FusedActivationFunctionType::kRelu || in FullyConnected() 670 Ac == FusedActivationFunctionType::kRelu || in Conv() 697 Ac == FusedActivationFunctionType::kRelu || in Conv() 750 Ac == FusedActivationFunctionType::kRelu || in ConvAsGemm() 955 Ac == FusedActivationFunctionType::kRelu || in Add() 1024 Ac == FusedActivationFunctionType::kRelu || in BroadcastAdd() 1063 Ac == FusedActivationFunctionType::kRelu || in BroadcastAddFivefold() 1118 Ac == FusedActivationFunctionType::kRelu || in Add() 1272 Ac == FusedActivationFunctionType::kRelu || in AveragePool() 1363 Ac == FusedActivationFunctionType::kRelu || in MaxPool()
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | depthwiseconv_float_test.cc | 86 FusedActivationFunctionType::kRelu, in TryTestOneDepthwiseConv()
|
D | common.h | 61 case FusedActivationFunctionType::kRelu: in GetActivationMinMax()
|
/external/tensorflow/tensorflow/stream_executor/ |
D | dnn.proto | 50 kRelu = 2; enumerator
|
D | dnn.cc | 70 case ActivationMode::kRelu: in ActivationModeString()
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/ |
D | common.h | 63 case FusedActivationFunctionType::kRelu: in GetActivationMinMax()
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | cudnn_fused_conv_rewriter.cc | 233 static_cast<int64>(se::dnn::ActivationMode::kRelu)); in TryRewriteToCudnnForwardRelu()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | legacy_reference_ops.h | 340 Ac == FusedActivationFunctionType::kRelu || in Conv() 505 Ac == FusedActivationFunctionType::kRelu || in FullyConnected() 1175 Ac == FusedActivationFunctionType::kRelu || in Add() 1231 Ac == FusedActivationFunctionType::kRelu || in BroadcastAdd() 1298 Ac == FusedActivationFunctionType::kRelu || in BroadcastAddFivefold() 1352 Ac == FusedActivationFunctionType::kRelu || in Add() 1522 Ac == FusedActivationFunctionType::kRelu || in AveragePool() 1613 Ac == FusedActivationFunctionType::kRelu || in MaxPool()
|
/external/tensorflow/tensorflow/lite/toco/ |
D | model.h | 74 kRelu, enumerator 714 ReluOperator() : Operator(OperatorType::kRelu) {}
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | fully_connected.cc | 285 macro_name(target_namespace, kRelu); \
|