/external/tensorflow/tensorflow/lite/toco/graph_transformations/ |
D | fuse_activation_functions.cc | 35 if (ac_op->type != OperatorType::kRelu6 && in Run() 92 if (ac_op->type == OperatorType::kRelu6) { in Run() 93 op->fused_activation_function = FusedActivationFunctionType::kRelu6; in Run()
|
D | remove_trivial_quantized_activation_func.cc | 47 case OperatorType::kRelu6: in IsTrivialUnfusedActivationFunc() 77 case FusedActivationFunctionType::kRelu6: in IsTrivialFusedActivationFunc()
|
D | propagate_activation_function_into_constants.cc | 34 if (ac_op->type != OperatorType::kRelu6 && in Run() 104 case OperatorType::kRelu6: { in Run()
|
D | resolve_constant_unary.cc | 134 case OperatorType::kRelu6: in Run() 315 } else if (unary_op->type == OperatorType::kRelu6 || in Run() 333 case OperatorType::kRelu6: { in Run()
|
D | unfuse_activation_functions.cc | 45 case FusedActivationFunctionType::kRelu6: in Run()
|
D | reorder_elementwise_unary.cc | 39 case OperatorType::kRelu6: in IsElementwiseOperator()
|
D | propagate_fake_quant_num_bits.cc | 115 case OperatorType::kRelu6: in DoesOpBlockBackwardPropagation()
|
D | quantize.cc | 64 type == OperatorType::kRelu1 || type == OperatorType::kRelu6 || in SupportsQuantization() 366 op.type == OperatorType::kRelu6 || op.type == OperatorType::kPRelu) { in ChooseQuantizationForOperatorOutput()
|
/external/tensorflow/tensorflow/lite/toco/tflite/ |
D | operator_test.cc | 119 CheckSimpleOperator<Relu6Operator>("RELU6", OperatorType::kRelu6); in TEST_F() 163 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F() 231 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F() 349 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F() 371 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F() 387 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F() 400 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F() 409 op.fused_activation_function = FusedActivationFunctionType::kRelu6; in TEST_F()
|
D | types.cc | 217 case FusedActivationFunctionType::kRelu6: in Serialize() 234 return FusedActivationFunctionType::kRelu6; in Deserialize()
|
D | types_test.cc | 204 {FusedActivationFunctionType::kRelu6, in TEST()
|
/external/tensorflow/tensorflow/core/kernels/neon/ |
D | depthwiseconv_float.h | 561 Ac == FusedActivationFunctionType::kRelu6 || 674 } else if (Ac == FusedActivationFunctionType::kRelu6) { 695 } else if (Ac == FusedActivationFunctionType::kRelu6) { 710 } else if (Ac == FusedActivationFunctionType::kRelu6) {
|
D | types.h | 23 enum class FusedActivationFunctionType { kNone, kRelu6, kRelu1, kRelu }; enumerator
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
D | legacy_optimized_ops.h | 417 Ac == FusedActivationFunctionType::kRelu6 || in FullyConnected() 671 Ac == FusedActivationFunctionType::kRelu6 || in Conv() 698 Ac == FusedActivationFunctionType::kRelu6 || in Conv() 751 Ac == FusedActivationFunctionType::kRelu6 || in ConvAsGemm() 956 Ac == FusedActivationFunctionType::kRelu6 || in Add() 1025 Ac == FusedActivationFunctionType::kRelu6 || in BroadcastAdd() 1064 Ac == FusedActivationFunctionType::kRelu6 || in BroadcastAddFivefold() 1119 Ac == FusedActivationFunctionType::kRelu6 || in Add() 1273 Ac == FusedActivationFunctionType::kRelu6 || in AveragePool() 1364 Ac == FusedActivationFunctionType::kRelu6 || in MaxPool()
|
/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | depthwiseconv_float_test.cc | 88 FusedActivationFunctionType::kRelu6})); in TryTestOneDepthwiseConv()
|
D | common.h | 69 case FusedActivationFunctionType::kRelu6: in GetActivationMinMax()
|
D | types.h | 25 enum class FusedActivationFunctionType : uint8 { kNone, kRelu6, kRelu1, kRelu }; enumerator
|
/external/tensorflow/tensorflow/stream_executor/ |
D | dnn.proto | 52 kRelu6 = 3; enumerator
|
D | dnn.cc | 72 case ActivationMode::kRelu6: in ActivationModeString()
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/ |
D | common.h | 71 case FusedActivationFunctionType::kRelu6: in GetActivationMinMax()
|
D | types.h | 31 enum class FusedActivationFunctionType : uint8 { kNone, kRelu6, kRelu1, kRelu }; enumerator
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | legacy_reference_ops.h | 341 Ac == FusedActivationFunctionType::kRelu6 || in Conv() 506 Ac == FusedActivationFunctionType::kRelu6 || in FullyConnected() 1176 Ac == FusedActivationFunctionType::kRelu6 || in Add() 1232 Ac == FusedActivationFunctionType::kRelu6 || in BroadcastAdd() 1299 Ac == FusedActivationFunctionType::kRelu6 || in BroadcastAddFivefold() 1353 Ac == FusedActivationFunctionType::kRelu6 || in Add() 1523 Ac == FusedActivationFunctionType::kRelu6 || in AveragePool() 1614 Ac == FusedActivationFunctionType::kRelu6 || in MaxPool()
|
/external/tensorflow/tensorflow/lite/toco/ |
D | model.h | 76 kRelu6, enumerator 737 Relu6Operator() : Operator(OperatorType::kRelu6) {}
|
D | dump_graphviz.cc | 361 case FusedActivationFunctionType::kRelu6: in GetOpAttributes()
|
/external/tensorflow/tensorflow/lite/kernels/ |
D | fully_connected.cc | 288 macro_name(target_namespace, kRelu6); \
|