Home
last modified time | relevance | path

Searched refs:kRelu (Results 1 – 24 of 24) sorted by relevance

/external/tensorflow/tensorflow/contrib/lite/toco/graph_transformations/
Dfuse_activation_functions.cc34 ac_op->type != OperatorType::kRelu) { in Run()
84 } else if (ac_op->type == OperatorType::kRelu) { in Run()
85 op->fused_activation_function = FusedActivationFunctionType::kRelu; in Run()
Dremove_trivial_quantized_activation_func.cc32 if (op->fused_activation_function != FusedActivationFunctionType::kRelu && in Run()
48 if (op->fused_activation_function == FusedActivationFunctionType::kRelu || in Run()
Dunfuse_activation_functions.cc39 case FusedActivationFunctionType::kRelu: in Run()
Dreorder_activation_functions.cc35 ac_op->type != OperatorType::kRelu) { in Run()
Dpropagate_fixed_sizes.cc1349 case OperatorType::kRelu: in Run()
/external/tensorflow/tensorflow/contrib/lite/toco/tflite/
Dtypes.cc150 case FusedActivationFunctionType::kRelu: in Serialize()
167 return FusedActivationFunctionType::kRelu; in Deserialize()
Doperator_test.cc104 CheckSimpleOperator<ReluOperator>("RELU", OperatorType::kRelu); in TEST_F()
343 op.fused_activation_function = FusedActivationFunctionType::kRelu; in TEST_F()
Dtypes_test.cc171 {FusedActivationFunctionType::kRelu, in TEST()
Doperator.cc852 new SimpleOperator<ReluOperator>("RELU", OperatorType::kRelu)); in BuildOperatorList()
/external/tensorflow/tensorflow/core/kernels/neon/
Ddepthwiseconv_float.h560 Ac == FusedActivationFunctionType::kRelu ||
670 if (Ac == FusedActivationFunctionType::kRelu) {
693 if (Ac == FusedActivationFunctionType::kRelu) {
708 if (Ac == FusedActivationFunctionType::kRelu) {
Dtypes.h23 enum class FusedActivationFunctionType { kNone, kRelu6, kRelu1, kRelu }; enumerator
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/
Dcommon.h63 case FusedActivationFunctionType::kRelu: in GetActivationMinMax()
Dtypes.h22 enum class FusedActivationFunctionType : uint8 { kNone, kRelu6, kRelu1, kRelu }; enumerator
/external/tensorflow/tensorflow/contrib/fused_conv/kernels/
Dfused_conv2d_bias_activation_op.cc561 dnn::ActivationMode::kRelu, output_desc, &output_ptr, in launch()
599 bias_desc, bias_ptr, dnn::ActivationMode::kRelu, output_desc, in launch()
/external/tensorflow/tensorflow/contrib/lite/toco/
Dmodel.h63 kRelu, enumerator
534 ReluOperator() : Operator(OperatorType::kRelu) {}
Ddump_graphviz.cc230 case FusedActivationFunctionType::kRelu: in GetPropertiesForOperator()
Dexport_tensorflow.cc1630 } else if (src_op.type == OperatorType::kRelu) { in ConvertOperator()
Dimport_tensorflow.cc1509 op->fused_activation_function = FusedActivationFunctionType::kRelu; in ConvertSvdfOperator()
/external/tensorflow/tensorflow/contrib/lite/kernels/
Dfully_connected.cc171 macro_name(target_namespace, kRelu); \
/external/tensorflow/tensorflow/stream_executor/
Ddnn.cc62 case ActivationMode::kRelu: in ActivationModeString()
Ddnn.h858 kRelu,
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/
Dreference_ops.h330 Ac == FusedActivationFunctionType::kRelu || in Conv()
548 Ac == FusedActivationFunctionType::kRelu || in FullyConnected()
839 Ac == FusedActivationFunctionType::kRelu || in Add()
1013 Ac == FusedActivationFunctionType::kRelu || in BroadcastAdd()
1815 Ac == FusedActivationFunctionType::kRelu || in AveragePool()
2038 Ac == FusedActivationFunctionType::kRelu || in MaxPool()
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/
Doptimized_ops.h625 Ac == FusedActivationFunctionType::kRelu || in FullyConnected()
922 Ac == FusedActivationFunctionType::kRelu || in Conv()
949 Ac == FusedActivationFunctionType::kRelu || in Conv()
1034 Ac == FusedActivationFunctionType::kRelu || in ConvAsGemm()
1412 Ac == FusedActivationFunctionType::kRelu || in Add()
1662 Ac == FusedActivationFunctionType::kRelu || in BroadcastAdd()
2845 Ac == FusedActivationFunctionType::kRelu || in AveragePool()
3044 Ac == FusedActivationFunctionType::kRelu || in MaxPool()
/external/tensorflow/tensorflow/stream_executor/cuda/
Dcuda_dnn.cc837 case dnn::ActivationMode::kRelu: in ScopedActivationDescriptor()
2554 if (activation_mode != dnn::ActivationMode::kRelu) { in DoFusedConvolveImpl()
3915 case dnn::ActivationMode::kRelu: in DoActivate()