Home
last modified time | relevance | path

Searched refs:kRelu (Results 1 – 25 of 37) sorted by relevance

12

/external/tensorflow/tensorflow/lite/toco/graph_transformations/
Dfuse_activation_functions.cc37 ac_op->type != OperatorType::kRelu) { in Run()
96 } else if (ac_op->type == OperatorType::kRelu) { in Run()
97 op->fused_activation_function = FusedActivationFunctionType::kRelu; in Run()
Didentify_prelu.cc58 if (relu_input_op == nullptr || relu_input_op->type != OperatorType::kRelu || in Run()
86 FusedActivationFunctionType::kRelu) { in Run()
94 relu_neg_input_op->type != OperatorType::kRelu || in Run()
Dremove_trivial_quantized_activation_func.cc39 case OperatorType::kRelu: in IsTrivialUnfusedActivationFunc()
69 case FusedActivationFunctionType::kRelu: in IsTrivialFusedActivationFunc()
Dpropagate_activation_function_into_constants.cc36 ac_op->type != OperatorType::kRelu) { in Run()
93 case OperatorType::kRelu: { in Run()
Dresolve_constant_unary.cc136 case OperatorType::kRelu: in Run()
317 unary_op->type == OperatorType::kRelu) { in Run()
322 case OperatorType::kRelu: { in Run()
Dunfuse_activation_functions.cc42 case FusedActivationFunctionType::kRelu: in Run()
Dreorder_elementwise_unary.cc37 case OperatorType::kRelu: in IsElementwiseOperator()
Dpropagate_fake_quant_num_bits.cc113 case OperatorType::kRelu: in DoesOpBlockBackwardPropagation()
Dquantize.cc63 type == OperatorType::kArgMax || type == OperatorType::kRelu || in SupportsQuantization()
365 op.type == OperatorType::kRelu || op.type == OperatorType::kRelu1 || in ChooseQuantizationForOperatorOutput()
Dhardcode_min_max.cc410 case OperatorType::kRelu: in Run()
/external/tensorflow/tensorflow/lite/toco/tflite/
Dtypes.cc215 case FusedActivationFunctionType::kRelu: in Serialize()
232 return FusedActivationFunctionType::kRelu; in Deserialize()
Dtypes_test.cc202 {FusedActivationFunctionType::kRelu, in TEST()
Doperator_test.cc117 CheckSimpleOperator<ReluOperator>("RELU", OperatorType::kRelu); in TEST_F()
436 op.fused_activation_function = FusedActivationFunctionType::kRelu; in TEST_F()
/external/tensorflow/tensorflow/core/kernels/neon/
Ddepthwiseconv_float.h560 Ac == FusedActivationFunctionType::kRelu ||
670 if (Ac == FusedActivationFunctionType::kRelu) {
693 if (Ac == FusedActivationFunctionType::kRelu) {
708 if (Ac == FusedActivationFunctionType::kRelu) {
Dtypes.h23 enum class FusedActivationFunctionType { kNone, kRelu6, kRelu1, kRelu }; enumerator
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/
Dlegacy_optimized_ops.h416 Ac == FusedActivationFunctionType::kRelu || in FullyConnected()
670 Ac == FusedActivationFunctionType::kRelu || in Conv()
697 Ac == FusedActivationFunctionType::kRelu || in Conv()
750 Ac == FusedActivationFunctionType::kRelu || in ConvAsGemm()
955 Ac == FusedActivationFunctionType::kRelu || in Add()
1024 Ac == FusedActivationFunctionType::kRelu || in BroadcastAdd()
1063 Ac == FusedActivationFunctionType::kRelu || in BroadcastAddFivefold()
1118 Ac == FusedActivationFunctionType::kRelu || in Add()
1272 Ac == FusedActivationFunctionType::kRelu || in AveragePool()
1363 Ac == FusedActivationFunctionType::kRelu || in MaxPool()
/external/tensorflow/tensorflow/lite/kernels/internal/
Ddepthwiseconv_float_test.cc86 FusedActivationFunctionType::kRelu, in TryTestOneDepthwiseConv()
Dcommon.h61 case FusedActivationFunctionType::kRelu: in GetActivationMinMax()
/external/tensorflow/tensorflow/stream_executor/
Ddnn.proto50 kRelu = 2; enumerator
Ddnn.cc70 case ActivationMode::kRelu: in ActivationModeString()
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/
Dcommon.h63 case FusedActivationFunctionType::kRelu: in GetActivationMinMax()
/external/tensorflow/tensorflow/compiler/xla/service/gpu/
Dcudnn_fused_conv_rewriter.cc233 static_cast<int64>(se::dnn::ActivationMode::kRelu)); in TryRewriteToCudnnForwardRelu()
/external/tensorflow/tensorflow/lite/kernels/internal/reference/
Dlegacy_reference_ops.h340 Ac == FusedActivationFunctionType::kRelu || in Conv()
505 Ac == FusedActivationFunctionType::kRelu || in FullyConnected()
1175 Ac == FusedActivationFunctionType::kRelu || in Add()
1231 Ac == FusedActivationFunctionType::kRelu || in BroadcastAdd()
1298 Ac == FusedActivationFunctionType::kRelu || in BroadcastAddFivefold()
1352 Ac == FusedActivationFunctionType::kRelu || in Add()
1522 Ac == FusedActivationFunctionType::kRelu || in AveragePool()
1613 Ac == FusedActivationFunctionType::kRelu || in MaxPool()
/external/tensorflow/tensorflow/lite/toco/
Dmodel.h74 kRelu, enumerator
714 ReluOperator() : Operator(OperatorType::kRelu) {}
/external/tensorflow/tensorflow/lite/kernels/
Dfully_connected.cc285 macro_name(target_namespace, kRelu); \

12