Home
last modified time | relevance | path

Searched full:selu (Results 1 – 25 of 124) sorted by relevance

12345

/external/pytorch/test/onnx/
Dtest_onnxscript_no_runtime.py27 # 1. Register Selu onnxscript function as custom Op
31 def Selu(X): function
43 return g.onnxscript_op(Selu, X).setType(X.type())
46 symbolic_name="aten::selu",
88 model_selu = torch.nn.SELU()
106 self.assertEqual(selu_proto.functions[0].name, "Selu")
115 self.selu = torch.nn.SELU()
122 y = self.selu(x)
135 def Selu(X): function
150 return g.onnxscript_op(Selu, X).setType(X.type())
[all …]
Dtest_onnxscript_runtime.py24 model = torch.nn.SELU()
31 def Selu( function
45 return g.onnxscript_op(Selu, X).setType(X.type())
48 symbolic_name="aten::selu",
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_SeluGrad.pbtxt7 The backpropagated gradients to the corresponding Selu operation.
13 The outputs of the corresponding Selu operation.
23 summary: "Computes gradients for the scaled exponential linear (Selu) operation."
Dapi_def_Selu.pbtxt2 graph_op_name: "Selu"
/external/tensorflow/tensorflow/core/kernels/mlir_generated/
Dgpu_op_selu.cc21 GENERATE_AND_REGISTER_UNARY_GPU_KERNEL(Selu, DT_HALF);
22 GENERATE_AND_REGISTER_UNARY_GPU_KERNEL(Selu, DT_FLOAT);
23 GENERATE_AND_REGISTER_UNARY_GPU_KERNEL(Selu, DT_DOUBLE);
/external/tensorflow/tensorflow/core/api_def/python_api/
Dapi_def_Selu.pbtxt2 graph_op_name: "Selu"
4 name: "nn.selu"
/external/tensorflow/tensorflow/core/api_def/java_api/
Dapi_def_Selu.pbtxt2 graph_op_name: "Selu"
4 name: "nn.Selu"
/external/tensorflow/tensorflow/python/ops/
Dnn_grad_test.py228 selu = gen_nn_ops.selu(inputs)
229 selu_grad = gradients_impl.gradients(selu, inputs, grad_ys=dummy)[0]
245 selu = gen_nn_ops.selu(inputs)
246 selu_grad = gradients_impl.gradients(selu, inputs, grad_ys=dummy)[0]
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v1/
DSelu.pbtxt2 name: "Selu"
24 name: "Selu"
/external/tensorflow/tensorflow/core/ops/compat/ops_history_v2/
DSelu.pbtxt2 name: "Selu"
24 name: "Selu"
/external/tensorflow/tensorflow/core/kernels/
Drelu_op_functor.h174 struct Selu { struct
175 // Computes Selu activation.
198 // gradients: gradients backpropagated to the Selu op. argument
199 // activations: outputs of the Selu op.
200 // backprops: gradients to backpropagate to the Selu inputs.
Drelu_op.cc75 Name("Selu").Device(DEVICE_CPU).TypeConstraint<type>("T"), \
81 // Elu and Selu only make sense with float or double.
104 void Selu<GPUDevice, T>::operator()( \
107 extern template struct Selu<GPUDevice, T>;
120 Name("Selu").Device(DEVICE_GPU).TypeConstraint<type>("T"), \
Dunary_ops_composition.cc171 // Register compute function for the Relu/Relu6/Elu/Selu.
204 auto selu = functor::Selu<Eigen::DefaultDevice, T>(); \
205 selu(Eigen::DefaultDevice(), in, *out); \
255 REGISTER_COMPUTE_FN(Selu); in UnaryOpsCompositionSupport()
320 REGISTER_COMPUTE_FN(Selu); in UnaryOpsCompositionSupport()
384 REGISTER_COMPUTE_FN(Selu); in UnaryOpsCompositionSupport()
/external/tensorflow/tensorflow/python/keras/
Dactivations.py146 @keras_export('keras.activations.selu')
148 def selu(x): function
149 """Scaled Exponential Linear Unit (SELU).
151 The Scaled Exponential Linear Unit (SELU) activation function is defined as:
159 Basically, the SELU activation function multiplies `scale` (> 1) with the
175 ... activation='selu'))
177 ... activation='selu'))
179 ... activation='selu'))
197 return nn.selu(x)
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/
Delu_op.cc35 XlaOp Selu(XlaOp x) { in Selu() function
82 ctx->SetOutput(0, xla::Selu(ctx->Input(0))); in Compile()
107 REGISTER_XLA_OP(Name("Selu"), SeluOp);
Delu_op.h23 XlaOp Selu(XlaOp x);
/external/pytorch/torch/nn/modules/
D__init__.py21 SELU,
300 "SELU",
/external/pytorch/torch/_refs/nn/functional/
D__init__.py57 "selu",
119 # alpha = - SELU.alpha * SELU.scale, here
120 # SELU.alpha = 1.6732632423543772848170429916717 and
121 # SELU.scale = 1.0507009873554804934193349852946
400 @register_decomposition(aten.selu)
407 def selu(a: TensorLikeType, inplace: bool = False) -> TensorLikeType: function
409 Reference implementation of torch.nn.functional.selu
1278 selu_ = _make_inplace(selu)
/external/tensorflow/tensorflow/core/kernels/mlir_generated/op_definitions/
Dselu.mlir.tmpl3 %0 = "tf.Selu"(%arg0) : (tensor<*xelem_type>) -> tensor<*xoutput_type>
/external/pytorch/test/onnx/expect/
DTestOperators.test_selu.expect9 op_type: "Selu"
/external/pytorch/functorch/dim/
Dop_properties.py271 torch.selu,
272 torch.nn.functional.selu,
/external/pytorch/torch/ao/pruning/_experimental/pruner/
Dbase_structured_sparsifier.py53 F.selu,
80 nn.SELU,
/external/pytorch/torch/csrc/api/include/torch/nn/modules/
Dactivation.h49 // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ SELU ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
51 /// Applies the selu function element-wise.
52 /// See https://pytorch.org/docs/main/nn.html#torch.nn.SELU to learn
60 /// SELU model(SELUOptions().inplace(true));
70 /// Pretty prints the `SELU` module into the given `stream`.
79 /// provides, and examples of how to use `SELU` with `torch::nn::SELUOptions`.
82 TORCH_MODULE(SELU);
/external/tensorflow/tensorflow/python/kernel_tests/nn_ops/
Drelu_op_test.py519 tf_selu = nn_ops.selu(np_features)
537 nn_ops.selu, [x], delta=1.0 / 1024))
545 *gradient_checker_v2.compute_gradient(nn_ops.selu, [x]))
555 y = nn_ops.selu(x)
573 y = nn_ops.selu(x)
/external/pytorch/torch/csrc/jit/passes/
Drestore_mutation.h28 {aten::selu, false},

12345