Home
last modified time | relevance | path

Searched full:native_layer_norm (Results 1 – 25 of 130) sorted by relevance

123456

/external/executorch/backends/vulkan/runtime/graph/ops/impl/
DNativeLayerNorm.cpp67 VK_THROW("native_layer_norm only supports normalized_shape with dim == 1"); in add_native_layer_norm_node()
71 VK_THROW("native_layer_norm requires weight to be non-None"); in add_native_layer_norm_node()
75 VK_THROW("native_layer_norm requires bias to be non-None"); in add_native_layer_norm_node()
94 std::string kernel_name("native_layer_norm"); in add_native_layer_norm_node()
124 void native_layer_norm(ComputeGraph& graph, const std::vector<ValueRef>& args) { in native_layer_norm() function
130 VK_REGISTER_OP(aten.native_layer_norm.default, native_layer_norm);
/external/executorch/backends/vulkan/runtime/graph/ops/glsl/
Dnative_layer_norm.yaml7 native_layer_norm:
16 - NAME: native_layer_norm
/external/pytorch/aten/src/ATen/native/vulkan/ops/
DLayernorm.cpp13 #include <ATen/ops/native_layer_norm.h>
74 // We invoke native_layer_norm which returns a tuple of tensors: <layer_norm, in run_layernorm_context()
77 at::native_layer_norm(input, normalized_shape, weight_opt, bias_opt, eps); in run_layernorm_context()
DNativeLayerNorm.cpp55 std::tuple<Tensor, Tensor, Tensor> native_layer_norm( in native_layer_norm() function
107 TORCH_SELECTIVE_NAME("aten::native_layer_norm"), in TORCH_LIBRARY_IMPL()
108 TORCH_FN(native_layer_norm)); in TORCH_LIBRARY_IMPL()
/external/executorch/exir/dialects/edge/test/
Dtest_edge_yaml.py157 # Two of three tensor inputs of native_layer_norm are in optional tensor type.
158 ret = gen_op_yaml("native_layer_norm.default")
160 self.assertEqual(ret.func_name, "aten::native_layer_norm")
161 self.assertEqual(ret.inherits, "aten::native_layer_norm")
/external/executorch/backends/arm/_passes/
Ddecompose_layernorm_pass.py18 if op == exir_ops.edge.aten.native_layer_norm.default:
61 exir_ops.edge.aten.native_layer_norm.default,
/external/executorch/kernels/portable/cpu/
Dop_native_layer_norm.cpp94 // native_layer_norm.out(Tensor input, int[] normalized_shape, Tensor? weight,
171 input.scalar_type(), ctx, "native_layer_norm.out", CTYPE, [&]() { in native_layer_norm_out()
/external/pytorch/torch/onnx/
Dsymbolic_opset18.py113 @_onnx_symbolic("aten::native_layer_norm")
124 return opset9.native_layer_norm(g, input, normalized_shape, weight, bias, eps)
/external/executorch/backends/qualcomm/builders/
DREADME.md53 KeyError: 'aten.native_layer_norm.default'
62 if node.op == "call_function" and node.target.__name__ == 'aten.native_layer_norm.default':
188 target = ["aten.native_layer_norm.default"]
Dop_layer_norm.py23 target = ["aten.native_layer_norm.default"]
/external/executorch/kernels/optimized/
Doptimized-oss.yaml68 - op: native_layer_norm.out
Doptimized.yaml80 - op: native_layer_norm.out
/external/pytorch/torch/nn/utils/_expanded_weights/
Dlayer_norm_expanded_weights.py32 torch.native_layer_norm, expanded_args, expanded_kwargs
/external/executorch/backends/apple/mps/operators/
Dnormalization_ops.py66 target = "aten.native_layer_norm.default"
/external/pytorch/aten/src/ATen/native/
Dts_native_functions.yaml91 - native_layer_norm
/external/pytorch/test/inductor/
Dtest_standalone_compile.py107 torch.ops.aten.native_layer_norm.default, (x1, x2, x3, x4, x5), {}
/external/pytorch/test/expect/
DHasDecompTest.test_aten_core_operators.expect369 aten::native_layer_norm
370 aten::native_layer_norm.out
/external/executorch/backends/arm/operator_support/
Dtosa_supported_operators.py91 exir_ops.edge.aten.native_layer_norm.default,
/external/executorch/backends/vulkan/partitioner/
Dsupported_ops.py102 exir_ops.edge.aten.native_layer_norm.default,
/external/executorch/kernels/optimized/cpu/
Dop_native_layer_norm.cpp159 input.scalar_type(), ctx, "native_layer_norm.out", CTYPE, [&]() { in opt_native_layer_norm_out()
/external/pytorch/functorch/op_analysis/
Dpublic_api439 native_layer_norm
/external/executorch/kernels/aten/
Dfunctions.yaml278 - op: native_layer_norm.out
/external/pytorch/aten/src/ATen/test/
Dmath_kernel_test.cpp66 auto out = at::native_layer_norm( in TEST()
/external/pytorch/test/onnx/
Dtest_pytorch_jit_onnx.py128 … Float(2, 3, 2), %12 : Float(2, 1, 1), %13 : Float(2, 1, 1) = aten::native_layer_norm(%x, %7, %w, …
/external/pytorch/benchmarks/dynamo/microbenchmarks/
Doperatorbench.py124 # maybe disable aten.native_layer_norm.default

123456