Home
last modified time | relevance | path

Searched full:native_batch_norm (Results 1 – 25 of 120) sorted by relevance

12345

/external/executorch/backends/vulkan/runtime/graph/ops/impl/
DBatchNorm.cpp29 "native_batch_norm requires ", in check_and_prepack_arg()
101 void native_batch_norm(ComputeGraph& graph, const std::vector<ValueRef>& args) { in native_batch_norm() function
109 aten._native_batch_norm_legit_no_training.default, native_batch_norm);
/external/executorch/exir/tests/
Ddynamic_shape_models.py14 batch_norm_op = torch.ops.aten.native_batch_norm.default
40 if nd.target == torch.ops.aten.native_batch_norm.out
/external/executorch/backends/xnnpack/_passes/
Dfuse_batch_norm_with_conv.py45 bn.target != exir_ops.edge.aten.native_batch_norm.default
64 bn.target == exir_ops.edge.aten.native_batch_norm.default
80 # args[7] for native_batch_norm, but args[6] for
/external/executorch/backends/transforms/
Dfuse_batch_norm_with_conv.py47 bn.target != exir_ops.edge.aten.native_batch_norm.default
66 bn.target == exir_ops.edge.aten.native_batch_norm.default
82 # args[7] for native_batch_norm, but args[6] for
/external/pytorch/test/lazy/
Dtest_reuse_ir.py116 # so we call `torch.ops.aten.native_batch_norm` to bypass the checks.
117 z, _, _ = torch.ops.aten.native_batch_norm(
129 z_lazy, _, _ = torch.ops.aten.native_batch_norm(
/external/pytorch/aten/src/ATen/functorch/
DBatchRulesNorm.cpp848 …OT good. In the ideal world, we do NOT want to convert the new legit op back into native_batch_norm
849 // as native_batch_norm has a problematic schema--it promises it is functional when it is not. Howe…
851 // make native_batch_norm composite implicit within a few weeks and we can fix this before vmap wor…
855 …return at::native_batch_norm(self, weight_opt, bias_opt, running_mean, running_var, train, momentu… in _native_batch_norm_legit_batch()
861 …return at::native_batch_norm(self, weight_opt, bias_opt, Tensor(), Tensor(), train, momentum, eps); in _native_batch_norm_legit_no_stats_batch()
865 VMAP_SUPPORT(native_batch_norm, NATIVE_BATCH_NORM_BATCH_RULE(native_batch_norm)); in TORCH_LIBRARY_IMPL()
/external/pytorch/aten/src/ATen/native/
Dgroup_norm.cpp15 #include <ATen/ops/native_batch_norm.h>
230 auto outputs = at::native_batch_norm( in math_group_norm()
Dts_native_functions.yaml87 - native_batch_norm
Dlayer_norm.cpp18 #include <ATen/ops/native_batch_norm.h>
240 auto outputs = at::native_batch_norm( in math_native_layer_norm()
/external/pytorch/functorch/op_analysis/
Dpublic_api437 native_batch_norm
/external/executorch/kernels/aten/
Dfunctions.yaml274 - op: native_batch_norm.out
/external/pytorch/test/
Dtest_decomp.py194 (torch.bfloat16, torch.ops.aten.native_batch_norm.default): 1e-5,
195 (torch.float16, torch.ops.aten.native_batch_norm.default): 1e-5,
411 # native_batch_norm is only implicit when python dispatcher is on (and noncomposite otherwise)
412 (None, None, "native_batch_norm"),
716 res = torch._decomp.decompositions.native_batch_norm(
Dtest_meta.py738 torch.native_batch_norm: {bf16, f16},
755 torch.native_batch_norm: {f32, f64},
881 aten.native_batch_norm.default: {bf16, f16},
908 aten.native_batch_norm.default: {f32, f64},
916 aten.native_batch_norm.out: {bf16, f16, f32, f64}
/external/pytorch/torch/jit/
D_shape_functions.py1135 def native_batch_norm( function
1435 …"aten::native_batch_norm(Tensor input, Tensor? weight, Tensor? bias, Tensor? running_mean, Tensor?…
1436 native_batch_norm,
1440 native_batch_norm,
1444 native_batch_norm,
/external/pytorch/benchmarks/dynamo/microbenchmarks/operator_inp_logs/timm_train/
Dconvmixer_768_32_training.txt32 Operator: aten.native_batch_norm.default
/external/pytorch/test/quantization/eager/
Dtest_fuse_eager.py92 …with self.assertRaisesRegex(RuntimeError, "Could not run 'aten::native_batch_norm' with arguments …
101 …with self.assertRaisesRegex(RuntimeError, "Could not run 'aten::native_batch_norm' with arguments …
/external/pytorch/torch/csrc/utils/
Dschema_info.cpp286 …{"aten::native_batch_norm(Tensor input, Tensor? weight, Tensor? bias, Tensor? running_mean, Tensor… in getTrainingOps()
288 …{"aten::native_batch_norm.out(Tensor input, Tensor? weight, Tensor? bias, Tensor? running_mean, Te… in getTrainingOps()
/external/pytorch/test/edge/
Dselected_operators.yaml266 aten::native_batch_norm.out:
/external/pytorch/benchmarks/dynamo/microbenchmarks/operator_inp_logs/torchbench_train/
Ddcgan_training.txt29 Operator: aten.native_batch_norm.default
Dmaml_omniglot_training.txt28 Operator: aten.native_batch_norm.default
Dpytorch_CycleGAN_and_pix2pix_training.txt31 Operator: aten.native_batch_norm.default
/external/pytorch/test/functorch/
Dtest_ops.py563 "native_batch_norm"
1037 xfail("native_batch_norm"),
1152 skip("native_batch_norm"),
1314 xfail("native_batch_norm"),
1561 xfail("native_batch_norm"),
1676 xfail("native_batch_norm"),
2060 xfail("native_batch_norm"),
/external/pytorch/torchgen/
Dmodel.py2503 # (native_batch_norm is a good example, although this isn't the case today).
2510 # native_batch_norm (mutable variant)
2511 # native_batch_norm.functional (functional variant)
2516 # native_batch_norm.mutable (mutable variant)
2517 # native_batch_norm (functional variant)
2558 # for now on ops that have a functional + mutable variant (like native_batch_norm).
/external/pytorch/aten/src/ATen/native/cudnn/
DBatchNorm.cpp241 // This keeps a consistent output with native_batch_norm in cudnn_batch_norm()
/external/pytorch/test/expect/
DHasDecompTest.test_aten_core_operators.expect365 aten::native_batch_norm.out

12345