Home
last modified time | relevance | path

Searched full:hardtanh_ (Results 1 – 25 of 48) sorted by relevance

12

/external/pytorch/aten/src/ATen/native/metal/ops/
DMetalClamp.mm13 static Tensor& hardtanh_(Tensor& input, const Scalar& min_val, const Scalar& max_val) {
62 m.impl(TORCH_SELECTIVE_NAME("aten::hardtanh_"), TORCH_FN(hardtanh_));
/external/pytorch/test/jit/
Dtest_optimize_for_mobile_preserve_debug_info.py238 conv2d_activation=F.hardtanh_,
239 conv2d_activation_kind="aten::hardtanh_",
245 linear_activation=F.hardtanh_,
246 linear_activation_kind="aten::hardtanh_",
/external/pytorch/ios/TestApp/custom_build/
Dmobilenetv2.yaml6 - aten::hardtanh_
/external/pytorch/torch/ao/quantization/pt2e/
Dgraph_utils.py27 {torch.nn.Hardtanh, torch.nn.functional.hardtanh, torch.nn.functional.hardtanh_},
/external/pytorch/aten/src/ATen/native/vulkan/ops/
DClamp.cpp290 Tensor& hardtanh_(Tensor& self, const Scalar& min, const Scalar& max) { in hardtanh_() function
608 m.impl(TORCH_SELECTIVE_NAME("aten::hardtanh_"), hardtanh_); in TORCH_LIBRARY_IMPL()
/external/pytorch/docs/source/
Dnn.functional.rst77 hardtanh_
/external/pytorch/torch/csrc/jit/passes/
Dmetal_rewrite.cpp189 %r = aten::hardtanh_(%r, %output_min, %output_max) in fuseHardtanhWithPackedOps()
197 %res = aten::hardtanh_(%linear_res, %output_min, %output_max) in fuseHardtanhWithPackedOps()
Dxnnpack_rewrite.cpp229 %res = aten::hardtanh_(%linear_res, %output_min, %output_max) in fuseHardtanhWithPackedOps()
239 %res = aten::hardtanh_(%conv2d_res, %output_min, %output_max) in fuseHardtanhWithPackedOps()
/external/pytorch/aten/src/ATen/test/
Dvulkan_api_test.cpp2714 TEST_F(VulkanAPITest, hardtanh_) { in TEST_F() argument
2718 at::hardtanh_(a_cpu, 3, 7); in TEST_F()
2719 at::hardtanh_(a_vulkan, 3, 7); in TEST_F()
7159 hardtanh_, enumerator
7238 class Hardtanh_ final : public BaseOp { class
7240 Hardtanh_() : BaseOp(OpType::hardtanh_) {} in Hardtanh_() function in __anon0035c3ed0211::Hardtanh_
7243 return at::hardtanh_(input, 0, 6); in run()
7247 return "hardtanh_"; in toString()
7301 ops_.emplace_back(new Hardtanh_()); in MobileNetV2()
7303 ops_.emplace_back(new Hardtanh_()); in MobileNetV2()
[all …]
/external/pytorch/torch/ao/ns/fx/
Dmappings.py252 F.hardtanh_,
548 F.hardtanh_,
/external/executorch/backends/arm/quantizer/
Darm_quantizer_utils.py141 torch.ops.aten.hardtanh_.default,
/external/pytorch/torch/csrc/jit/codegen/onednn/
Dinterface.cpp57 aten::hardtanh_, in fuseGraph()
/external/pytorch/test/
Dtest_xnnpack_integration.py1086 "aten::hardtanh_": 2,
1093 M(F.hardtanh_), pattern_count_map, data_shape, prepack_removal=True
1097 pattern_count_map["aten::hardtanh_"] = -1
1099 M(F.hardtanh_),
/external/pytorch/torch/csrc/jit/passes/quantization/
Dhelper.cpp162 "hardtanh_",
169 "hardtanh_",
Dquantization_patterns.h887 auto hardtanh_ = getClampOpFusionInfo("aten::hardtanh_", {"%min", "%max"}); in quant_fusion_pattern_and_replacements() local
1070 std::move(hardtanh_), in quant_fusion_pattern_and_replacements()
/external/pytorch/test/mobile/model_test/
Dtensor_ops.py88 a.hardtanh_(),
Dmodel_ops.yaml176 aten::hardtanh_: 5
Dcoverage.yaml281 - aten::hardtanh_
839 aten::hardtanh_: 3
/external/pytorch/tools/pyi/
Dgen_pyi.py390 "hardtanh_": [
391 "def hardtanh_({}) -> Tensor: ...".format(
516 "hardtanh_",
/external/pytorch/torch/_decomp/
D__init__.py324 aten.hardtanh_,
/external/pytorch/aten/src/ATen/native/
DActivation.cpp457 Tensor& hardtanh_(Tensor& self, const Scalar& min, const Scalar& max) { in hardtanh_() function
534 return at::hardtanh_(self, /*min_val=*/0, /*max_val=*/6); in relu6_()
/external/pytorch/torch/csrc/jit/mobile/model_tracer/
DTracerRunner.cpp38 "aten::hardtanh_",
/external/pytorch/test/quantization/pt2e/
Dtest_x86inductor_quantizer.py629 torch.ops.aten.hardtanh_.default,
634 torch.ops.aten.hardtanh_.default,
1762 torch.ops.aten.hardtanh_.default,
1767 torch.ops.aten.hardtanh_.default,
/external/pytorch/torch/ao/quantization/backend_config/
D_common_operator_config_utils.py642 torch.nn.functional.hardtanh_,
/external/pytorch/torch/_functorch/
Dtop_operators_github_usage.py382 ("nn.functional.hardtanh_", 26),

12