Home
last modified time | relevance | path

Searched full:hardsigmoid (Results 1 – 25 of 115) sorted by relevance

12345

/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/
Dhardsigmoid.c42 … "failed to create Hardsigmoid operator with %zu channels: number of channels must be non-zero", in pytorch_qnnp_create_hardsigmoid_nc_q8()
49 … "failed to create Hardsigmoid operator with %.7g input scale: scale must be finite and positive", in pytorch_qnnp_create_hardsigmoid_nc_q8()
56 … "failed to create Hardsigmoid operator with %.7g output scale: scale must be finite and positive", in pytorch_qnnp_create_hardsigmoid_nc_q8()
63 "failed to create Hardsigmoid operator with [%" PRIu8 ", %" PRIu8 in pytorch_qnnp_create_hardsigmoid_nc_q8()
74 …"failed to create Hardsigmoid operator with %.7g output scale: only output scale of 1/256 is suppo… in pytorch_qnnp_create_hardsigmoid_nc_q8()
81 "failed to create Hardsigmoid operator with %" PRIu8 in pytorch_qnnp_create_hardsigmoid_nc_q8()
100 "failed to allocate 256 bytes for Hardsigmoid lookup table"); in pytorch_qnnp_create_hardsigmoid_nc_q8()
111 // hardsigmoid, no min/max functions in C in pytorch_qnnp_create_hardsigmoid_nc_q8()
140 pytorch_qnnp_operator_t hardsigmoid, in pytorch_qnnp_setup_hardsigmoid_nc_q8() argument
153 hardsigmoid->batch_size = 0; in pytorch_qnnp_setup_hardsigmoid_nc_q8()
[all …]
/external/executorch/backends/vulkan/runtime/graph/ops/glsl/
Dactivations.h34 float hardsigmoid(float x) { in hardsigmoid() function
38 vec4 hardsigmoid(vec4 tex) { in hardsigmoid() function
40 hardsigmoid(tex.x), in hardsigmoid()
41 hardsigmoid(tex.y), in hardsigmoid()
42 hardsigmoid(tex.z), in hardsigmoid()
43 hardsigmoid(tex.w)); in hardsigmoid()
Dunary_op.yaml43 - NAME: hardsigmoid
44 OPERATOR: hardsigmoid(X)
/external/pytorch/benchmarks/operator_benchmark/pt/
Dhardsigmoid_test.py8 Microbenchmarks for the hardsigmoid operator.
12 # Configs for hardsigmoid ops
34 ["Hardsigmoid", nn.Hardsigmoid],
Dqactivation_test.py49 ("functional.hardsigmoid", qF.hardsigmoid),
/external/pytorch/aten/src/ATen/native/metal/
DMetalNeuronType.h16 HardSigmoid, enumerator
45 } else if (type == NeuronType::HardSigmoid) { in neuron()
46 return [MPSCNNNeuronOp hardSigmoid]; in neuron()
60 } else if (type == NeuronType::HardSigmoid) { in neuronDescriptor()
/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/bench/
Dhardsigmoid.cc49 state.SkipWithError("failed to create Hardsigmoid operator"); in hardsigmoid_q8()
60 state.SkipWithError("failed to setup Hardsigmoid operator"); in hardsigmoid_q8()
67 state.SkipWithError("failed to run Hardsigmoid operator"); in hardsigmoid_q8()
81 state.SkipWithError("failed to delete Hardsigmoid operator"); in hardsigmoid_q8()
/external/pytorch/aten/src/ATen/native/quantized/cpu/
Dqhardsigmoid.cpp59 "failed to create QNNPACK Hardsigmoid operator"); in qnnpack_hardsigmoid()
75 "failed to setup QNNPACK Hardsigmoid operator"); in qnnpack_hardsigmoid()
84 "failed to run QNNPACK Hardsigmoid operator"); in qnnpack_hardsigmoid()
103 // Note: we create a new temporary tensor because the output of hardsigmoid in hardsigmoid_out_quantized_cpu()
/external/executorch/backends/qualcomm/builders/
Dop_hardsigmoid.py21 target = ["aten.hardsigmoid.default"]
58 # The operation enum of hardsigmoid in QNN
65 # The parameter used in Pytorch definition for hardsigmoid
/external/pytorch/torch/ao/ns/fx/
Dmappings.py241 # F.hardsigmoid
244 "hardsigmoid",
245 F.hardsigmoid,
246 nn.Hardsigmoid,
541 F.hardsigmoid,
679 nn.Hardsigmoid,
703 "hardsigmoid",
/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/
DCMakeLists.txt164 src/hardsigmoid.c
548 add_executable(hardsigmoid-test test/hardsigmoid.cc)
549 set_target_properties(hardsigmoid-test PROPERTIES
553 target_include_directories(hardsigmoid-test PRIVATE src test)
554 target_link_libraries(hardsigmoid-test PRIVATE pytorch_qnnpack cpuinfo gtest gtest_main)
555 add_test(hardsigmoid-test hardsigmoid-test)
803 add_executable(hardsigmoid-bench bench/hardsigmoid.cc)
804 set_target_properties(hardsigmoid-bench PROPERTIES
808 target_link_libraries(hardsigmoid-bench PRIVATE pytorch_qnnpack benchmark)
Dconfigure.py89 build.cc("hardsigmoid.c"),
225 build.unittest("hardsigmoid-test", build.cxx("hardsigmoid.cc"))
264 build.benchmark("hardsigmoid-bench", build.cxx("hardsigmoid.cc"))
Dbuckbuild.bzl271 "src/hardsigmoid.c",
579 "test/hardsigmoid.cc",
617 "hardsigmoid-operator-tester.h": "test/hardsigmoid-operator-tester.h",
/external/executorch/backends/vulkan/runtime/graph/ops/impl/
DUnaryOp.cpp139 DEFINE_ACTIVATION_FN(hardsigmoid);
157 VK_REGISTER_OP(aten.hardsigmoid.default, hardsigmoid);
/external/pytorch/torch/nn/modules/
D__init__.py9 Hardsigmoid,
226 "Hardsigmoid",
/external/pytorch/functorch/op_analysis/
Dpublic_api104 nn.functional.hardsigmoid
474 nn.functional.hardsigmoid
/external/pytorch/torch/ao/quantization/
Dqconfig_mapping.py45 torch.nn.Hardsigmoid: default_fixed_qparams_range_0to1_observer,
46 torch.nn.functional.hardsigmoid: default_fixed_qparams_range_0to1_observer,
47 "hardsigmoid": default_fixed_qparams_range_0to1_observer,
/external/pytorch/test/cpp/jit/
Dtest_subgraph_utils.cpp64 %q5 : Tensor = aten::hardsigmoid(%q4) in TEST()
104 ->check("aten::hardsigmoid") in TEST()
/external/executorch/backends/xnnpack/partition/
Dconfigs.py72 torch.nn.Hardsigmoid, # we can handle decomposition
120 torch.nn.Hardsigmoid,
/external/pytorch/aten/src/ATen/native/metal/mpscnn/
DMPSCNNNeuronOp.h5 + (MPSCNNNeuronHardSigmoid*)hardSigmoid API_AVAILABLE(ios(11.0), macos(10.13));
/external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/include/
Dpytorch_qnnpack.h418 pytorch_qnnp_operator_t* hardsigmoid);
421 pytorch_qnnp_operator_t hardsigmoid,
/external/pytorch/aten/src/ATen/native/vulkan/ops/
DClamp.cpp324 Tensor hardsigmoid(const Tensor& self) { in hardsigmoid() function
325 return ops::activation(self, VK_KERNEL(hardsigmoid)); in hardsigmoid()
601 m.impl(TORCH_SELECTIVE_NAME("aten::hardsigmoid"), hardsigmoid); in TORCH_LIBRARY_IMPL()
/external/pytorch/torch/ao/pruning/_experimental/pruner/
Dbase_structured_sparsifier.py46 F.hardsigmoid,
73 nn.Hardsigmoid,
/external/pytorch/torch/ao/nn/quantized/
Dfunctional.py35 "hardsigmoid",
650 def hardsigmoid(input: Tensor, inplace: bool = False) -> Tensor: function
651 r"""This is the quantized version of :func:`~torch.nn.functional.hardsigmoid`."""
653 raise ValueError("Input to 'quantized.hardsigmoid' must be quantized!")
656 return torch._C._nn.hardsigmoid(input)
/external/pytorch/torch/csrc/jit/passes/
Drestore_mutation.h20 {aten::hardsigmoid, false},

12345