Home
last modified time | relevance | path

Searched full:relu_ (Results 1 – 25 of 167) sorted by relevance

1234567

/external/pytorch/torch/csrc/jit/passes/
Dfuse_relu.cpp27 %res = aten::relu_(%add_res) in fuseAddReluImpl()
34 %res = aten::relu_(%add_res) in fuseAddReluImpl()
45 %res = aten::relu_(%add_res) in fuseAddReluImpl()
Dmetal_rewrite.cpp119 %res = aten::relu_(%linear_res) in fuseReluWithPackedOps()
129 %r = aten::relu_(%r) in fuseReluWithPackedOps()
Dfrozen_conv_add_relu_fusion_cuda.cpp25 std::array<std::string, 2> relu_operators = {"relu", "relu_"}; in fuseFrozenConvAddReluImpl()
/external/pytorch/test/jit/
Dtest_optimize_for_mobile_preserve_debug_info.py256 conv2d_activation=F.relu_,
257 conv2d_activation_kind="aten::relu_",
263 linear_activation=F.relu_,
264 linear_activation_kind="aten::relu_",
Dtest_batch_mm.py108 torch.relu_(T1)
136 torch.relu_(T1)
172 torch.relu_(T1)
236 torch.relu_(T1)
272 torch.relu_(A)
Dtest_convert_activation.py113 FileCheck().check_not("aten::relu_").run(fn.graph)
128 FileCheck().check_not("aten::relu_").run(fn.graph)
168 torch.relu_,
Dtest_device_analysis.py115 def relu_(x): function
116 return torch.nn.functional.relu_(x)
118 functions = [add_self, relu_]
/external/pytorch/aten/src/ATen/native/metal/ops/
DMetalNeurons.mm59 static Tensor& relu_(Tensor& input) {
84 m.impl(TORCH_SELECTIVE_NAME("aten::relu_"), TORCH_FN(relu_));
/external/pytorch/torch/csrc/jit/passes/quantization/
Dquantization_patterns.h307 // aten::conv1d - aten::relu_ in quant_fusion_pattern_and_replacements()
314 %r = aten::relu_(%conv_out) in quant_fusion_pattern_and_replacements()
351 // aten::conv2d - aten::relu_ in quant_fusion_pattern_and_replacements()
358 %r = aten::relu_(%conv_out) in quant_fusion_pattern_and_replacements()
395 // aten::conv3d - aten::relu_ in quant_fusion_pattern_and_replacements()
402 %r = aten::relu_(%conv_out) in quant_fusion_pattern_and_replacements()
464 %r_relu = aten::relu_(%r_add) in quant_fusion_pattern_and_replacements()
482 %r_relu = aten::relu_(%r_add) in quant_fusion_pattern_and_replacements()
517 %r = aten::relu_(%linear_out) in quant_fusion_pattern_and_replacements()
595 %r = aten::relu_(%r_add) in quant_fusion_pattern_and_replacements()
[all …]
Dinsert_observers.cpp565 // nn.Linear + aten::relu_
570 %second_output = aten::relu_(%first_output)
600 // aten::linear + aten::relu_
605 %second_output = aten::relu_(%first_output)
636 %second_output = aten::relu_(%first_output)
668 %second_output = aten::relu_(%first_output)
700 %second_output = aten::relu_(%first_output)
745 %second_output = aten::relu_(%first_output)
757 %second_output = aten::relu_(%first_output)
788 %second_output = aten::relu_(%first_output)
[all …]
/external/executorch/backends/qualcomm/quantizer/
DREADME.md57 @register_annotator([torch.ops.aten.relu.default, torch.ops.aten.relu_.default])
59 Where `torch.ops.aten.relu.default` / `torch.ops.aten.relu_.default` map to `copy` / `in-place` ver…
162 @register_annotator([torch.ops.aten.relu.default, torch.ops.aten.relu_.default])
/external/pytorch/benchmarks/operator_benchmark/pt/
Dunary_test.py124 ["relu_", torch.relu_],
Dqunary_test.py94 ["q_relu_", torch.relu_],
/external/pytorch/torch/ao/quantization/backend_config/
D_qnnpack_pt2e.py98 (torch.ops.aten.convolution.default, torch.ops.aten.relu_.default)
160 (op_with_quantized_bop_scalar_variant, torch.ops.aten.relu_.default),
/external/pytorch/test/distributed/_tools/
Dtest_mod_tracker.py20 x = x["a"].relu_()
88 return self.foo(x).relu_()
/external/pytorch/test/dynamo/
Dtest_subclasses.py957 return x.add_(1.0) + torch.nn.functional.relu_(x)
974 relu_: "f32[3, 4]" = torch.relu_(l_x_); l_x_ = None
975 add: "f32[3, 4]" = add_ + relu_; add_ = relu_ = None
996 relu_: "f32[3, 4]" = torch.relu_(l_x_); l_x_ = None
997 add: "f32[3, 4]" = add_ + relu_; add_ = relu_ = None
1038 relu_: "f32[3, 4]" = torch.relu_(l_x_); l_x_ = None
1039 add: "f32[3, 4]" = add_ + relu_; add_ = relu_ = None
/external/pytorch/torch/ao/quantization/quantizer/
Dxnnpack_quantizer_utils.py237 torch.ops.aten.relu_.default,
343 torch.ops.aten.relu_.default,
486 output = F.relu_(bn) if relu_is_inplace else F.relu(bn)
720 torch.ops.aten.relu_.default,
842 torch.ops.aten.relu_.default,
/external/pytorch/torch/ao/quantization/pt2e/
Dgraph_utils.py25 {torch.nn.ReLU, torch.nn.functional.relu, torch.nn.functional.relu_},
/external/pytorch/aten/src/ATen/native/vulkan/ops/
DClamp.cpp305 Tensor& relu_(Tensor& self) { in relu_() function
618 m.impl(TORCH_SELECTIVE_NAME("aten::relu_"), relu_); in TORCH_LIBRARY_IMPL()
/external/pytorch/docs/source/
Dnn.functional.rst75 relu_
/external/pytorch/test/
Dtest_module_tracker.py21 x = x["a"].relu_()
/external/pytorch/torch/fx/passes/dialect/common/
Dcse_pass.py17 inplace_ops = {aten.add_, aten.sub_, aten.mul_, aten.div_, aten.pow_, aten.lerp_, aten.relu_, aten.…
/external/pytorch/torch/ao/ns/fx/
Dmappings.py105 "relu_",
704 "relu_",
/external/pytorch/aten/src/ATen/native/nested/
DNestedTensorUnaryOps.cpp96 at::relu_(buffer); in NestedTensor_relu_()
/external/pytorch/torch/csrc/jit/codegen/onednn/
Dinterface.cpp52 aten::relu_, in fuseGraph()

1234567