/external/pytorch/torch/csrc/jit/passes/ |
D | fuse_relu.cpp | 27 %res = aten::relu_(%add_res) in fuseAddReluImpl() 34 %res = aten::relu_(%add_res) in fuseAddReluImpl() 45 %res = aten::relu_(%add_res) in fuseAddReluImpl()
|
D | metal_rewrite.cpp | 119 %res = aten::relu_(%linear_res) in fuseReluWithPackedOps() 129 %r = aten::relu_(%r) in fuseReluWithPackedOps()
|
D | frozen_conv_add_relu_fusion_cuda.cpp | 25 std::array<std::string, 2> relu_operators = {"relu", "relu_"}; in fuseFrozenConvAddReluImpl()
|
/external/pytorch/test/jit/ |
D | test_optimize_for_mobile_preserve_debug_info.py | 256 conv2d_activation=F.relu_, 257 conv2d_activation_kind="aten::relu_", 263 linear_activation=F.relu_, 264 linear_activation_kind="aten::relu_",
|
D | test_batch_mm.py | 108 torch.relu_(T1) 136 torch.relu_(T1) 172 torch.relu_(T1) 236 torch.relu_(T1) 272 torch.relu_(A)
|
D | test_convert_activation.py | 113 FileCheck().check_not("aten::relu_").run(fn.graph) 128 FileCheck().check_not("aten::relu_").run(fn.graph) 168 torch.relu_,
|
D | test_device_analysis.py | 115 def relu_(x): function 116 return torch.nn.functional.relu_(x) 118 functions = [add_self, relu_]
|
/external/pytorch/aten/src/ATen/native/metal/ops/ |
D | MetalNeurons.mm | 59 static Tensor& relu_(Tensor& input) { 84 m.impl(TORCH_SELECTIVE_NAME("aten::relu_"), TORCH_FN(relu_));
|
/external/pytorch/torch/csrc/jit/passes/quantization/ |
D | quantization_patterns.h | 307 // aten::conv1d - aten::relu_ in quant_fusion_pattern_and_replacements() 314 %r = aten::relu_(%conv_out) in quant_fusion_pattern_and_replacements() 351 // aten::conv2d - aten::relu_ in quant_fusion_pattern_and_replacements() 358 %r = aten::relu_(%conv_out) in quant_fusion_pattern_and_replacements() 395 // aten::conv3d - aten::relu_ in quant_fusion_pattern_and_replacements() 402 %r = aten::relu_(%conv_out) in quant_fusion_pattern_and_replacements() 464 %r_relu = aten::relu_(%r_add) in quant_fusion_pattern_and_replacements() 482 %r_relu = aten::relu_(%r_add) in quant_fusion_pattern_and_replacements() 517 %r = aten::relu_(%linear_out) in quant_fusion_pattern_and_replacements() 595 %r = aten::relu_(%r_add) in quant_fusion_pattern_and_replacements() [all …]
|
D | insert_observers.cpp | 565 // nn.Linear + aten::relu_ 570 %second_output = aten::relu_(%first_output) 600 // aten::linear + aten::relu_ 605 %second_output = aten::relu_(%first_output) 636 %second_output = aten::relu_(%first_output) 668 %second_output = aten::relu_(%first_output) 700 %second_output = aten::relu_(%first_output) 745 %second_output = aten::relu_(%first_output) 757 %second_output = aten::relu_(%first_output) 788 %second_output = aten::relu_(%first_output) [all …]
|
/external/executorch/backends/qualcomm/quantizer/ |
D | README.md | 57 @register_annotator([torch.ops.aten.relu.default, torch.ops.aten.relu_.default]) 59 Where `torch.ops.aten.relu.default` / `torch.ops.aten.relu_.default` map to `copy` / `in-place` ver… 162 @register_annotator([torch.ops.aten.relu.default, torch.ops.aten.relu_.default])
|
/external/pytorch/benchmarks/operator_benchmark/pt/ |
D | unary_test.py | 124 ["relu_", torch.relu_],
|
D | qunary_test.py | 94 ["q_relu_", torch.relu_],
|
/external/pytorch/torch/ao/quantization/backend_config/ |
D | _qnnpack_pt2e.py | 98 (torch.ops.aten.convolution.default, torch.ops.aten.relu_.default) 160 (op_with_quantized_bop_scalar_variant, torch.ops.aten.relu_.default),
|
/external/pytorch/test/distributed/_tools/ |
D | test_mod_tracker.py | 20 x = x["a"].relu_() 88 return self.foo(x).relu_()
|
/external/pytorch/test/dynamo/ |
D | test_subclasses.py | 957 return x.add_(1.0) + torch.nn.functional.relu_(x) 974 relu_: "f32[3, 4]" = torch.relu_(l_x_); l_x_ = None 975 add: "f32[3, 4]" = add_ + relu_; add_ = relu_ = None 996 relu_: "f32[3, 4]" = torch.relu_(l_x_); l_x_ = None 997 add: "f32[3, 4]" = add_ + relu_; add_ = relu_ = None 1038 relu_: "f32[3, 4]" = torch.relu_(l_x_); l_x_ = None 1039 add: "f32[3, 4]" = add_ + relu_; add_ = relu_ = None
|
/external/pytorch/torch/ao/quantization/quantizer/ |
D | xnnpack_quantizer_utils.py | 237 torch.ops.aten.relu_.default, 343 torch.ops.aten.relu_.default, 486 output = F.relu_(bn) if relu_is_inplace else F.relu(bn) 720 torch.ops.aten.relu_.default, 842 torch.ops.aten.relu_.default,
|
/external/pytorch/torch/ao/quantization/pt2e/ |
D | graph_utils.py | 25 {torch.nn.ReLU, torch.nn.functional.relu, torch.nn.functional.relu_},
|
/external/pytorch/aten/src/ATen/native/vulkan/ops/ |
D | Clamp.cpp | 305 Tensor& relu_(Tensor& self) { in relu_() function 618 m.impl(TORCH_SELECTIVE_NAME("aten::relu_"), relu_); in TORCH_LIBRARY_IMPL()
|
/external/pytorch/docs/source/ |
D | nn.functional.rst | 75 relu_
|
/external/pytorch/test/ |
D | test_module_tracker.py | 21 x = x["a"].relu_()
|
/external/pytorch/torch/fx/passes/dialect/common/ |
D | cse_pass.py | 17 inplace_ops = {aten.add_, aten.sub_, aten.mul_, aten.div_, aten.pow_, aten.lerp_, aten.relu_, aten.…
|
/external/pytorch/torch/ao/ns/fx/ |
D | mappings.py | 105 "relu_", 704 "relu_",
|
/external/pytorch/aten/src/ATen/native/nested/ |
D | NestedTensorUnaryOps.cpp | 96 at::relu_(buffer); in NestedTensor_relu_()
|
/external/pytorch/torch/csrc/jit/codegen/onednn/ |
D | interface.cpp | 52 aten::relu_, in fuseGraph()
|