| /external/pytorch/torch/csrc/jit/operator_upgraders/ |
| D | version_map.cpp | 58 {"aten::div_.Tensor", 61 "aten::div_.Tensor(Tensor(a!) self, Tensor other) -> Tensor(a!)"}}}, 62 {"aten::div_.Tensor_mode", 65 … "aten::div_.Tensor_mode(Tensor(a!) self, Tensor other, *, str? rounding_mode) -> Tensor(a!)"}}}, 66 {"aten::div_.Scalar", 69 "aten::div_.Scalar(Tensor(a!) self, Scalar other) -> Tensor(a!)"}}}, 70 {"aten::div_.Scalar_mode", 73 … "aten::div_.Scalar_mode(Tensor(a!) self, Scalar other, *, str? rounding_mode) -> Tensor(a!)"}}},
|
| /external/strace/ |
| D | macros.h | 52 # define ROUNDUP(val_, div_) ((((val_) + (div_) - 1) / (div_)) * (div_)) argument
|
| /external/pytorch/torch/distributed/algorithms/_comm_hooks/ |
| D | default_hooks.py | 107 grad.div_(state.gradient_predivide_factor) 111 grad.div_(state.gradient_postdivide_factor) 128 grad.div_(state.gradient_predivide_factor) 132 output.div_(state.gradient_postdivide_factor)
|
| /external/executorch/backends/mediatek/quantizer/ |
| D | annotator.py | 129 torch.ops.aten.div_.Scalar, 130 torch.ops.aten.div_.Tensor, 224 torch.ops.aten.div_.Scalar, 225 torch.ops.aten.div_.Tensor,
|
| /external/pytorch/torch/csrc/jit/mobile/ |
| D | upgrader_mobile.cpp | 49 {std::string("aten::div_.Scalar"), in getOperatorVersionMapForMobile() 53 {std::string("aten::div_.Scalar_mode"), in getOperatorVersionMapForMobile() 57 {std::string("aten::div_.Tensor"), in getOperatorVersionMapForMobile() 61 {std::string("aten::div_.Tensor_mode"), in getOperatorVersionMapForMobile() 263 OperatorString({"aten::div_", "Scalar", 2}), in getUpgraderBytecodeList() 265 OperatorString({"aten::div_", "Scalar_mode", 3}), in getUpgraderBytecodeList() 284 OperatorString({"aten::div_", "Scalar_mode", 3}), in getUpgraderBytecodeList() 325 OperatorString({"aten::div_", "Tensor", 2}), in getUpgraderBytecodeList() 326 OperatorString({"aten::div_", "Tensor_mode", 3}), in getUpgraderBytecodeList() 345 OperatorString({"aten::div_", "Tensor_mode", 3}), in getUpgraderBytecodeList()
|
| /external/pytorch/test/mobile/ |
| D | test_upgrader_bytecode_table_example.cpp | 41 {std::string("aten::div_.Scalar"), in getOperatorVersionMapForMobile() 45 {std::string("aten::div_.Tensor"), in getOperatorVersionMapForMobile() 193 OperatorString({"aten::div_", "Scalar", 2}), in getUpgraderBytecodeList() 195 OperatorString({"aten::div_", "Scalar_mode", 3}), in getUpgraderBytecodeList() 236 OperatorString({"aten::div_", "Tensor", 2}), in getUpgraderBytecodeList() 237 OperatorString({"aten::div_", "Tensor_mode", 3}), in getUpgraderBytecodeList()
|
| /external/pytorch/torch/distributed/algorithms/ddp_comm_hooks/ |
| D | default_hooks.py | 24 tensor.div_(group_to_use.size()) 80 compressed_tensor = buffer.to(torch.float16).div_(world_size) 129 compressed_tensor = buffer.to(torch.bfloat16).div_(world_size)
|
| D | mixed_precision_hooks.py | 52 bucket.buffer().div_(process_group.size())
|
| D | optimizer_overlap_hooks.py | 80 bucket.buffer().div_(process_group.size())
|
| /external/pytorch/test/expect/ |
| D | HasDecompTest.test_aten_core_operators.expect | 194 aten::div_.Scalar 195 aten::div_.Scalar_mode 196 aten::div_.Tensor 197 aten::div_.Tensor_mode
|
| /external/pytorch/torch/csrc/jit/frontend/ |
| D | versioned_symbols.cpp | 73 {Symbol::fromQualString("aten::div_"), 81 {aten::div_, 4},
|
| /external/pytorch/aten/src/ATen/functorch/ |
| D | BatchRulesBinaryOps.cpp | 473 …VMAP_SUPPORT2(div_, Tensor, SINGLE_ARG(binary_pointwise_inplace_batch_rule<TensorInplaceT, &Tensor… in TORCH_LIBRARY_IMPL() 474 …VMAP_SUPPORT2(div_, Tensor_mode, SINGLE_ARG(binary_pointwise_inplace_batch_rule<TensorInplaceModeT… in TORCH_LIBRARY_IMPL() 475 …VMAP_SUPPORT2(div_, Scalar, SINGLE_ARG(unary_inplace_batch_rule<ScalarInplaceT, &Tensor::div_, con… in TORCH_LIBRARY_IMPL()
|
| /external/pytorch/torch/csrc/jit/passes/ |
| D | remove_inplace_ops.cpp | 9 {aten::div_, aten::div}, 110 (it->kind() == aten::mul_) || (it->kind() == aten::div_)) { in ImplicitCastForBinaryInplaceOps()
|
| D | normalize_ops.cpp | 123 {aten::divide_, aten::div_}, in getOperatorAliasMap() 129 {aten::true_divide_, aten::div_}, in getOperatorAliasMap()
|
| /external/pytorch/test/distributed/_composable/fsdp/ |
| D | test_fully_shard_mixed_precision.py | 107 param.grad.div_(predivide_factor) 109 param.grad.div_(self.world_size) 114 param.grad.div_(postdivide_factor) 170 param.grad.div_(self.world_size)
|
| /external/executorch/backends/arm/_passes/ |
| D | decompose_div_pass.py | 14 aten_div_ops = (torch.ops.aten.div.Tensor, torch.ops.aten.div_.Tensor)
|
| D | scalars_to_attribute_pass.py | 34 torch.ops.aten.div_.Tensor,
|
| /external/pytorch/aten/src/ATen/ |
| D | TensorOperators.h | 22 ::at::empty_like(y, at::MemoryFormat::Preserve).fill_(x).div_(y)) \
|
| /external/pytorch/aten/src/ATen/native/ |
| D | BinaryOps.cpp | 895 Tensor& div_(Tensor& self, const Scalar& other) { in div_() function 896 return self.div_(wrapped_scalar_tensor(other)); // redispatch! in div_() 903 Tensor& div_(Tensor& self, const Scalar& other, std::optional<c10::string_view> rounding_mode) { in div_() function 904 return self.div_(wrapped_scalar_tensor(other), std::move(rounding_mode)); // redispatch! in div_() 917 return self.div_(other); in divide_() 925 return self.div_(other); in divide_() 937 return self.div_(other, std::move(rounding_mode)); in divide_() 945 return self.div_(other, std::move(rounding_mode)); in divide_() 958 return self.div_(divisor); in true_divide_() 966 return self.div_(divisor); in true_divide_()
|
| /external/pytorch/torch/distributed/tensor/_ops/ |
| D | _pointwise_ops.py | 52 …aten.div_.Scalar, # this op is linear on the first argument, and the second argument is scalar, s… 177 aten.div_.Tensor, 178 aten.div_.Tensor_mode,
|
| /external/pytorch/torch/optim/ |
| D | _functional.py | 84 param.add_(make_sparse(-step_size * numer.div_(denom)))
|
| /external/pytorch/test/jit/ |
| D | test_save_load_for_op_version.py | 169 _helper(current_mobile_module, torch.Tensor.div_) 420 _helper(current_mobile_module_float, torch.Tensor.div_) 422 _helper(current_mobile_module_int, torch.Tensor.div_)
|
| /external/pytorch/aten/src/ATen/native/cpu/ |
| D | HistogramKernel.cpp | 240 hist.div_(hist_sum); in histogramdd_out_cpu_template() 252 hist.div_(bin_lengths.reshape(shape)); in histogramdd_out_cpu_template()
|
| /external/pytorch/torch/fx/passes/dialect/common/ |
| D | cse_pass.py | 17 inplace_ops = {aten.add_, aten.sub_, aten.mul_, aten.div_, aten.pow_, aten.lerp_, aten.relu_, aten.…
|
| /external/pytorch/torch/csrc/api/src/data/datasets/ |
| D | mnist.cpp | 79 return tensor.to(torch::kFloat32).div_(255); in read_images()
|