| /external/pytorch/test/autograd/ |
| D | test_complex.py | 16 x2.mul_(2) 20 y0.mul_(2) 36 x0.mul_(2) 56 x0.mul_(2)
|
| /external/pytorch/test/cpp/api/ |
| D | grad_mode.cpp | 24 c.mul_(2); in TEST() 66 view_out.mul_( in TEST() 70 view_out.mul_(2); in TEST()
|
| /external/libchrome-gestures/include/ |
| D | accel_filter_interpreter.h | 52 CurveSegment() : x_(INFINITY), sqr_(0.0), mul_(1.0), int_(0.0) {} in CurveSegment() 54 : x_(x), sqr_(s), mul_(m), int_(b) {} in CurveSegment() 56 : x_(that.x_), sqr_(that.sqr_), mul_(that.mul_), int_(that.int_) {} in CurveSegment() 61 double mul_; // Slope of line (x multiplier) member
|
| /external/pytorch/aten/src/ATen/native/ |
| D | Dropout.cpp | 52 return input.mul_(noise); in multiply() 78 b = noise.add(-1).mul_(alpha * a).add_(alpha * a * p); in _dropout_impl() 79 noise.mul_(a); in _dropout_impl() 119 output = input.mul(mask).mul_(scale); in native_dropout_cpu()
|
| D | Loss.cpp | 301 loss.mul_(weight); in binary_cross_entropy_out_cpu() 354 grad_input.mul_(weight); in binary_cross_entropy_backward_out_cpu() 373 log_sigmoid_input.mul_(log_weight); in binary_cross_entropy_with_logits() 376 Tensor loss = (1 - target).mul_(input).sub_(log_sigmoid_input); in binary_cross_entropy_with_logits() 379 loss.mul_(weight); in binary_cross_entropy_with_logits() 406 at::mul_out(grad_input, target, z).mul_(-norm); in soft_margin_loss_backward_out() 408 grad_input.div_(z).mul_(grad_output); in soft_margin_loss_backward_out() 423 at::neg_out(output, input).mul_(target).exp_().log1p_(); in soft_margin_loss_out()
|
| /external/executorch/backends/mediatek/quantizer/ |
| D | annotator.py | 135 torch.ops.aten.mul_.Scalar, 136 torch.ops.aten.mul_.Tensor, 235 torch.ops.aten.mul_.Scalar, 236 torch.ops.aten.mul_.Tensor,
|
| /external/pytorch/aten/src/ATen/test/ |
| D | operators_test.cpp | 32 decltype(&ATEN_FN2(mul_, Tensor)), &ATEN_FN2(mul_, Tensor), in TEST()
|
| /external/pytorch/torch/csrc/api/src/nn/modules/ |
| D | _functions.cpp | 71 .mul_( in forward() 79 output.mul_(input); in forward() 115 .mul_(grad_output); in backward()
|
| /external/OpenCL-CTS/test_conformance/subgroups/ |
| D | test_subgroup_non_uniform_arithmetic.cpp | 50 error |= rft.run_impl<T, SCIN_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type() 58 error |= rft.run_impl<T, SCEX_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type() 66 error |= rft.run_impl<T, RED_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type()
|
| D | subgroup_common_templates.h | 284 case ArithmeticOp::mul_: return a * b; in calculate() 310 case ArithmeticOp::mul_: { in calculate() 332 case ArithmeticOp::mul_: { in calculate() 351 case ArithmeticOp::mul_: return to_half(to_float(a) * to_float(b)); in calculate() 369 if (operation == ArithmeticOp::mul_ || operation == ArithmeticOp::add_) in generate_inputs() 384 if (operation == ArithmeticOp::mul_ in generate_inputs()
|
| /external/pytorch/test/inductor/ |
| D | test_layout_optim.py | 207 y.mul_(2) 225 x.mul_(2) 242 x.mul_(2) 260 y.mul_(2)
|
| /external/pytorch/torch/csrc/api/src/optim/ |
| D | rmsprop.cpp | 116 square_avg.mul_(alpha).addcmul_(grad, grad, 1 - alpha); in step() 121 grad_avg.mul_(alpha).add_(grad, 1 - alpha); in step() 131 buf.mul_(options.momentum()).addcdiv_(grad, avg); in step()
|
| D | adamw.cpp | 92 p.mul_(1 - options.lr() * options.weight_decay()); in step() 124 exp_avg.mul_(beta1).add_(grad, 1 - beta1); in step() 125 exp_avg_sq.mul_(beta2).addcmul_(grad, grad, 1 - beta2); in step()
|
| D | adam.cpp | 123 exp_avg.mul_(beta1).add_(grad, 1 - beta1); in step() 124 exp_avg_sq.mul_(beta2).addcmul_(grad, grad, 1 - beta2); in step()
|
| /external/pytorch/torch/optim/ |
| D | _functional.py | 66 exp_avg_update_values = grad_values.sub(old_exp_avg_values).mul_(1 - beta1) 70 grad_values.pow(2).sub_(old_exp_avg_sq_values).mul_(1 - beta2)
|
| /external/pytorch/test/functorch/ |
| D | test_aotdispatch.py | 557 a.mul_(2) 599 a_.view(-1).mul_(2) 616 b.mul_(2) 664 a_view.mul_(2) 684 a.mul_(2) 812 a.mul_(2) 1055 out_ref[0].mul_(3) 1056 out_test[0].mul_(3) 1063 a.mul_(2) 1081 a.mul_(2) [all …]
|
| /external/pytorch/torch/nn/modules/ |
| D | _functions.py | 263 ctx.scale.mul_(ctx.alpha / ctx.size).add_(ctx.k) 266 output.mul_(input) 288 torch.pow(ctx.scale, -ctx.beta, out=grad_input).mul_(grad_output)
|
| /external/pytorch/torch/_export/db/examples/ |
| D | user_input_mutation.py | 11 x.mul_(2)
|
| /external/pytorch/test/expect/ |
| D | TestJit.test_pretty_printer-while_test.expect | 7 a1 = torch.mul_(a0, a0)
|
| /external/pytorch/aten/src/ATen/native/sparse/cuda/ |
| D | SparseBlas.cpp | 59 result.mul_(beta); in sparse_sampled_addmm_out_sparse_csr_cuda() 151 result.mul_(beta); in addmm_out_sparse_compressed_cuda() 188 result.mul_(beta); in baddbmm_out_sparse_csr_cuda()
|
| /external/pytorch/aten/src/ATen/functorch/ |
| D | PyTorchOperatorHacks.cpp | 155 return input.mul_(noise); in multiply() 191 b = noise.add(-1).mul_(alpha * a).add_(alpha * a * p); in _dropout_impl() 192 noise.mul_(a); in _dropout_impl()
|
| /external/libchrome-gestures/src/ |
| D | accel_filter_interpreter_unittest.cc | 952 ASSERT_EQ(segs[0].mul_, 1.0); in TEST_F() 978 ASSERT_EQ(segs[0].mul_, linear_until_x / divisor); in TEST_F() 993 ASSERT_EQ(segs[1].mul_, 0.0); in TEST_F() 1010 ASSERT_EQ(segs[2].mul_, slope); in TEST_F() 1014 // return seg.mul_ + seg.int_ / speed;; in TEST_F() 1042 ASSERT_EQ(segs[0].mul_, 1.0); in TEST_F() 1068 ASSERT_EQ(segs[0].mul_, linear_until_x / divisor); in TEST_F() 1083 ASSERT_EQ(segs[1].mul_, 0.0); in TEST_F() 1100 ASSERT_EQ(segs[2].mul_, slope); in TEST_F()
|
| /external/executorch/examples/portable/custom_ops/ |
| D | custom_ops_2.cpp | 23 out.mul_(4); in mul4_impl()
|
| /external/pytorch/torch/csrc/jit/passes/ |
| D | remove_inplace_ops.cpp | 10 {aten::mul_, aten::mul}, 110 (it->kind() == aten::mul_) || (it->kind() == aten::div_)) { in ImplicitCastForBinaryInplaceOps()
|
| /external/pytorch/aten/src/ATen/native/sparse/ |
| D | SparseBlasImpl.cpp | 266 result.mul_(alpha); in _compressed_row_strided_addmm_out() 282 result.mul_(beta); in _compressed_row_strided_addmm_out() 287 mm.mul_(alpha); in _compressed_row_strided_addmm_out() 294 result.mul_(alpha); in _compressed_row_strided_addmm_out()
|