Home
last modified time | relevance | path

Searched full:mul_ (Results 1 – 25 of 209) sorted by relevance

123456789

/external/pytorch/test/autograd/
Dtest_complex.py16 x2.mul_(2)
20 y0.mul_(2)
36 x0.mul_(2)
56 x0.mul_(2)
/external/pytorch/test/cpp/api/
Dgrad_mode.cpp24 c.mul_(2); in TEST()
66 view_out.mul_( in TEST()
70 view_out.mul_(2); in TEST()
/external/libchrome-gestures/include/
Daccel_filter_interpreter.h52 CurveSegment() : x_(INFINITY), sqr_(0.0), mul_(1.0), int_(0.0) {} in CurveSegment()
54 : x_(x), sqr_(s), mul_(m), int_(b) {} in CurveSegment()
56 : x_(that.x_), sqr_(that.sqr_), mul_(that.mul_), int_(that.int_) {} in CurveSegment()
61 double mul_; // Slope of line (x multiplier) member
/external/pytorch/aten/src/ATen/native/
DDropout.cpp52 return input.mul_(noise); in multiply()
78 b = noise.add(-1).mul_(alpha * a).add_(alpha * a * p); in _dropout_impl()
79 noise.mul_(a); in _dropout_impl()
119 output = input.mul(mask).mul_(scale); in native_dropout_cpu()
DLoss.cpp301 loss.mul_(weight); in binary_cross_entropy_out_cpu()
354 grad_input.mul_(weight); in binary_cross_entropy_backward_out_cpu()
373 log_sigmoid_input.mul_(log_weight); in binary_cross_entropy_with_logits()
376 Tensor loss = (1 - target).mul_(input).sub_(log_sigmoid_input); in binary_cross_entropy_with_logits()
379 loss.mul_(weight); in binary_cross_entropy_with_logits()
406 at::mul_out(grad_input, target, z).mul_(-norm); in soft_margin_loss_backward_out()
408 grad_input.div_(z).mul_(grad_output); in soft_margin_loss_backward_out()
423 at::neg_out(output, input).mul_(target).exp_().log1p_(); in soft_margin_loss_out()
/external/executorch/backends/mediatek/quantizer/
Dannotator.py135 torch.ops.aten.mul_.Scalar,
136 torch.ops.aten.mul_.Tensor,
235 torch.ops.aten.mul_.Scalar,
236 torch.ops.aten.mul_.Tensor,
/external/pytorch/aten/src/ATen/test/
Doperators_test.cpp32 decltype(&ATEN_FN2(mul_, Tensor)), &ATEN_FN2(mul_, Tensor), in TEST()
/external/pytorch/torch/csrc/api/src/nn/modules/
D_functions.cpp71 .mul_( in forward()
79 output.mul_(input); in forward()
115 .mul_(grad_output); in backward()
/external/OpenCL-CTS/test_conformance/subgroups/
Dtest_subgroup_non_uniform_arithmetic.cpp50 error |= rft.run_impl<T, SCIN_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type()
58 error |= rft.run_impl<T, SCEX_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type()
66 error |= rft.run_impl<T, RED_NU<T, ArithmeticOp::mul_>>( in run_functions_add_mul_max_min_for_type()
Dsubgroup_common_templates.h284 case ArithmeticOp::mul_: return a * b; in calculate()
310 case ArithmeticOp::mul_: { in calculate()
332 case ArithmeticOp::mul_: { in calculate()
351 case ArithmeticOp::mul_: return to_half(to_float(a) * to_float(b)); in calculate()
369 if (operation == ArithmeticOp::mul_ || operation == ArithmeticOp::add_) in generate_inputs()
384 if (operation == ArithmeticOp::mul_ in generate_inputs()
/external/pytorch/test/inductor/
Dtest_layout_optim.py207 y.mul_(2)
225 x.mul_(2)
242 x.mul_(2)
260 y.mul_(2)
/external/pytorch/torch/csrc/api/src/optim/
Drmsprop.cpp116 square_avg.mul_(alpha).addcmul_(grad, grad, 1 - alpha); in step()
121 grad_avg.mul_(alpha).add_(grad, 1 - alpha); in step()
131 buf.mul_(options.momentum()).addcdiv_(grad, avg); in step()
Dadamw.cpp92 p.mul_(1 - options.lr() * options.weight_decay()); in step()
124 exp_avg.mul_(beta1).add_(grad, 1 - beta1); in step()
125 exp_avg_sq.mul_(beta2).addcmul_(grad, grad, 1 - beta2); in step()
Dadam.cpp123 exp_avg.mul_(beta1).add_(grad, 1 - beta1); in step()
124 exp_avg_sq.mul_(beta2).addcmul_(grad, grad, 1 - beta2); in step()
/external/pytorch/torch/optim/
D_functional.py66 exp_avg_update_values = grad_values.sub(old_exp_avg_values).mul_(1 - beta1)
70 grad_values.pow(2).sub_(old_exp_avg_sq_values).mul_(1 - beta2)
/external/pytorch/test/functorch/
Dtest_aotdispatch.py557 a.mul_(2)
599 a_.view(-1).mul_(2)
616 b.mul_(2)
664 a_view.mul_(2)
684 a.mul_(2)
812 a.mul_(2)
1055 out_ref[0].mul_(3)
1056 out_test[0].mul_(3)
1063 a.mul_(2)
1081 a.mul_(2)
[all …]
/external/pytorch/torch/nn/modules/
D_functions.py263 ctx.scale.mul_(ctx.alpha / ctx.size).add_(ctx.k)
266 output.mul_(input)
288 torch.pow(ctx.scale, -ctx.beta, out=grad_input).mul_(grad_output)
/external/pytorch/torch/_export/db/examples/
Duser_input_mutation.py11 x.mul_(2)
/external/pytorch/test/expect/
DTestJit.test_pretty_printer-while_test.expect7 a1 = torch.mul_(a0, a0)
/external/pytorch/aten/src/ATen/native/sparse/cuda/
DSparseBlas.cpp59 result.mul_(beta); in sparse_sampled_addmm_out_sparse_csr_cuda()
151 result.mul_(beta); in addmm_out_sparse_compressed_cuda()
188 result.mul_(beta); in baddbmm_out_sparse_csr_cuda()
/external/pytorch/aten/src/ATen/functorch/
DPyTorchOperatorHacks.cpp155 return input.mul_(noise); in multiply()
191 b = noise.add(-1).mul_(alpha * a).add_(alpha * a * p); in _dropout_impl()
192 noise.mul_(a); in _dropout_impl()
/external/libchrome-gestures/src/
Daccel_filter_interpreter_unittest.cc952 ASSERT_EQ(segs[0].mul_, 1.0); in TEST_F()
978 ASSERT_EQ(segs[0].mul_, linear_until_x / divisor); in TEST_F()
993 ASSERT_EQ(segs[1].mul_, 0.0); in TEST_F()
1010 ASSERT_EQ(segs[2].mul_, slope); in TEST_F()
1014 // return seg.mul_ + seg.int_ / speed;; in TEST_F()
1042 ASSERT_EQ(segs[0].mul_, 1.0); in TEST_F()
1068 ASSERT_EQ(segs[0].mul_, linear_until_x / divisor); in TEST_F()
1083 ASSERT_EQ(segs[1].mul_, 0.0); in TEST_F()
1100 ASSERT_EQ(segs[2].mul_, slope); in TEST_F()
/external/executorch/examples/portable/custom_ops/
Dcustom_ops_2.cpp23 out.mul_(4); in mul4_impl()
/external/pytorch/torch/csrc/jit/passes/
Dremove_inplace_ops.cpp10 {aten::mul_, aten::mul},
110 (it->kind() == aten::mul_) || (it->kind() == aten::div_)) { in ImplicitCastForBinaryInplaceOps()
/external/pytorch/aten/src/ATen/native/sparse/
DSparseBlasImpl.cpp266 result.mul_(alpha); in _compressed_row_strided_addmm_out()
282 result.mul_(beta); in _compressed_row_strided_addmm_out()
287 mm.mul_(alpha); in _compressed_row_strided_addmm_out()
294 result.mul_(alpha); in _compressed_row_strided_addmm_out()

123456789