Home
last modified time | relevance | path

Searched defs:learning_rate (Results 1 – 25 of 75) sorted by relevance

123

/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/
Dmomentum_impl.cu19 …mentumUpdateVariableKernel(const size_t size, T *variable, T *accumulation, const S *learning_rate, in MomentumUpdateVariableKernel()
35const float *learning_rate, const half *gradient, const float *momentum, in MomentumUpdateVariableKernel()
52const float *learning_rate, const half *gradient, const float *momentum, in MomentumUpdateVariableKernel()
67 void MomentumUpdateVariable(const size_t size, T *variable, T *accumulation, const S *learning_rate in MomentumUpdateVariable()
75 … T *accumulation, const T *learning_rate, const S *gradient, in FusedMomentumWeightDecayScaleKernel()
86 const T *learning_rate, const S *gradient, const T *momentum, in FusedWeightDecayScaleMomentum()
96const T *learning_rate, const S *gradient, const T *momentum) { in FusedMomentumScaleKernel()
104 …eMomentum(const size_t element_num, T *scale, T *variable, T *accumulation, const T *learning_rate, in FusedScaleMomentum()
114const T *learning_rate, const S *gradient, const T *momentum) { in FusedWeightDecayMomentumKernel()
124const T *learning_rate, const S *gradient, const T *momentum, cudaStream_t cuda_stream) { in FusedWeightDecayMomentum()
[all …]
Dadagrad_impl.cu32 const S *learning_rate, in ApplyAdagradKernel()
47 const float *learning_rate, in ApplyAdagradKernel()
62 const float *learning_rate, in ApplyAdagradKernel()
77 const half *learning_rate, in ApplyAdagradKernel()
92 const float *learning_rate, in ApplyAdagradKernel()
107 const S *learning_rate, in ApplyAdagrad()
Drmsprop_impl.cu22 __global__ void RmsPropKernel(const T* learning_rate, const T decay, const T momentum, const T epsi… in RmsPropKernel()
32 void RmsProp(const T* learning_rate, const T decay, const T momentum, const T epsilon, in RmsProp()
39 __global__ void RmsPropCenterKernel(const T* learning_rate, const T* decay, const T* momentum, cons… in RmsPropCenterKernel()
52 void RmsPropCenter(const T* learning_rate, const T* decay, const T* momentum, const T* epsilon, T* … in RmsPropCenter()
Dadam_impl.cu31const T *learning_rate, const T *beta1, const T *beta2, const T *epsilon, T *variable, in ApplyAdamKernel()
44 …obal__ void AdamWeightDecayKernel(const size_t size, const T *gradient, const float *learning_rate, in AdamWeightDecayKernel()
58 …l__ void AdamWeightDecayKernel(const size_t size, const half *gradient, const float *learning_rate, in AdamWeightDecayKernel()
72 …size_t size, const T *gradient, const T *beta1_power, const T *beta2_power, const T *learning_rate, in ApplyAdam()
78 void AdamWeightDecayOp(const size_t size, const T *gradient, const float *learning_rate, const floa… in AdamWeightDecayOp()
Dftrl_impl.cu50 __global__ void ApplyFtrlKernel(const size_t size, const T *gradient, const T *learning_rate, in ApplyFtrlKernel()
72 void ApplyFtrl(const size_t size, const T *gradient, const T *learning_rate, const T *l1_regulariza… in ApplyFtrl()
Dsparse_apply_proximal_adagrad_impl.cu60 …rseApplyProximalAdagradUpdate(const size_t size, const size_t indices_size, const T *learning_rate, in SparseApplyProximalAdagradUpdate()
84 …CalSparseApplyProximalAdagrad(const size_t size, const size_t indices_size, const T *learning_rate, in CalSparseApplyProximalAdagrad()
/third_party/mindspore/mindspore/nn/
Ddynamic_lr.py65 def _check_inputs(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_stair): argument
76 def exponential_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_sta… argument
119 def natural_exp_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_sta… argument
163 def inverse_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_stair=F… argument
257 def polynomial_decay_lr(learning_rate, end_learning_rate, total_step, step_per_epoch, decay_epoch, … argument
333 def warmup_lr(learning_rate, total_step, step_per_epoch, warmup_epoch): argument
Dlearning_rate_schedule.py48 def _check_inputs(learning_rate, decay_rate, decay_steps, is_stair, cls_name): argument
107 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument
174 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument
242 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument
376 …def __init__(self, learning_rate, end_learning_rate, decay_steps, power, update_decay_steps=False): argument
453 def __init__(self, learning_rate, warmup_steps): argument
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/sponge/nvtit/
Dmd_iteration_gradient_descent_impl.cu22 const float learning_rate) { in MD_Iteration_Gradient_Descent()
35 …IterationGradientDescent(const int atom_numbers, float *crd, float *frc, const float learning_rate, in MDIterationGradientDescent()
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/
Drmsprop_cpu_kernel.cc31 float *learning_rate) { in LaunchRMSPropUnuseCenter()
52 … T *mean_gradients, float *momentum, float *learning_rate, float *decay, in LaunchRMSPropUseCenter()
105 float *learning_rate = reinterpret_cast<float *>(inputs[3]->addr); in Launch() local
118 float *learning_rate = reinterpret_cast<float *>(inputs[5]->addr); in Launch() local
/third_party/mindspore/mindspore/nn/optim/
Doptimizer.py129 def __init__(self, learning_rate, parameters, weight_decay=0.0, loss_scale=1.0): argument
368 def _preprocess_single_lr(self, learning_rate): argument
392 def _build_single_lr(self, learning_rate, name): argument
439 def _parse_group_params(self, parameters, learning_rate): argument
468 def _init_group_params(self, parameters, learning_rate, weight_decay, grad_centralization): argument
577 def get_lr_value(learning_rate): argument
758 def __init__(self, learning_rate): argument
770 def __init__(self, learning_rate, name): argument
Drmsprop.py26 def _rmsprop_opt_(opt, decay, epsilon, momentum, learning_rate, weight, ms, mom, grad): argument
35 def _centered_rmsprop_opt_(opt, decay, epsilon, momentum, learning_rate, weight, mg, ms, mom, grad): argument
181 def __init__(self, params, learning_rate=0.1, decay=0.9, momentum=0.0, epsilon=1e-10, argument
Dproximal_ada_grad.py28 def _tensor_run_opt_with_sparse(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument
37 def _tensor_run_opt(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument
172 def __init__(self, params, accum=0.1, learning_rate=0.001, l1=0.0, l2=0.0, argument
Dmomentum.py28 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, moment, ps_parameter, cache… argument
151 …def __init__(self, params, learning_rate, momentum, weight_decay=0.0, loss_scale=1.0, use_nesterov… argument
Dsgd.py28 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, accum, stat): argument
139 …def __init__(self, params, learning_rate=0.1, momentum=0.0, dampening=0.0, weight_decay=0.0, neste… argument
Dada_grad.py25 def _tensor_run_opt(opt, learning_rate, weight, accum, gradient): argument
149 def __init__(self, params, accum=0.1, learning_rate=0.001, argument
Dftrl.py29 def _tensor_run_opt_with_sparse(opt, spars_opt, push, pull, l1, l2, lr_power, learning_rate, linear, argument
46 def _tensor_run_opt(opt, spars_opt, push, pull, l1, l2, lr_power, learning_rate, linear, argument
197 … def __init__(self, params, initial_accum=0.1, learning_rate=0.001, lr_power=-0.5, l1=0.0, l2=0.0, argument
Dadafactor.py30 def _get_lr(step, rms, learning_rate, relative_step, warmup_init, scale_parameter, eps): argument
71 learning_rate, step, grad, param, argument
290 learning_rate=None, argument
/third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/fp32_grad/
Dsgd.cc33 int DoSgd(float *weight, float *accumulate, const float *gradient, float learning_rate, float dampe… in DoSgd()
55 int DoSgdInit(float *weight, float *accumulate, float *gradient, float *stat, float learning_rate, … in DoSgdInit()
76 float learning_rate = lr_; in Execute() local
100 float learning_rate = lr_; in ExecuteInit() local
207 float learning_rate = lr_; in OptimizerStep() local
Dapply_momentum.cc32 static int DoApplyMomentum(float *weight, float *accumulate, float learning_rate, const float *grad… in DoApplyMomentum()
54 float learning_rate = lr_; in Execute() local
124 float learning_rate = lr_; in OptimizerStep() local
Dadam.cc35 … float beta2_power, float eps, float learning_rate, bool nesterov, int start, int end) { in DoAdam()
71 auto learning_rate = lr_; in Execute() local
142 auto learning_rate = lr_; in OptimizerStep() local
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/nn/
Drmsprop_gpu_kernel.h47 T *learning_rate = GetDeviceAddress<T>(inputs, 3); in Launch() local
58 T *learning_rate = GetDeviceAddress<T>(inputs, 5); in Launch() local
/third_party/mindspore/tests/mindspore_test_framework/utils/
Dmodel_util.py49 def update_opt_step(learning_rate, batch_size, parameter, gradient): argument
69 def __init__(self, parameters, learning_rate=0.001, batch_size=1): argument
/third_party/mindspore/tests/st/networks/models/resnet50/src_thor/
Dthor.py56 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, moment): argument
147 def THOR(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0, batch_size=32, argument
159 …def __init__(self, net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0, batch_… argument
/third_party/mindspore/tests/st/ops/gpu/
Dtest_rmsprop.py67 learning_rate, decay, momentum, epsilon): argument
75 learning_rate, decay, momentum, epsilon): argument

123