| /third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/ |
| D | momentum_impl.cu | 19 …mentumUpdateVariableKernel(const size_t size, T *variable, T *accumulation, const S *learning_rate, in MomentumUpdateVariableKernel() 35 … const float *learning_rate, const half *gradient, const float *momentum, in MomentumUpdateVariableKernel() 52 … const float *learning_rate, const half *gradient, const float *momentum, in MomentumUpdateVariableKernel() 67 void MomentumUpdateVariable(const size_t size, T *variable, T *accumulation, const S *learning_rate… in MomentumUpdateVariable() 75 … T *accumulation, const T *learning_rate, const S *gradient, in FusedMomentumWeightDecayScaleKernel() 86 const T *learning_rate, const S *gradient, const T *momentum, in FusedWeightDecayScaleMomentum() 96 … const T *learning_rate, const S *gradient, const T *momentum) { in FusedMomentumScaleKernel() 104 …eMomentum(const size_t element_num, T *scale, T *variable, T *accumulation, const T *learning_rate, in FusedScaleMomentum() 114 … const T *learning_rate, const S *gradient, const T *momentum) { in FusedWeightDecayMomentumKernel() 124 … const T *learning_rate, const S *gradient, const T *momentum, cudaStream_t cuda_stream) { in FusedWeightDecayMomentum() [all …]
|
| D | adagrad_impl.cu | 32 const S *learning_rate, in ApplyAdagradKernel() 47 const float *learning_rate, in ApplyAdagradKernel() 62 const float *learning_rate, in ApplyAdagradKernel() 77 const half *learning_rate, in ApplyAdagradKernel() 92 const float *learning_rate, in ApplyAdagradKernel() 107 const S *learning_rate, in ApplyAdagrad()
|
| D | rmsprop_impl.cu | 22 __global__ void RmsPropKernel(const T* learning_rate, const T decay, const T momentum, const T epsi… in RmsPropKernel() 32 void RmsProp(const T* learning_rate, const T decay, const T momentum, const T epsilon, in RmsProp() 39 __global__ void RmsPropCenterKernel(const T* learning_rate, const T* decay, const T* momentum, cons… in RmsPropCenterKernel() 52 void RmsPropCenter(const T* learning_rate, const T* decay, const T* momentum, const T* epsilon, T* … in RmsPropCenter()
|
| D | adam_impl.cu | 31 … const T *learning_rate, const T *beta1, const T *beta2, const T *epsilon, T *variable, in ApplyAdamKernel() 44 …obal__ void AdamWeightDecayKernel(const size_t size, const T *gradient, const float *learning_rate, in AdamWeightDecayKernel() 58 …l__ void AdamWeightDecayKernel(const size_t size, const half *gradient, const float *learning_rate, in AdamWeightDecayKernel() 72 …size_t size, const T *gradient, const T *beta1_power, const T *beta2_power, const T *learning_rate, in ApplyAdam() 78 void AdamWeightDecayOp(const size_t size, const T *gradient, const float *learning_rate, const floa… in AdamWeightDecayOp()
|
| D | ftrl_impl.cu | 50 __global__ void ApplyFtrlKernel(const size_t size, const T *gradient, const T *learning_rate, in ApplyFtrlKernel() 72 void ApplyFtrl(const size_t size, const T *gradient, const T *learning_rate, const T *l1_regulariza… in ApplyFtrl()
|
| D | sparse_apply_proximal_adagrad_impl.cu | 60 …rseApplyProximalAdagradUpdate(const size_t size, const size_t indices_size, const T *learning_rate, in SparseApplyProximalAdagradUpdate() 84 …CalSparseApplyProximalAdagrad(const size_t size, const size_t indices_size, const T *learning_rate, in CalSparseApplyProximalAdagrad()
|
| /third_party/mindspore/mindspore/nn/ |
| D | dynamic_lr.py | 65 def _check_inputs(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_stair): argument 76 def exponential_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_sta… argument 119 def natural_exp_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_sta… argument 163 def inverse_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_stair=F… argument 257 def polynomial_decay_lr(learning_rate, end_learning_rate, total_step, step_per_epoch, decay_epoch, … argument 333 def warmup_lr(learning_rate, total_step, step_per_epoch, warmup_epoch): argument
|
| D | learning_rate_schedule.py | 48 def _check_inputs(learning_rate, decay_rate, decay_steps, is_stair, cls_name): argument 107 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument 174 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument 242 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument 376 …def __init__(self, learning_rate, end_learning_rate, decay_steps, power, update_decay_steps=False): argument 453 def __init__(self, learning_rate, warmup_steps): argument
|
| /third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/sponge/nvtit/ |
| D | md_iteration_gradient_descent_impl.cu | 22 const float learning_rate) { in MD_Iteration_Gradient_Descent() 35 …IterationGradientDescent(const int atom_numbers, float *crd, float *frc, const float learning_rate, in MDIterationGradientDescent()
|
| /third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ |
| D | rmsprop_cpu_kernel.cc | 31 float *learning_rate) { in LaunchRMSPropUnuseCenter() 52 … T *mean_gradients, float *momentum, float *learning_rate, float *decay, in LaunchRMSPropUseCenter() 105 float *learning_rate = reinterpret_cast<float *>(inputs[3]->addr); in Launch() local 118 float *learning_rate = reinterpret_cast<float *>(inputs[5]->addr); in Launch() local
|
| /third_party/mindspore/mindspore/nn/optim/ |
| D | optimizer.py | 129 def __init__(self, learning_rate, parameters, weight_decay=0.0, loss_scale=1.0): argument 368 def _preprocess_single_lr(self, learning_rate): argument 392 def _build_single_lr(self, learning_rate, name): argument 439 def _parse_group_params(self, parameters, learning_rate): argument 468 def _init_group_params(self, parameters, learning_rate, weight_decay, grad_centralization): argument 577 def get_lr_value(learning_rate): argument 758 def __init__(self, learning_rate): argument 770 def __init__(self, learning_rate, name): argument
|
| D | rmsprop.py | 26 def _rmsprop_opt_(opt, decay, epsilon, momentum, learning_rate, weight, ms, mom, grad): argument 35 def _centered_rmsprop_opt_(opt, decay, epsilon, momentum, learning_rate, weight, mg, ms, mom, grad): argument 181 def __init__(self, params, learning_rate=0.1, decay=0.9, momentum=0.0, epsilon=1e-10, argument
|
| D | proximal_ada_grad.py | 28 def _tensor_run_opt_with_sparse(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument 37 def _tensor_run_opt(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument 172 def __init__(self, params, accum=0.1, learning_rate=0.001, l1=0.0, l2=0.0, argument
|
| D | momentum.py | 28 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, moment, ps_parameter, cache… argument 151 …def __init__(self, params, learning_rate, momentum, weight_decay=0.0, loss_scale=1.0, use_nesterov… argument
|
| D | sgd.py | 28 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, accum, stat): argument 139 …def __init__(self, params, learning_rate=0.1, momentum=0.0, dampening=0.0, weight_decay=0.0, neste… argument
|
| D | ada_grad.py | 25 def _tensor_run_opt(opt, learning_rate, weight, accum, gradient): argument 149 def __init__(self, params, accum=0.1, learning_rate=0.001, argument
|
| D | ftrl.py | 29 def _tensor_run_opt_with_sparse(opt, spars_opt, push, pull, l1, l2, lr_power, learning_rate, linear, argument 46 def _tensor_run_opt(opt, spars_opt, push, pull, l1, l2, lr_power, learning_rate, linear, argument 197 … def __init__(self, params, initial_accum=0.1, learning_rate=0.001, lr_power=-0.5, l1=0.0, l2=0.0, argument
|
| D | adafactor.py | 30 def _get_lr(step, rms, learning_rate, relative_step, warmup_init, scale_parameter, eps): argument 71 learning_rate, step, grad, param, argument 290 learning_rate=None, argument
|
| /third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/fp32_grad/ |
| D | sgd.cc | 33 int DoSgd(float *weight, float *accumulate, const float *gradient, float learning_rate, float dampe… in DoSgd() 55 int DoSgdInit(float *weight, float *accumulate, float *gradient, float *stat, float learning_rate, … in DoSgdInit() 76 float learning_rate = lr_; in Execute() local 100 float learning_rate = lr_; in ExecuteInit() local 207 float learning_rate = lr_; in OptimizerStep() local
|
| D | apply_momentum.cc | 32 static int DoApplyMomentum(float *weight, float *accumulate, float learning_rate, const float *grad… in DoApplyMomentum() 54 float learning_rate = lr_; in Execute() local 124 float learning_rate = lr_; in OptimizerStep() local
|
| D | adam.cc | 35 … float beta2_power, float eps, float learning_rate, bool nesterov, int start, int end) { in DoAdam() 71 auto learning_rate = lr_; in Execute() local 142 auto learning_rate = lr_; in OptimizerStep() local
|
| /third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/nn/ |
| D | rmsprop_gpu_kernel.h | 47 T *learning_rate = GetDeviceAddress<T>(inputs, 3); in Launch() local 58 T *learning_rate = GetDeviceAddress<T>(inputs, 5); in Launch() local
|
| /third_party/mindspore/tests/mindspore_test_framework/utils/ |
| D | model_util.py | 49 def update_opt_step(learning_rate, batch_size, parameter, gradient): argument 69 def __init__(self, parameters, learning_rate=0.001, batch_size=1): argument
|
| /third_party/mindspore/tests/st/networks/models/resnet50/src_thor/ |
| D | thor.py | 56 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, moment): argument 147 def THOR(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0, batch_size=32, argument 159 …def __init__(self, net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0, batch_… argument
|
| /third_party/mindspore/tests/st/ops/gpu/ |
| D | test_rmsprop.py | 67 learning_rate, decay, momentum, epsilon): argument 75 learning_rate, decay, momentum, epsilon): argument
|