Home
last modified time | relevance | path

Searched defs:learning_rate (Results 1 – 25 of 109) sorted by relevance

12345

/third_party/mindspore/mindspore-src/source/mindspore/ccsrc/plugin/device/gpu/kernel/cuda_impl/cuda_ops/
Dadagrad_v2_impl.cu32 const S *learning_rate, const T *gradient) { in ApplyAdagradV2Kernel()
43 const half *learning_rate, const half *gradient) { in ApplyAdagradV2Kernel()
54 const float *learning_rate, const half *gradient) { in ApplyAdagradV2Kernel()
65 const double *learning_rate, const half *gradient) { in ApplyAdagradV2Kernel()
76 const half *learning_rate, const double *gradient) { in ApplyAdagradV2Kernel()
87 const float *learning_rate, const double *gradient) { in ApplyAdagradV2Kernel()
98 const half *learning_rate, const float *gradient) { in ApplyAdagradV2Kernel()
109 const double *learning_rate, const float *gradient) { in ApplyAdagradV2Kernel()
120 const S *learning_rate, const T *gradient) { in ApplyAdagradV2Kernel_()
128 const half *learning_rate, const half *gradient) { in ApplyAdagradV2Kernel_()
[all …]
Dadagrad_impl.cu84 …lobal__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const S *learning_rate, in ApplyAdagradKernel()
95 …l__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const float *learning_rate, in ApplyAdagradKernel()
106 …al__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const half *learning_rate, in ApplyAdagradKernel()
117 …__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const double *learning_rate, in ApplyAdagradKernel()
128 …__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const double *learning_rate, in ApplyAdagradKernel()
139 …__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const double *learning_rate, in ApplyAdagradKernel()
150 …al__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const half *learning_rate, in ApplyAdagradKernel()
161 …l__ void ApplyAdagradKernel(const size_t size, const bool update_slots, const float *learning_rate, in ApplyAdagradKernel()
172 cudaError_t ApplyAdagrad(const size_t size, const bool update_slots, const S *learning_rate, const … in ApplyAdagrad()
Dadadelta_impl.cu31 __global__ void ApplyAdadeltaKernal(const size_t size, const S *learning_rate, const S *rho, const … in ApplyAdadeltaKernal()
44 __global__ void ApplyAdadeltaKernal(const size_t size, const float *learning_rate, const float *rho, in ApplyAdadeltaKernal()
60 __global__ void ApplyAdadeltaKernal(const size_t size, const float *learning_rate, const float *rho, in ApplyAdadeltaKernal()
75 __global__ void ApplyAdadeltaKernal(const size_t size, const half *learning_rate, const half *rho, … in ApplyAdadeltaKernal()
91 __global__ void ApplyAdadeltaKernal(const size_t size, const float *learning_rate, const float *rho, in ApplyAdadeltaKernal()
108 __global__ void ApplyAdadeltaKernal(const size_t size, const half *learning_rate, const half *rho, … in ApplyAdadeltaKernal()
123 __global__ void ApplyAdadeltaKernal(const size_t size, const half *learning_rate, const half *rho, … in ApplyAdadeltaKernal()
140 __global__ void ApplyAdadeltaKernal(const size_t size, const half *learning_rate, const half *rho, … in ApplyAdadeltaKernal()
157 __global__ void ApplyAdadeltaKernal(const size_t size, const float *learning_rate, const float *rho, in ApplyAdadeltaKernal()
171 __global__ void ApplyAdadeltaKernal(const size_t size, const float *learning_rate, const float *rho, in ApplyAdadeltaKernal()
[all …]
Dapply_add_sign_impl.cu26 …__ void ApplyAddSignKernel(const size_t size, T *variable, T *accumulation, const S *learning_rate, in ApplyAddSignKernel()
36 …plyAddSignKernel(const size_t size, half *variable, half *accumulation, const float *learning_rate, in ApplyAddSignKernel()
51 …yAddSignKernel(const size_t size, float *variable, float *accumulation, const float *learning_rate, in ApplyAddSignKernel()
63 …lyAddSignKernel(const size_t size, float *variable, float *accumulation, const half *learning_rate, in ApplyAddSignKernel()
75 …lyAddSignKernel(const size_t size, float *variable, float *accumulation, const half *learning_rate, in ApplyAddSignKernel()
88 …pplyAddSignKernel(const size_t size, half *variable, half *accumulation, const half *learning_rate, in ApplyAddSignKernel()
101 cudaError_t ApplyAddSign(const size_t size, T *variable, T *accumulation, const S *learning_rate, c… in ApplyAddSign()
Dapply_power_sign_impl.cu42 …_ void ApplyPowerSignKernel(const size_t size, T *variable, T *accumulation, const S learning_rate, in ApplyPowerSignKernel()
52 …lyPowerSignKernel(const size_t size, half *variable, half *accumulation, const float learning_rate, in ApplyPowerSignKernel()
67 …PowerSignKernel(const size_t size, float *variable, float *accumulation, const float learning_rate, in ApplyPowerSignKernel()
79 …plyPowerSignKernel(const size_t size, half *variable, half *accumulation, const half learning_rate, in ApplyPowerSignKernel()
92 …yPowerSignKernel(const size_t size, float *variable, float *accumulation, const half learning_rate, in ApplyPowerSignKernel()
105 …yPowerSignKernel(const size_t size, float *variable, float *accumulation, const half learning_rate, in ApplyPowerSignKernel()
118 cudaError_t ApplyPowerSign(const size_t size, T *variable, T *accumulation, const S learning_rate, … in ApplyPowerSign()
Dadamax_impl.cu42 __global__ void ApplyAdamaxKernal(const size_t size, const S *b1_power, const S *learning_rate, con… in ApplyAdamaxKernal()
53 …id ApplyAdamaxKernal(const size_t size, const float *b1_power, const float *learning_rate, const f… in ApplyAdamaxKernal()
66 …id ApplyAdamaxKernal(const size_t size, const float *b1_power, const float *learning_rate, const f… in ApplyAdamaxKernal()
78 …void ApplyAdamaxKernal(const size_t size, const half *b1_power, const half *learning_rate, const h… in ApplyAdamaxKernal()
91 …void ApplyAdamaxKernal(const size_t size, const half *b1_power, const half *learning_rate, const h… in ApplyAdamaxKernal()
104 …id ApplyAdamaxKernal(const size_t size, const float *b1_power, const float *learning_rate, const f… in ApplyAdamaxKernal()
116 …id ApplyAdamaxKernal(const size_t size, const float *b1_power, const float *learning_rate, const f… in ApplyAdamaxKernal()
129 …id ApplyAdamaxKernal(const size_t size, const float *b1_power, const float *learning_rate, const f… in ApplyAdamaxKernal()
141 …void ApplyAdamaxKernal(const size_t size, const half *b1_power, const half *learning_rate, const h… in ApplyAdamaxKernal()
154 …void ApplyAdamaxKernal(const size_t size, const half *b1_power, const half *learning_rate, const h… in ApplyAdamaxKernal()
[all …]
Drmsprop_impl.cu71 …__ void RmsPropKernel(const size_t batch_size, const size_t input_elements, const T *learning_rate, in RmsPropKernel()
87 const Complex<double> *learning_rate, const Complex<double> *decay, in RmsPropKernel()
102 …opKernel(const size_t batch_size, const size_t input_elements, const Complex<float> *learning_rate, in RmsPropKernel()
117 cudaError_t RmsProp(const size_t batch_size, const size_t input_elements, const T *learning_rate, c… in RmsProp()
127 …d RmsPropCenterKernel(const size_t batch_size, const size_t input_elements, const T *learning_rate, in RmsPropCenterKernel()
145 …sPropCenterKernel(const size_t batch_size, const size_t input_elements, const float *learning_rate, in RmsPropCenterKernel()
163const Complex<double> *learning_rate, const Complex<double> *decay, in RmsPropCenterKernel()
184const Complex<float> *learning_rate, const Complex<float> *decay, in RmsPropCenterKernel()
204 …msPropCenterKernel(const size_t batch_size, const size_t input_elements, const half *learning_rate, in RmsPropCenterKernel()
224 …PropCenter(const size_t batch_size, const size_t input_elements, const T *learning_rate, const T *… in RmsPropCenter()
Dadam_impl.cu32 … const T *beta2_power, const T *learning_rate, const T *beta1, const T *beta2, in ApplyAdamKernel()
51 …__ void AdamWeightDecayKernel(const size_t size, const float *gradient, const float *learning_rate, in AdamWeightDecayKernel()
65 …l__ void AdamWeightDecayKernel(const size_t size, const half *gradient, const float *learning_rate, in AdamWeightDecayKernel()
79 …l__ void AdamWeightDecayKernel(const size_t size, const half *gradient, const float *learning_rate, in AdamWeightDecayKernel()
98 … const T *beta2_power, const T *learning_rate, const T *beta1, const T *beta2, const T *epsilon, in ApplyAdam()
106 cudaError_t AdamWeightDecayOp(const size_t size, const S *gradient, const float *learning_rate, con… in AdamWeightDecayOp()
Dmomentum_impl.cu197 …r_t MomentumUpdateVariable(const size_t size, T *variable, T *accumulation, const S *learning_rate, in MomentumUpdateVariable()
211const S *learning_rate, const G *gradient, const S *momentum, in FusedWeightDecayScaleMomentum()
219 …sedScaleMomentum(const size_t size, S *scale, T *variable, T *accumulation, const S *learning_rate, in FusedScaleMomentum()
228 const S *learning_rate, const G *gradient, const S *momentum, in FusedWeightDecayMomentum()
238 … T **accumulation, S **learning_rate, G **gradient, S **momentum) { in CombineFusedMomentumScaleKernel()
249 … T **variable, T **accumulation, S **learning_rate, G **gradient, S **momentum, in CombineFusedScaleMomentum()
263S **learning_rate, G **gradient, S **momentum) { in CombineFusedMomentumWeightDecayScaleKernel()
276 S **learning_rate, G **gradient, S **momentum, in CombineFusedWeightDecayScaleMomentum()
Dsparse_apply_centered_rms_prop_impl.cu37T *learning_rate, T *decay_rate, T *epsilon, T *momentum, in SparseApplyCenteredRMSPropUpdate()
59double *learning_rate, double *decay_rate, double *epsilon, in SparseApplyCenteredRMSPropUpdate()
82half *learning_rate, half *decay_rate, half *epsilon, half *momentum, in SparseApplyCenteredRMSPropUpdate()
112T *learning_rate, T *decay_rate, T *epsilon, T *momentum, const T *gradient, in CalSparseApplyCenteredRMSProp()
Dsparse_apply_adagrad_impl.cu31 …d SparseApplyAdagradUpdate(const size_t size, const size_t indices_size, const float learning_rate, in SparseApplyAdagradUpdate()
51 …or_t CalSparseApplyAdagrad(const size_t size, const size_t indices_size, const float learning_rate, in CalSparseApplyAdagrad()
/third_party/mindspore/mindspore-src/source/mindspore/python/mindspore/nn/
Ddynamic_lr.py87 def _check_inputs(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_stair): argument
98 def exponential_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_sta… argument
153 def natural_exp_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_sta… argument
209 def inverse_decay_lr(learning_rate, decay_rate, total_step, step_per_epoch, decay_epoch, is_stair=F… argument
328 def polynomial_decay_lr(learning_rate, end_learning_rate, total_step, step_per_epoch, decay_epoch, … argument
417 def warmup_lr(learning_rate, total_step, step_per_epoch, warmup_epoch): argument
Dlearning_rate_schedule.py51 def _check_inputs(learning_rate, decay_rate, decay_steps, is_stair, cls_name): argument
114 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument
185 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument
257 def __init__(self, learning_rate, decay_rate, decay_steps, is_stair=False): argument
404 …def __init__(self, learning_rate, end_learning_rate, decay_steps, power, update_decay_steps=False): argument
488 def __init__(self, learning_rate, warmup_steps): argument
/third_party/mindspore/mindspore-src/source/mindspore/python/mindspore/nn/optim/
Dftrl.py31 def _tensor_run_opt_with_sparse_dist(opt, spars_opt, l1, l2, lr_power, learning_rate, linear, argument
45 def _apply_map_tensor_ftrl(l1, l2, lr_power, learning_rate, linear, weight, moment, indices, values… argument
83 def _run_map_tensor_opt_with_sparse_dist(opt, spars_opt, l1, l2, lr_power, learning_rate, linear, argument
102 def _tensor_run_opt_dist(opt, spars_opt, l1, l2, lr_power, learning_rate, linear, argument
116 def _tensor_run_opt_with_sparse(opt, spars_opt, l1, l2, lr_power, learning_rate, linear, argument
128 def _run_map_tensor_opt_with_sparse(opt, spars_opt, l1, l2, lr_power, learning_rate, linear, argument
139 def _tensor_run_opt(opt, spars_opt, l1, l2, lr_power, learning_rate, linear, argument
147 def _check_param(initial_accum, learning_rate, lr_power, l1, l2, use_locking, prim_name=None): argument
301 … def __init__(self, params, initial_accum=0.1, learning_rate=0.001, lr_power=-0.5, l1=0.0, l2=0.0, argument
368 def _init_distributed_opts(self, use_locking, learning_rate, l1, l2, lr_power): argument
Dmomentum.py33 def _tensor_run_opt_ext(opt, momentum, learning_rate, gradient, weight, moment): argument
40 def _tensor_run_opt_ext_dist(opt, momentum, learning_rate, gradient, weight, moment, argument
186 …def __init__(self, params, learning_rate, momentum, weight_decay=0.0, loss_scale=1.0, use_nesterov… argument
Drmsprop.py29 def _rmsprop_opt_(opt, decay, epsilon, momentum, learning_rate, weight, ms, mom, grad): argument
42 def _centered_rmsprop_opt_(opt, decay, epsilon, momentum, learning_rate, weight, mg, ms, mom, grad): argument
212 def __init__(self, params, learning_rate=0.1, decay=0.9, momentum=0.0, epsilon=1e-10, argument
Doptimizer.py200 def __init__(self, learning_rate, parameters, weight_decay=0.0, loss_scale=1.0): argument
242 def _init_opt_attrs(self, learning_rate, parameters, weight_decay): argument
540 def _preprocess_single_lr(self, learning_rate): argument
569 def _build_single_lr(self, learning_rate, name): argument
609 def _parse_group_params(self, parameters, learning_rate): argument
640 def _init_group_params(self, parameters, learning_rate, weight_decay, grad_centralization): argument
789 def get_lr_value(learning_rate): argument
994 def __init__(self, learning_rate): argument
1008 def __init__(self, learning_rate, name): argument
Dproximal_ada_grad.py30 def _tensor_run_opt_with_sparse(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument
38 def _tensor_run_opt(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument
191 def __init__(self, params, accum=0.1, learning_rate=0.001, l1=0.0, l2=0.0, argument
Dada_grad.py29 def _tensor_run_opt(opt, learning_rate, weight, accum, gradient): argument
194 def __init__(self, params, accum=0.1, learning_rate=0.001, argument
Dsgd.py31 def _tensor_run_opt_ext(momentum, learning_rate, gradient, weight, accum, stat, opt): argument
160 …def __init__(self, params, learning_rate=0.1, momentum=0.0, dampening=0.0, weight_decay=0.0, neste… argument
Dadadelta.py28 def _tensor_run_opt(opt, rho, epsilon, learning_rate, weight, accum, accum_update, gradient): argument
182 …def __init__(self, params, learning_rate=1.0, rho=0.9, epsilon=1e-6, loss_scale=1.0, weight_decay=… argument
Dadamax.py35 def _tensor_run_opt(opt, beta1, beta2, beta1_power, eps, learning_rate, weight, moment1, moment2, g… argument
188 def __init__(self, params, learning_rate=0.001, beta1=0.9, beta2=0.999, eps=1e-08, argument
/third_party/mindspore/mindspore-src/source/mindspore/lite/src/litert/kernel/cpu/fp32_grad/
Dsgd.cc33 int DoSgd(float *weight, float *accumulate, float *gradient, float learning_rate, float dampening, … in DoSgd()
60 int DoSgdInit(float *weight, float *accumulate, float *gradient, float learning_rate, float moment,… in DoSgdInit()
91 float learning_rate = lr_; in DoExecute() local
115 float learning_rate = lr_; in ExecuteInit() local
231 float learning_rate = lr_; in OptimizerStep() local
Dapply_momentum.cc32 static int DoApplyMomentum(float *weight, float *accumulate, float learning_rate, const float *grad… in DoApplyMomentum()
54 float learning_rate = lr_; in DoExecute() local
129 float learning_rate = lr_; in OptimizerStep() local
/third_party/mindspore/mindspore-src/source/mindspore/ccsrc/plugin/device/cpu/kernel/
Drmsprop_cpu_kernel.cc48float *learning_rate, float *decay, float *momentum, in LaunchRMSPropUnuseCenter()
80 … T *mean_gradients, T *momentum, T *learning_rate, T *decay, in LaunchRMSPropUseCenter()
112std::complex<float> *learning_rate, std::complex<float> *decay, in LaunchRMSPropUseCenter()
145std::complex<double> *learning_rate, std::complex<double> *decay, in LaunchRMSPropUseCenter()
267 float *learning_rate = reinterpret_cast<float *>(inputs[kNumberThree]->device_ptr()); in LaunchKernel() local
283 T *learning_rate = static_cast<T *>(inputs[kNumberFive]->device_ptr()); in LaunchKernel() local

12345