| /third_party/flutter/skia/src/gpu/gradients/generated/ |
| D | GrTextureGradientColorizer.h | 19 static std::unique_ptr<GrFragmentProcessor> Make(sk_sp<GrTextureProxy> gradient) { in Make() 25 TextureSampler gradient; variable 28 GrTextureGradientColorizer(sk_sp<GrTextureProxy> gradient) in GrTextureGradientColorizer()
|
| /third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/gpu/cuda_impl/ |
| D | momentum_impl.cu | 20 … const G *gradient, const S *momentum, bool use_nesterov) { in MomentumUpdateVariableKernel() 35 … const float *learning_rate, const half *gradient, const float *momentum, in MomentumUpdateVariableKernel() 52 … const float *learning_rate, const half *gradient, const float *momentum, in MomentumUpdateVariableKernel() 67 …ariable(const size_t size, T *variable, T *accumulation, const S *learning_rate, const G *gradient, in MomentumUpdateVariable() 75 … T *accumulation, const T *learning_rate, const S *gradient, in FusedMomentumWeightDecayScaleKernel() 86 const T *learning_rate, const S *gradient, const T *momentum, in FusedWeightDecayScaleMomentum() 96 … const T *learning_rate, const S *gradient, const T *momentum) { in FusedMomentumScaleKernel() 105 const S *gradient, const T *momentum, cudaStream_t cuda_stream) { in FusedScaleMomentum() 114 … const T *learning_rate, const S *gradient, const T *momentum) { in FusedWeightDecayMomentumKernel() 124 … const T *learning_rate, const S *gradient, const T *momentum, cudaStream_t cuda_stream) { in FusedWeightDecayMomentum() [all …]
|
| D | adagrad_impl.cu | 33 const G *gradient, in ApplyAdagradKernel() 48 const half *gradient, in ApplyAdagradKernel() 63 const half *gradient, in ApplyAdagradKernel() 78 const float *gradient, in ApplyAdagradKernel() 93 const float *gradient, in ApplyAdagradKernel() 108 const G *gradient, in ApplyAdagrad()
|
| D | adam_impl.cu | 30 __global__ void ApplyAdamKernel(const size_t size, const T *gradient, const T *beta1_power, const T… in ApplyAdamKernel() 44 __global__ void AdamWeightDecayKernel(const size_t size, const T *gradient, const float *learning_r… in AdamWeightDecayKernel() 58 __global__ void AdamWeightDecayKernel(const size_t size, const half *gradient, const float *learnin… in AdamWeightDecayKernel() 72 void ApplyAdam(const size_t size, const T *gradient, const T *beta1_power, const T *beta2_power, co… in ApplyAdam() 78 void AdamWeightDecayOp(const size_t size, const T *gradient, const float *learning_rate, const floa… in AdamWeightDecayOp()
|
| D | adam_weight_decay_impl.cu | 24 T *param, T *gradient) { in AdamWeightDecayKernel() 41 … const float *weight_decay, T *m, T *v, T *param, T *gradient, cudaStream_t stream) { in AdamWeightDecay()
|
| D | ftrl_impl.cu | 50 __global__ void ApplyFtrlKernel(const size_t size, const T *gradient, const T *learning_rate, in ApplyFtrlKernel() 72 void ApplyFtrl(const size_t size, const T *gradient, const T *learning_rate, const T *l1_regulariza… in ApplyFtrl()
|
| D | fake_learned_scale_quant_perlayer_impl.cu | 33 …void FakeLearnedScaleQuantPerLayerGrad(float *grad_input, float *grad_alpha, const float *gradient, in FakeLearnedScaleQuantPerLayerGrad() 79 …arnedScaleQuantPerLayerGrad(float *grad_input, float *grad_alpha, const float *gradient, const int… in CalFakeLearnedScaleQuantPerLayerGrad()
|
| D | sparse_apply_proximal_adagrad_impl.cu | 62 … const T *gradient, const int *indices, T *variable, T *accumulation, in SparseApplyProximalAdagradUpdate() 85 … const T *l1_regularization, const T *l2_regularization, const T *gradient, in CalSparseApplyProximalAdagrad()
|
| D | fake_quant_perlayer_impl.cu | 45 __global__ void FakeQuantPerLayerGrad(const float *input, const float *gradient, float *output, con… in FakeQuantPerLayerGrad() 98 void CalFakeQuantPerLayerGrad(const float *input, const float *gradient, float *output, const int s… in CalFakeQuantPerLayerGrad()
|
| D | sparse_ftrl_impl.cu | 52 __global__ void SparseApplyFtrlKernel(const T *gradient, const S *indices, const int num_index, con… in SparseApplyFtrlKernel() 84 void CalSparseApplyFtrl(const T *gradient, const S *indices, const int num_index, const size_t n_st… in CalSparseApplyFtrl()
|
| /third_party/mindspore/tests/st/ops/graph_kernel/ |
| D | test_fused_adam.py | 47 …def construct(self, beta1, beta2, one_sub_beta_1, one_sub_beta_2, gradient, eps, weight_decay_tens… argument 91 …def construct(self, beta1, beta2, one_sub_beta_1, one_sub_beta_2, gradient, eps, weight_decay_tens… argument 119 def CalFusedAdam(beta1, beta2, one_sub_beta_1, one_sub_beta_2, gradient, eps, weight_decay_tensor, … argument
|
| /third_party/mindspore/tests/st/fl/mobile/src/ |
| D | adam.py | 33 def _update_run_kernel(beta1, beta2, eps, lr, weight_decay, param, m, v, gradient, decay_flags, opt… argument 48 def _update_run_op(beta1, beta2, eps, lr, overflow, weight_decay, param, m, v, gradient, decay_flag… argument 109 … beta2_power, beta1, beta2, eps, lr, gradient, param, m, v, ps_parameter, cache_enable): argument 176 beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, argument 192 def _run_off_load_opt(opt, beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, moment… argument
|
| /third_party/mindspore/tests/st/fl/hybrid_lenet/src/ |
| D | adam.py | 33 def _update_run_kernel(beta1, beta2, eps, lr, weight_decay, param, m, v, gradient, decay_flags, opt… argument 48 def _update_run_op(beta1, beta2, eps, lr, overflow, weight_decay, param, m, v, gradient, decay_flag… argument 109 … beta2_power, beta1, beta2, eps, lr, gradient, param, m, v, ps_parameter, cache_enable): argument 175 beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, argument 191 def _run_off_load_opt(opt, beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, moment… argument
|
| /third_party/mindspore/tests/st/fl/albert/src/ |
| D | adam.py | 34 def _update_run_kernel(beta1, beta2, eps, lr, weight_decay, param, m, v, gradient, decay_flags, opt… argument 49 def _update_run_op(beta1, beta2, eps, lr, overflow, weight_decay, param, m, v, gradient, decay_flag… argument 110 … beta2_power, beta1, beta2, eps, lr, gradient, param, m, v, ps_parameter, cache_enable): argument 176 beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, argument 192 def _run_off_load_opt(opt, beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, moment… argument
|
| /third_party/mindspore/mindspore/lite/src/runtime/kernel/arm/fp32_grad/ |
| D | sgd.cc | 33 int DoSgd(float *weight, float *accumulate, const float *gradient, float learning_rate, float dampe… in DoSgd() 55 int DoSgdInit(float *weight, float *accumulate, float *gradient, float *stat, float learning_rate, … in DoSgdInit() 77 auto gradient = reinterpret_cast<float *>(in_tensors_.at(1)->MutableData()); in Execute() local 101 auto gradient = reinterpret_cast<float *>(in_tensors_.at(1)->MutableData()); in ExecuteInit() local
|
| D | apply_momentum.cc | 32 …lyMomentum(float *weight, float *accumulate, float learning_rate, const float *gradient, float mom… in DoApplyMomentum() 55 auto gradient = reinterpret_cast<float *>(in_tensors_.at(FOURTH_INPUT)->data()); in Execute() local
|
| /third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ |
| D | apply_adagrad_cpu_kernel.cc | 71 const auto *gradient = reinterpret_cast<T *>(inputs[3]->addr); in LaunchKernel() local 92 void ApplyAdagradCPUKernel::LaunchApplyAdagrad(T *var, T *accum, const T *lr, const T *gradient, si… in LaunchApplyAdagrad()
|
| D | adam_weight_decay_cpu_kernel.cc | 46 T *gradient = reinterpret_cast<T *>(inputs[GRAD]->addr); in LaunchAdamWeightDecay() local 77 auto gradient = reinterpret_cast<float *>(inputs[GRAD]->addr); in LaunchAdamWeightDecayNnacl() local
|
| D | adam_cpu_kernel.cc | 42 T *gradient = reinterpret_cast<T *>(inputs[GRAD]->addr); in LaunchAdam() local 77 float *gradient = reinterpret_cast<float *>(inputs[GRAD]->addr); in LaunchAdamNnacl() local
|
| D | apply_momentum_cpu_kernel.cc | 42 const auto *gradient = reinterpret_cast<float *>(inputs[3]->addr); in Launch() local
|
| /third_party/skia/tests/ |
| D | ShaderImageFilterTest.cpp | 36 sk_sp<SkShader> gradient = SkGradientShader::MakeRadial( in test_unscaled() local 84 sk_sp<SkShader> gradient = SkGradientShader::MakeRadial( in test_scaled() local
|
| /third_party/mindspore/mindspore/nn/optim/ |
| D | proximal_ada_grad.py | 28 def _tensor_run_opt_with_sparse(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument 37 def _tensor_run_opt(opt, sparse_opt, l1, l2, learning_rate, gradient, weight, accum): argument
|
| D | adam.py | 37 def _update_run_op(beta1, beta2, eps, lr, weight_decay, param, m, v, gradient, decay_flag, optim_fi… argument 94 … beta2_power, beta1, beta2, eps, lr, gradient, param, m, v, ps_parameter, cache_enable): argument 155 beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, argument 171 def _run_off_load_opt(opt, beta1_power, beta2_power, beta1, beta2, eps, lr, gradient, param, moment… argument
|
| D | lamb.py | 39 def _update_run_op(beta1, beta2, eps, global_step, lr, weight_decay, param, m, v, gradient, decay_f… argument 121 def _update_run_op_ascend(beta1, beta2, eps, global_step, lr, weight_decay, param, m, v, gradient, … argument
|
| /third_party/skia/modules/skottie/src/layers/shapelayer/ |
| D | FillStroke.cpp | 131 sk_sp<AnimatablePropertyContainer> gradient) { in AttachFill() 143 sk_sp<AnimatablePropertyContainer> gradient) { in AttachStroke()
|