Home
last modified time | relevance | path

Searched defs:learning_rate (Results 1 – 25 of 36) sorted by relevance

12

/external/tensorflow/tensorflow/python/training/
Dlearning_rate_decay.py31 def exponential_decay(learning_rate, argument
183 def polynomial_decay(learning_rate, argument
284 def natural_exp_decay(learning_rate, argument
372 def inverse_time_decay(learning_rate, argument
455 def cosine_decay(learning_rate, global_step, decay_steps, alpha=0.0, name=None): argument
518 def cosine_decay_restarts(learning_rate, argument
595 def linear_cosine_decay(learning_rate, argument
680 def noisy_linear_cosine_decay(learning_rate, argument
Dadagrad_test.py53 learning_rate = lambda: 3.0 function
326 learning_rate = lambda: 3.0 function
Dgradient_descent.py34 def __init__(self, learning_rate, use_locking=False, name="GradientDescent"): argument
Dproximal_gradient_descent.py41 def __init__(self, learning_rate, l1_regularization_strength=0.0, argument
Dmomentum.py46 def __init__(self, learning_rate, momentum, argument
Dadadelta.py38 def __init__(self, learning_rate=0.001, rho=0.95, epsilon=1e-8, argument
Dproximal_adagrad.py43 def __init__(self, learning_rate, initial_accumulator_value=0.1, argument
Dadagrad_da.py49 learning_rate, argument
Drmsprop.py66 learning_rate, argument
Dadagrad.py41 def __init__(self, learning_rate, initial_accumulator_value=0.1, argument
Dftrl.py43 learning_rate, argument
/external/tensorflow/tensorflow/examples/tutorials/mnist/
Dmnist.py101 def training(loss, learning_rate): argument
/external/tensorflow/tensorflow/core/kernels/boosted_trees/
Dtraining_ops.cc79 const auto learning_rate = learning_rate_t->scalar<float>()(); in Compute() local
170 OpKernelContext* const context, const float learning_rate, in FindBestSplitsPerNode()
281 const auto learning_rate = learning_rate_t->scalar<float>()(); in Compute() local
390 OpKernelContext* const context, const float learning_rate, in FindBestSplitsPerNode()
/external/tensorflow/tensorflow/python/tpu/
Dtpu_embedding.py64 learning_rate=None, argument
260 def __init__(self, learning_rate, use_gradient_accumulation, argument
289 learning_rate, argument
335 learning_rate, argument
405 learning_rate, argument
480 def __init__(self, learning_rate, clip_weight_min=None, argument
/external/tensorflow/tensorflow/python/keras/mixed_precision/experimental/
Dloss_scale_optimizer.py305 def learning_rate(self): member in LossScaleOptimizer
309 def learning_rate(self, lr): member in LossScaleOptimizer
/external/tensorflow/tensorflow/python/keras/optimizer_v2/
Dadamax.py46 learning_rate=0.001, argument
Dftrl.py58 learning_rate=0.001, argument
Dadadelta.py62 learning_rate=0.001, argument
Dadam.py49 learning_rate=0.001, argument
Dnadam.py65 learning_rate=0.001, argument
Dgradient_descent.py64 learning_rate=0.01, argument
Drmsprop.py64 learning_rate=0.001, argument
Dadagrad.py58 learning_rate=0.001, argument
/external/webrtc/webrtc/base/
Drollingaccumulator.h126 double ComputeWeightedMean(double learning_rate) const { in ComputeWeightedMean()
/external/tensorflow/tensorflow/core/grappler/optimizers/
Dauto_parallel_test.cc41 Output learning_rate = ops::Const(s.WithOpName("learning_rate"), 0.01f, {1}); in TEST_F() local

12