Searched refs:_set_hyper (Results 1 – 9 of 9) sorted by relevance
/external/tensorflow/tensorflow/python/keras/optimizer_v2/ |
D | ftrl.py | 130 self._set_hyper('learning_rate', learning_rate) 131 self._set_hyper('decay', self._initial_decay) 132 self._set_hyper('learning_rate_power', learning_rate_power) 133 self._set_hyper('l1_regularization_strength', l1_regularization_strength) 134 self._set_hyper('l2_regularization_strength', l2_regularization_strength)
|
D | adamax.py | 102 self._set_hyper('learning_rate', kwargs.get('lr', learning_rate)) 103 self._set_hyper('decay', self._initial_decay) 104 self._set_hyper('beta_1', beta_1) 105 self._set_hyper('beta_2', beta_2)
|
D | adadelta.py | 99 self._set_hyper('learning_rate', kwargs.get('lr', learning_rate)) 100 self._set_hyper('decay', self._initial_decay) 101 self._set_hyper('rho', rho)
|
D | adam.py | 143 self._set_hyper('learning_rate', kwargs.get('lr', learning_rate)) 144 self._set_hyper('decay', self._initial_decay) 145 self._set_hyper('beta_1', beta_1) 146 self._set_hyper('beta_2', beta_2)
|
D | gradient_descent.py | 86 self._set_hyper("learning_rate", kwargs.get("lr", learning_rate)) 87 self._set_hyper("decay", self._initial_decay) 94 self._set_hyper("momentum", momentum)
|
D | nadam.py | 98 self._set_hyper('learning_rate', kwargs.get('lr', learning_rate)) 99 self._set_hyper('decay', self._initial_decay) 100 self._set_hyper('beta_1', beta_1) 101 self._set_hyper('beta_2', beta_2)
|
D | rmsprop.py | 108 self._set_hyper("learning_rate", kwargs.get("lr", learning_rate)) 109 self._set_hyper("decay", self._initial_decay) 110 self._set_hyper("rho", rho) 117 self._set_hyper("momentum", momentum)
|
D | adagrad.py | 95 self._set_hyper('learning_rate', kwargs.get('lr', learning_rate)) 96 self._set_hyper('decay', self._initial_decay)
|
D | optimizer_v2.py | 510 def _set_hyper(self, name, value): member in OptimizerV2 561 self._set_hyper(name, value) 1160 setter=RestoredOptimizer._set_hyper # pylint: disable=protected-access
|