Yogi
optimizer_yogi( learning_rate = 0.01, beta1 = 0.9, beta2 = 0.999, epsilon = 0.001, l1_regularization_strength = 0, l2_regularization_strength = 0, initial_accumulator_value = 1e-06, activation = "sign", name = "Yogi", clipnorm = NULL, clipvalue = NULL, decay = NULL, lr = NULL )
| learning_rate | A Tensor or a floating point value. The learning rate. |
|---|---|
| beta1 | A float value or a constant float tensor. The exponential decay rate for the 1st moment estimates. |
| beta2 | A float value or a constant float tensor. The exponential decay rate for the 2nd moment estimates. |
| epsilon | A constant trading off adaptivity and noise. |
| l1_regularization_strength | A float value, must be greater than or equal to zero. |
| l2_regularization_strength | A float value, must be greater than or equal to zero. |
| initial_accumulator_value | The starting value for accumulators. Only positive values are allowed. |
| activation | Use hard sign or soft tanh to determin sign. |
| name | Optional name for the operations created when applying gradients. Defaults to "Yogi". |
| clipnorm | is clip gradients by norm. |
| clipvalue | is clip gradients by value. |
| decay | is included for backward compatibility to allow time inverse decay of learning rate. |
| lr | is included for backward compatibility, recommended to use learning_rate instead. |
Optimizer for use with `keras::compile()`