Yogi
optimizer_yogi(
learning_rate = 0.01,
beta1 = 0.9,
beta2 = 0.999,
epsilon = 0.001,
l1_regularization_strength = 0,
l2_regularization_strength = 0,
initial_accumulator_value = 1e-06,
activation = "sign",
name = "Yogi",
clipnorm = NULL,
clipvalue = NULL,
decay = NULL,
lr = NULL
)
A Tensor or a floating point value. The learning rate.
A float value or a constant float tensor. The exponential decay rate for the 1st moment estimates.
A float value or a constant float tensor. The exponential decay rate for the 2nd moment estimates.
A constant trading off adaptivity and noise.
A float value, must be greater than or equal to zero.
A float value, must be greater than or equal to zero.
The starting value for accumulators. Only positive values are allowed.
Use hard sign or soft tanh to determin sign.
Optional name for the operations created when applying gradients. Defaults to "Yogi".
is clip gradients by norm.
is clip gradients by value.
is included for backward compatibility to allow time inverse decay of learning rate.
is included for backward compatibility, recommended to use learning_rate instead.
Optimizer for use with `keras::compile()`