Expand description
Trait and implementations for optimizers.
Structs§
- The AdaDelta optimizer with a learning rate
- Builder for
AdaDelta
. - The Adagrad optimizer.
- Builder for
AdaGrad
. - The Adafactor optimizer.
- Builder for
Adafactor
. - State of the Adafactor optimizer.
- The Adam optimizer.
- Builder for
Adam
. - The AdamW optimizer [1].
- Builder for
AdamW
. - The Adamax optimizer, a variant of Adam based on the infinity norm [1].
- Builder for
Adamax
. - The Lion optimizer [1].
- Builder for
Lion
. - The RMSprop optimizer [1].
- Builder for
RmsProp
. - Stochastic gradient descent optimizer.
- Builder for
Sgd
.
Traits§
- Trait for optimizers.
- Trait for optimizer states.
Functions§
- Clips the global norm of the gradients
Type Aliases§
- Type alias for the beta1 used in Adafactor
Option<f32>
Type alias for the beta1 used in Adafactor builder due to limitation in thegenerate_builder
macroOption<Array>
. Type alias for the learning rate used in Adafactor builder due to limitation in thegenerate_builder
macro- Type alias for the epsilon values used in Adafactor builder
- Type alias for the learning rate used in Adafactor
(f32, f32O)
. Type alias for betas in the Adam/AdamW/Adamax optimizer builders due to limitation in thegenerate_builder
macro- Type alias for clipped gradients that is returned by
clip_grad_norm
. - Type alias for common optimizer state.