nerva_numpy.optimizers
Optimizers used to adjusts the model’s parameters based on the gradients.
Only SGD, Momentum and Nesterov variants are provided. The parser creates factory callables from textual specifications like “Momentum(mu=0.9)”.
Functions
|
Parse a textual optimizer specification into a factory function. |
Classes
|
Combines multiple optimizers to update different parameter groups. |
|
Standard gradient descent optimizer: x -= eta * grad. |
|
Gradient descent with momentum for accelerated convergence. |
|
Nesterov accelerated gradient descent optimizer. |
Minimal optimizer interface used by layers to update parameters. |
- class nerva_numpy.optimizers.Optimizer[source]
Bases:
object
Minimal optimizer interface used by layers to update parameters.
- class nerva_numpy.optimizers.CompositeOptimizer(optimizers: List[Optimizer])[source]
Bases:
Optimizer
Combines multiple optimizers to update different parameter groups.
- class nerva_numpy.optimizers.GradientDescentOptimizer(x, Dx)[source]
Bases:
Optimizer
Standard gradient descent optimizer: x -= eta * grad.
- class nerva_numpy.optimizers.MomentumOptimizer(x, Dx, mu)[source]
Bases:
GradientDescentOptimizer
Gradient descent with momentum for accelerated convergence.
- class nerva_numpy.optimizers.NesterovOptimizer(x, Dx, mu)[source]
Bases:
MomentumOptimizer
Nesterov accelerated gradient descent optimizer.