Momentum#
Momentum accelerates each update step by accumulating velocity from past updates and adding a factor of the previous velocity to the current step. Momentum can help speed up training and escape bad local minima when compared with Stochastic Gradient Descent.
Parameters#
# | Name | Default | Type | Description |
---|---|---|---|---|
1 | rate | 0.001 | float | The learning rate that controls the global step size. |
2 | decay | 0.1 | float | The decay rate of the accumulated velocity. |
3 | lookahead | false | bool | Should we employ Nesterov's lookahead (NAG) when updating the parameters? |
Example#
use Rubix\ML\NeuralNet\Optimizers\Momentum;
$optimizer = new Momentum(0.01, 0.1, true);
References#
Last update:
2021-07-03