Optimizer¶
-
class
slayerSNN.optimizer.
Nadam
(params, lr=0.001, betas=0.9, 0.999, eps=1e-08, weight_decay=0, amsgrad=False)[source]¶ Implements Nadam algorithm. (Modified Adam from PyTorch)
It has been proposed in Incorporating Nesterov Momentum into Adam.
- Arguments:
params
(iterable): iterable of parameters to optimize or dicts defining parameter groups.lr
(float
, optional): learning rate (default: 1e-3).betas
(Tuple[float
,float
], optional): coefficients used for computing running averages of gradient and its square (default: (0.9, 0.999)).eps
(float
, optional): term added to the denominator to improve numerical stability (default: 1e-8).weight_decay
(float
, optional): weight decay (L2 penalty) (default: 0).amsgrad
(boolean
, optional): whether to use the AMSGrad variant of this algorithm from the paper On the Convergence of Adam and Beyond (default: False).