| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| from collections.abc import Callable, Iterable |
| from typing import TypeVar |
|
|
| import torch |
| from torch.optim import Optimizer |
|
|
| T = TypeVar("T") |
|
|
|
|
| class Novograd(Optimizer): |
| """ |
| Novograd based on `Stochastic Gradient Methods with Layer-wise Adaptive Moments for Training of Deep Networks |
| <https://arxiv.org/pdf/1905.11286.pdf>`_. |
| The code is adapted from the implementations in `Jasper for PyTorch |
| <https://github.com/NVIDIA/DeepLearningExamples/blob/master/PyTorch/SpeechRecognition/Jasper/common/optimizers.py>`_, |
| and `OpenSeq2Seq <https://github.com/NVIDIA/OpenSeq2Seq/blob/master/open_seq2seq/optimizers/novograd.py>`_. |
| |
| Args: |
| params: iterable of parameters to optimize or dicts defining parameter groups. |
| lr: learning rate. Defaults to 1e-3. |
| betas: coefficients used for computing running averages of gradient and its square. Defaults to (0.9, 0.98). |
| eps: term added to the denominator to improve numerical stability. Defaults to 1e-8. |
| weight_decay: weight decay (L2 penalty). Defaults to 0. |
| grad_averaging: gradient averaging. Defaults to ``False``. |
| amsgrad: whether to use the AMSGrad variant of this algorithm from the paper |
| `On the Convergence of Adam and Beyond <https://arxiv.org/pdf/1904.09237.pdf>`_. Defaults to ``False``. |
| """ |
|
|
| def __init__( |
| self, |
| params: Iterable, |
| lr: float = 1e-3, |
| betas: tuple[float, float] = (0.9, 0.98), |
| eps: float = 1e-8, |
| weight_decay: float = 0, |
| grad_averaging: bool = False, |
| amsgrad: bool = False, |
| ): |
| if 0.0 > lr: |
| raise ValueError(f"Invalid learning rate: {lr}") |
| if 0.0 > eps: |
| raise ValueError(f"Invalid epsilon value: {eps}") |
| if not 0.0 <= betas[0] < 1.0: |
| raise ValueError(f"Invalid beta parameter at index 0: {betas[0]}") |
| if not 0.0 <= betas[1] < 1.0: |
| raise ValueError(f"Invalid beta parameter at index 1: {betas[1]}") |
| if 0.0 > weight_decay: |
| raise ValueError(f"Invalid weight_decay value: {weight_decay}") |
| defaults = dict( |
| lr=lr, betas=betas, eps=eps, weight_decay=weight_decay, grad_averaging=grad_averaging, amsgrad=amsgrad |
| ) |
|
|
| super().__init__(params, defaults) |
|
|
| def __setstate__(self, state): |
| super().__setstate__(state) |
| for group in self.param_groups: |
| group.setdefault("amsgrad", False) |
|
|
| def step(self, closure: Callable[[], T] | None = None) -> T | None: |
| """Performs a single optimization step. |
| |
| Arguments: |
| closure: A closure that reevaluates the model and returns the loss. Defaults to ``None``. |
| """ |
| loss = None |
| if closure is not None: |
| loss = closure() |
|
|
| for group in self.param_groups: |
| for p in group["params"]: |
| if p.grad is None: |
| continue |
| grad = p.grad.data |
| if grad.is_sparse: |
| raise RuntimeError("Sparse gradients are not supported.") |
| amsgrad = group["amsgrad"] |
|
|
| state = self.state[p] |
|
|
| |
| if len(state) == 0: |
| state["step"] = 0 |
| |
| state["exp_avg"] = torch.zeros_like(p.data) |
| |
| state["exp_avg_sq"] = torch.zeros([]).to(state["exp_avg"].device) |
| if amsgrad: |
| |
| state["max_exp_avg_sq"] = torch.zeros([]).to(state["exp_avg"].device) |
|
|
| exp_avg, exp_avg_sq = state["exp_avg"], state["exp_avg_sq"] |
| if amsgrad: |
| max_exp_avg_sq = state["max_exp_avg_sq"] |
| beta1, beta2 = group["betas"] |
|
|
| state["step"] += 1 |
|
|
| norm = torch.sum(torch.pow(grad, 2)) |
|
|
| if exp_avg_sq == 0: |
| exp_avg_sq.copy_(norm) |
| else: |
| exp_avg_sq.mul_(beta2).add_(norm, alpha=1 - beta2) |
|
|
| if amsgrad: |
| |
| torch.max(max_exp_avg_sq, exp_avg_sq, out=max_exp_avg_sq) |
| |
| denom = max_exp_avg_sq.sqrt().add_(group["eps"]) |
| else: |
| denom = exp_avg_sq.sqrt().add_(group["eps"]) |
|
|
| grad.div_(denom) |
| if group["weight_decay"] != 0: |
| grad.add_(p.data, alpha=group["weight_decay"]) |
| if group["grad_averaging"]: |
| grad.mul_(1 - beta1) |
| exp_avg.mul_(beta1).add_(grad) |
|
|
| p.data.add_(exp_avg, alpha=-group["lr"]) |
|
|
| return loss |
|
|