You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Implementation of the proposed Adam-atan2 in Pytorch
1
+
## Adam-atan2 - Pytorch (wip)
2
+
3
+
Implementation of the proposed <ahref="https://arxiv.org/abs/2407.05872">Adam-atan2</a> optimizer in Pytorch
4
+
5
+
A multi-million dollar paper out of google deepmind basically proposes a small change to Adam (using `atan2`) for greater stability
6
+
7
+
## Citations
8
+
9
+
```bibtex
10
+
@inproceedings{Everett2024ScalingEA,
11
+
title = {Scaling Exponents Across Parameterizations and Optimizers},
12
+
author = {Katie Everett and Lechao Xiao and Mitchell Wortsman and Alex Alemi and Roman Novak and Peter J. Liu and Izzeddin Gur and Jascha Narain Sohl-Dickstein and Leslie Pack Kaelbling and Jaehoon Lee and Jeffrey Pennington},
0 commit comments