We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 771341a commit 16bdbcdCopy full SHA for 16bdbcd
adam_atan2_pytorch/adopt.py
@@ -98,15 +98,15 @@ def step(
98
next_m = grad.div(v.sqrt().clamp(min = eps)) # they claim that a max(value, eps) performs better than adding the epsilon
99
100
if steps > 1:
101
- m.lerp_(next_m, 1. - beta2)
+ m.lerp_(next_m, 1. - beta1)
102
103
# then update parameters
104
105
p.add_(m, alpha = -lr)
106
107
# update exp grad sq (v)
108
109
- v.lerp_(grad_sq, 1. - beta1)
+ v.lerp_(grad_sq, 1. - beta2)
110
111
# increment steps
112
pyproject.toml
@@ -1,6 +1,6 @@
1
[project]
2
name = "adam-atan2-pytorch"
3
-version = "0.1.2"
+version = "0.1.4"
4
description = "Adam-atan2 for Pytorch"
5
authors = [
6
{ name = "Phil Wang", email = "lucidrains@gmail.com" }
0 commit comments