Skip to content

Commit ef33fad

Browse files
authored
Revert "Drop radam optimizer. (#377)" (#388)
This reverts commit 4bb5e4b.
1 parent d2ff5c2 commit ef33fad

13 files changed

+238
-0
lines changed

README.rst

Lines changed: 30 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -125,6 +125,9 @@ Supported Optimizers
125125
| `QHM`_ | https://arxiv.org/abs/1810.06801 |
126126
+---------------+--------------------------------------------------------------------------------------------------------------------------------------+
127127
| | |
128+
| `RAdam`_ | https://arxiv.org/abs/1908.03265 |
129+
+---------------+--------------------------------------------------------------------------------------------------------------------------------------+
130+
| | |
128131
| `Ranger`_ | https://medium.com/@lessw/new-deep-learning-optimizer-ranger-synergistic-combination-of-radam-lookahead-for-the-best-of-2dc83f79a48d |
129132
+---------------+--------------------------------------------------------------------------------------------------------------------------------------+
130133
| | |
@@ -765,6 +768,33 @@ QHM
765768
**Reference Code**: https://github.com/facebookresearch/qhoptim
766769

767770

771+
RAdam
772+
-----
773+
774+
+---------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------+
775+
| .. image:: https://raw.githubusercontent.com/jettify/pytorch-optimizer/master/docs/rastrigin_RAdam.png | .. image:: https://raw.githubusercontent.com/jettify/pytorch-optimizer/master/docs/rosenbrock_RAdam.png |
776+
+---------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------+
777+
778+
.. code:: python
779+
780+
import torch_optimizer as optim
781+
782+
# model = ...
783+
optimizer = optim.RAdam(
784+
m.parameters(),
785+
lr= 1e-3,
786+
betas=(0.9, 0.999),
787+
eps=1e-8,
788+
weight_decay=0,
789+
)
790+
optimizer.step()
791+
792+
793+
**Paper**: *On the Variance of the Adaptive Learning Rate and Beyond* (2019) [https://arxiv.org/abs/1908.03265]
794+
795+
**Reference Code**: https://github.com/LiyuanLucasLiu/RAdam
796+
797+
768798
Ranger
769799
------
770800

docs/api.rst

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -97,6 +97,14 @@ QHM
9797
.. autoclass:: torch_optimizer.QHM
9898
:members:
9999

100+
.. _RAdam:
101+
102+
RAdam
103+
-----
104+
105+
.. autoclass:: torch_optimizer.RAdam
106+
:members:
107+
100108
.. _SGDP:
101109

102110
SGDP

docs/index.rst

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -84,6 +84,9 @@ Supported Optimizers
8484
| :ref:`QHM` | https://arxiv.org/abs/1810.06801 |
8585
+-----------------+-------------------------------------------------------------------------------+
8686
| | |
87+
| :ref:`RAdam` | https://arxiv.org/abs/1908.03265 |
88+
+-----------------+-------------------------------------------------------------------------------+
89+
| | |
8790
| :ref:`Ranger` | https://arxiv.org/abs/1908.00700v2 |
8891
+-----------------+-------------------------------------------------------------------------------+
8992
| | |

docs/rastrigin_RAdam.png

726 KB
Loading

docs/rosenbrock_RAdam.png

441 KB
Loading

examples/viz_optimizers.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -173,6 +173,7 @@ def LookaheadYogi(*a, **kw):
173173
(optim.Lamb, -8, -2.9),
174174
(optim.MADGRAD, -8, 0.5),
175175
(optim.NovoGrad, -8, -1.7),
176+
(optim.RAdam, -8, 0.5),
176177
(optim.Yogi, -8, 0.1),
177178
# SGD/Momentum based
178179
(optim.AccSGD, -8, -1.4),

setup.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -57,6 +57,7 @@ def _read_version():
5757
'pid',
5858
'qhadam',
5959
'qhm',
60+
'radam',
6061
'sgdw',
6162
'yogi',
6263
'ranger',

tests/test_basic.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -52,6 +52,7 @@ def build_lookahead(*a, **kw):
5252
{'lr': 2.9, 'betas': (0.9, 0.999), 'grad_averaging': True},
5353
900,
5454
),
55+
(optim.RAdam, {'lr': 0.01, 'betas': (0.9, 0.95), 'eps': 1e-3}, 800),
5556
(optim.SGDW, {'lr': 0.002, 'momentum': 0.91}, 900),
5657
(optim.DiffGrad, {'lr': 0.5}, 500),
5758
(optim.AdaMod, {'lr': 1.0}, 800),

tests/test_optimizer.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -85,6 +85,7 @@ def build_lookahead(*a, **kw):
8585
optim.PID,
8686
optim.QHAdam,
8787
optim.QHM,
88+
optim.RAdam,
8889
optim.Ranger,
8990
optim.RangerQH,
9091
optim.RangerVA,

tests/test_optimizer_with_nn.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -75,6 +75,7 @@ def build_lookahead(*a, **kw):
7575
(optim.PID, {'lr': 0.01, 'weight_decay': 1e-3, 'momentum': 0.1}, 200),
7676
(optim.QHAdam, {'lr': 0.1, 'weight_decay': 1e-3}, 200),
7777
(optim.QHM, {'lr': 0.1, 'weight_decay': 1e-5, 'momentum': 0.2}, 200),
78+
(optim.RAdam, {'lr': 1.0, 'weight_decay': 1e-3}, 200),
7879
(optim.Ranger, {'lr': 0.1, 'weight_decay': 1e-3}, 200),
7980
(optim.RangerQH, {'lr': 0.0124, 'weight_decay': 1e-3}, 1100),
8081
(optim.RangerVA, {'lr': 0.2214, 'weight_decay': 1e-3}, 500),

0 commit comments

Comments
 (0)