You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+6-3Lines changed: 6 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -10,7 +10,7 @@
10
10
11
11
**pytorch-optimizer** is optimizer & lr scheduler collections in PyTorch.
12
12
I just re-implemented (speed & memory tweaks, plug-ins) the algorithm while based on the original paper. Also, It includes useful and practical optimization ideas.
13
-
Currently, **61 optimizers (+ `bitsandbytes`)**, **10 lr schedulers**, and **13 loss functions** are supported!
13
+
Currently, **62 optimizers (+ `bitsandbytes`)**, **10 lr schedulers**, and **13 loss functions** are supported!
14
14
15
15
Highly inspired by [pytorch-optimizer](https://github.com/jettify/pytorch-optimizer).
16
16
@@ -27,9 +27,11 @@ So, please double-check the license before using it at your work.
27
27
$ pip3 install pytorch-optimizer
28
28
```
29
29
30
-
From `pytorch-optimizer v2.12.0`, you can install and import `bitsandbytes` optimizers.
30
+
From `v2.12.0`, you can install and import `bitsandbytes` optimizers.
31
31
please check [the requirements](https://github.com/TimDettmers/bitsandbytes?tab=readme-ov-file#tldr) before installing it.
32
32
33
+
From `v3.0.0`, drop `Python 3.7` support. However, you can still use this package with `Python 3.7` by installing with `--ignore-requires-python` option.
| LOMO |*Full Parameter Fine-tuning for Large Language Models with Limited Resources*|[github](https://github.com/OpenLMLab/LOMO)|<https://arxiv.org/abs/2306.09782>|[cite](https://github.com/OpenLMLab/LOMO#citation)|
157
159
| Tiger |*A Tight-fisted Optimizer, an optimizer that is extremely budget-conscious*|[github](https://github.com/bojone/tiger)||[cite](https://github.com/bojone/tiger/blob/main/README_en.md#citation)|
158
160
| CAME |*Confidence-guided Adaptive Memory Efficient Optimization*|[github](https://github.com/huawei-noah/Pretrained-Language-Model/tree/master/CAME)|<https://aclanthology.org/2023.acl-long.243/>|[cite](https://github.com/huawei-noah/Pretrained-Language-Model/tree/master/CAME#citation)|
159
-
| WSAM |*Sharpness-Aware Minimization Revisited: Weighted Sharpness as a Regularization Term*|[github](https://github.com/intelligent-machine-learning/dlrover/blob/master/atorch/atorch/optimizers/wsam.py)|<https://arxiv.org/abs/2305.15817>|[cite](https://github.com/intelligent-machine-learning/dlrover)|
161
+
| WSAM |*Sharpness-Aware Minimization Revisited: Weighted Sharpness as a Regularization Term*|[github](https://github.com/intelligent-machine-learning/dlrover/blob/master/atorch/atorch/optimizers/wsam.py)|<https://arxiv.org/abs/2305.15817>|[cite](https://github.com/intelligent-machine-learning/dlrover)|
162
+
| Aida |*A DNN Optimizer that Improves over AdaBelief by Suppression of the Adaptive Stepsize Range*|[github](https://github.com/guoqiang-zhang-x/Aida-Optimizer)|<https://arxiv.org/abs/2203.13273>|[cite](https://github.com/guoqiang-zhang-x/Aida-Optimizer?tab=readme-ov-file#1-brief-description-of-aida)|
0 commit comments