Skip to content

Commit 63f6623

Browse files
authored
[Feature] Implement Conda optimizer (#441)
* docs: v3.8.1 changelog * docs: README * feature: Conda optimizer * update: test cases * update: condition * update: condition * [skip ci] docs: update index.md * docs: Conda optimizer * chore: keyword * build(deps): dev deps * ci: bump pyright to 1.1.406 * update: test_galore_projection_type * update: test_galore_methods * update: test_galore_methods * update: test_galore_methods --------- Co-authored-by: kozistr <[email protected]>
1 parent aa70cc9 commit 63f6623

File tree

17 files changed

+435
-236
lines changed

17 files changed

+435
-236
lines changed

.github/workflows/static_analysis.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,4 +24,4 @@ jobs:
2424

2525
- uses: jakebailey/pyright-action@v2
2626
with:
27-
version: 1.1.404
27+
version: 1.1.406

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010

1111
## The reasons why you use `pytorch-optimizer`.
1212

13-
* Wide range of supported optimizers. Currently, **128 optimizers (+ `bitsandbytes`, `qgalore`, `torchao`)**, **16 lr schedulers**, and **13 loss functions** are supported!
13+
* Wide range of supported optimizers. Currently, **129 optimizers (+ `bitsandbytes`, `qgalore`, `torchao`)**, **16 lr schedulers**, and **13 loss functions** are supported!
1414
* Including many variants such as `ADOPT`, `Cautious`, `AdamD`, `StableAdamW`, and `Gradient Centrailiaztion`
1515
* Easy to use, clean, and tested codes
1616
* Active maintenance
@@ -222,6 +222,7 @@ get_supported_optimizers(['adam*', 'ranger*'])
222222
| Refined Schedule-Free | *Through the River: Understanding the Benefit of Schedule-Free Methods for Language Model Training* | | <https://arxiv.org/abs/2507.09846> | [cite](https://ui.adsabs.harvard.edu/abs/2025arXiv250709846S/exportcitation) |
223223
| FriendlySAM | *Friendly Sharpness-Aware Minimization* | [github](https://github.com/nblt/F-SAM) | <https://openaccess.thecvf.com/content/CVPR2024/papers/Li_Friendly_Sharpness-Aware_Minimization_CVPR_2024_paper.pdf> | [cite](https://github.com/nblt/F-SAM?tab=readme-ov-file#citation) |
224224
| AdaGO | *AdaGrad Meets Muon: Adaptive Stepsizes for Orthogonal Updates* | | <https://arxiv.org/abs/2509.02981> | [cite](https://ui.adsabs.harvard.edu/abs/2025arXiv250902981Z/exportcitation) |
225+
| Conda | *Column-Normalized Adam for Training Large Language Models Faster* | [github](https://github.com/jie040109/Conda) | <https://arxiv.org/abs/2509.24218> | [cite](https://ui.adsabs.harvard.edu/abs/2025arXiv250924218W/exportcitation) |
225226

226227
## Supported LR Scheduler
227228

docs/changelogs/v3.8.1.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@
88
* [AdaGrad Meets Muon: Adaptive Stepsizes for Orthogonal Updates](https://arxiv.org/abs/2509.02981)
99
* Update `EXAdam` optimizer to the latest version. (#438)
1010
* Update `EmoNavi` optimizer to the latest version. (#433, #439)
11+
* Implement `Conda` optimizer. (#440, #441)
12+
* [Conda: Column-Normalized Adam for Training Large Language Models Faster](https://arxiv.org/abs/2509.24218)
1113

1214
### Bug
1315

docs/index.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010

1111
## The reasons why you use `pytorch-optimizer`.
1212

13-
* Wide range of supported optimizers. Currently, **128 optimizers (+ `bitsandbytes`, `qgalore`, `torchao`)**, **16 lr schedulers**, and **13 loss functions** are supported!
13+
* Wide range of supported optimizers. Currently, **129 optimizers (+ `bitsandbytes`, `qgalore`, `torchao`)**, **16 lr schedulers**, and **13 loss functions** are supported!
1414
* Including many variants such as `ADOPT`, `Cautious`, `AdamD`, `StableAdamW`, and `Gradient Centrailiaztion`
1515
* Easy to use, clean, and tested codes
1616
* Active maintenance
@@ -222,6 +222,7 @@ get_supported_optimizers(['adam*', 'ranger*'])
222222
| Refined Schedule-Free | *Through the River: Understanding the Benefit of Schedule-Free Methods for Language Model Training* | | <https://arxiv.org/abs/2507.09846> | [cite](https://ui.adsabs.harvard.edu/abs/2025arXiv250709846S/exportcitation) |
223223
| FriendlySAM | *Friendly Sharpness-Aware Minimization* | [github](https://github.com/nblt/F-SAM) | <https://openaccess.thecvf.com/content/CVPR2024/papers/Li_Friendly_Sharpness-Aware_Minimization_CVPR_2024_paper.pdf> | [cite](https://github.com/nblt/F-SAM?tab=readme-ov-file#citation) |
224224
| AdaGO | *AdaGrad Meets Muon: Adaptive Stepsizes for Orthogonal Updates* | | <https://arxiv.org/abs/2509.02981> | [cite](https://ui.adsabs.harvard.edu/abs/2025arXiv250902981Z/exportcitation) |
225+
| Conda | *Column-Normalized Adam for Training Large Language Models Faster* | [github](https://github.com/jie040109/Conda) | <https://arxiv.org/abs/2509.24218> | [cite](https://ui.adsabs.harvard.edu/abs/2025arXiv250924218W/exportcitation) |
225226

226227
## Supported LR Scheduler
227228

docs/optimizer.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -160,6 +160,10 @@
160160
:docstring:
161161
:members:
162162

163+
::: pytorch_optimizer.Conda
164+
:docstring:
165+
:members:
166+
163167
::: pytorch_optimizer.DAdaptAdaGrad
164168
:docstring:
165169
:members:

0 commit comments

Comments
 (0)