pytorch-optimizer v3.6.1
Change Log
Feature
- Implement more cooldown types for WSD learning rate scheduler. (#382, #386)
- Implement
AdamWSNoptimizer. (#387, #389) - Implement
AdamCoptimizer. (#388, #390)
Update
- Change the default range of the
betaparameter from[0, 1]to[0, 1). (#392)
Fix
- Fix to use
momentum bufferinstead of the gradient to calculate LMO. (#385)