pytorch-optimizer v3.4.0
Change Log
Feature
- Implement
FOCUSoptimizer. (#330, #331) - Implement
PSGD Kronoptimizer. (#336, #337) - Implement
EXAdamoptimizer. (#338, #339)
Update
- Support
OrthoGradvariant toRanger25. (#332)Ranger25optimizer is my experimental-crafted optimizer, which mixes lots of optimizer variants such asADOPT+AdEMAMix+Cautious+StableAdamW+Adam-Atan2+OrthoGrad.
Fix
- Add the missing
stateproperty inOrthoGradoptimizer. (#326, #327) - Add the missing
state_dict, andload_state_dictmethods toTRACandOrthoGradoptimizers. (#332) - Skip when the gradient is sparse in
OrthoGradoptimizer. (#332) - Support alternative precision training in
SOAPoptimizer. (#333) - Store SOAP condition matrices as the dtype of their parameters. (#335)
Contributions
thanks to @Vectorrent, @kylevedder