Skip to content

Releases: kozistr/pytorch_optimizer

pytorch-optimizer v2.9.0

06 May 08:07
4dbfc23

Choose a tag to compare

Change Log

Feature

Docs

  • Fix readthedocs build issue, #156
  • Move citations into table, #156

Refactor

  • Refactor validation logic, #149, #150
  • Rename amsbound, amsgrad terms into ams_bound, #149
  • Return gradient instead of the parameter, AGC. #149
  • Refactor duplicates (e.g. rectified step size, AMSBound, AdamD, AdaNorm, weight decay) into re-usable functions, #150
  • Move pytorch_optimizer.experimental under pytorch_optimizer.*.experimental

Diff

2.8.0...2.9.0

pytorch-optimizer v2.8.0

29 Apr 08:51
cdfe807

Choose a tag to compare

pytorch-optimizer v2.7.0

26 Apr 06:31
7ded073

Choose a tag to compare

Change Log

Features

Refactor

  • Rename adamd_debias_term to adam_debias, #133
  • Merge the rectified version with the original, #133
    • diffRGrad + diffGrad -> diffGrad
    • RaLamb + Lamb -> Lamb
    • now you can simply use with rectify=True

Fix

  • Fix previous_grad deepcopy issue in Adan optimizer. #134

pytorch-optimizer v2.6.1

22 Apr 12:14
be0351d

Choose a tag to compare

Change Log

Fix

  • variables are not located on the same device with the gradients, #132 (related to #131) (thanks to @Bing-su)
  • fix approximate_sq_grad() in Adafactor optimizer, #132

pytorch-optimizer v2.6.0

22 Apr 07:56
19dcf2b

Choose a tag to compare

Change Log

Feature

  • Implement SM3 optimizer, #130
  • Tweak Scalable Shampoo optimizer, #128, #129
    • implement a new preconditioner type, OUTPUT.
    • optimize speed/memory usage of coupled Newton iteration and power iteration methods.
    • use in-place operation (SQRT-N Grafting).
    • clean-up shampoo_utils more readable.
    • support skip_preconditioning_rank_lt parameter to skip preconditioning in case of the low-rank gradient.
    • set default value for preconditioning_compute_steps to 1000.
    • set default value for start_preconditioning_step to 25.

pytorch-optimizer v2.5.2

11 Apr 13:47
e66435a

Choose a tag to compare

Feature

  • add eps to stabilize optimizing, Nero optimizer. #121

Fix

  • fix Ranger21 not to skip updates when the first parameter doesn't have a gradient, #125, #126 (thanks to @jdb78)
  • fix Lookahead optimizer, #122, #123

Dependency

  • upgrade to Pytorch 2.0, #123

pytorch-optimizer v2.5.1

12 Mar 05:48
df9e78d

Choose a tag to compare

Change Log

Feature

Bug

pytorch-optimizer v2.5.0

15 Feb 05:41
26b8b19

Choose a tag to compare

pytorch-optimizer v2.4.2

10 Feb 10:57
fff34af

Choose a tag to compare

Change Log

Bug

  • Fix to deep-copy inverse preconditioners

Deps

  • Support Pytorch 2.0, #106 (related to #105)

Docs

pytorch-optimizer v2.4.1

06 Feb 06:34
06dce18

Choose a tag to compare

Change Log

Feature

  • Rename the new Shampoo to ScalableShampoo. #103
  • Implement the old(?) version of Shampoo optimizer. #103
  • Support SVD method to calculate the inverse pth root matrix. #103
    • to boost the M^{-1/p} calculation, performs batched SVD when available.
  • Implement AdamS optimizer. #102
  • Support stable weight decay option for Adai optimizer. #102

Bug

  • Fix compute_power_svd() to get a singular value. #104