We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 59526a2 commit 71da75aCopy full SHA for 71da75a
pytorch_optimizer/optimizer/dadapt.py
@@ -246,7 +246,7 @@ def step(self, closure: CLOSURE = None) -> LOSS:
246
247
248
class DAdaptAdam(Optimizer, BaseOptimizer):
249
- r"""Adam with D-Adaptation. Leave LR set to 1 unless you encounter instability.
+ r"""Adam with D-Adaptation. Leave LR set to 1 unless you encounter instability. This implementation is based on V3.
250
251
:param params: PARAMETERS. iterable of parameters to optimize or dicts defining parameter groups.
252
:param lr: float. learning rate.
0 commit comments