We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b18d662 commit 1e67fa9Copy full SHA for 1e67fa9
pytorch_optimizer/adan.py
@@ -112,7 +112,7 @@ def step(self, closure: CLOSURE = None) -> LOSS:
112
state['previous_grad'] = grad.clone()
113
114
exp_avg.mul_(beta1).add_(grad, alpha=1.0 - beta1)
115
- exp_avg_var.mul_(beta2).add_(grad_diff, value=1.0 - beta2)
+ exp_avg_var.mul_(beta2).add_(grad_diff, alpha=1.0 - beta2)
116
exp_avg_nest.mul_(beta3).add_((grad + beta2 * grad_diff) ** 2, alpha=1.0 - beta3)
117
118
step_size = group['lr'] / math.sqrt(exp_avg_nest + self.eps)
0 commit comments