We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 1e67fa9 commit 46826b4Copy full SHA for 46826b4
pytorch_optimizer/adan.py
@@ -115,9 +115,9 @@ def step(self, closure: CLOSURE = None) -> LOSS:
115
exp_avg_var.mul_(beta2).add_(grad_diff, alpha=1.0 - beta2)
116
exp_avg_nest.mul_(beta3).add_((grad + beta2 * grad_diff) ** 2, alpha=1.0 - beta3)
117
118
- step_size = group['lr'] / math.sqrt(exp_avg_nest + self.eps)
+ step_size = group['lr'] / exp_avg_nest.add_(self.eps).sqrt_()
119
120
- p.sub_(exp_avg + beta2 * exp_avg_var, alpha=step_size)
+ p.sub_(step_size * (exp_avg + beta2 * exp_avg_var))
121
p.div_(1.0 + group['weight_decay'])
122
123
return loss
0 commit comments