We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 04b22d5 commit 300c539Copy full SHA for 300c539
pytorch_optimizer/optimizer/madgrad.py
@@ -101,7 +101,6 @@ def step(self, closure: CLOSURE = None) -> LOSS:
101
raise NoSparseGradientError(str(self), note='momentum > 0.0')
102
103
grad_sum_sq, s = state['grad_sum_sq'], state['s']
104
-
105
if weight_decay > 0.0 and not self.decouple_decay:
106
if grad.is_sparse:
107
raise NoSparseGradientError(str(self), note='weight_decay')
0 commit comments