We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 88601be commit 053eee5Copy full SHA for 053eee5
bmtrain/optim/optim_manager.py
@@ -136,6 +136,12 @@ def step(self):
136
self.zero_grad()
137
return
138
for optimizer, lr_scheduler in zip(self.optimizers, self.lr_schedulers):
139
+ try:
140
+ check_overflow(optimizer.param_groups)
141
+ except OverflowError:
142
+ has_overflow = True
143
+ print_rank("Gradient overflow, change scale from %lf to %lf" % (self.loss_scale, self.loss_scale / self.loss_scale_factor))
144
+ break
145
if hasattr(optimizer, "_bmtrain_optimizer") and optimizer._bmtrain_optimizer:
146
optimizer.step(scale=self.loss_scale)
147
else:
0 commit comments