Skip to content

Commit a7b7cb7

Browse files
Update distributed.py (bigscience-workshop#275)
1 parent 6a8b2f5 commit a7b7cb7

File tree

1 file changed

+0
-1
lines changed

1 file changed

+0
-1
lines changed

megatron/model/distributed.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,6 @@ def allreduce_gradients(self):
217217
if tp not in buckets:
218218
buckets[tp] = []
219219
buckets[tp].append(param)
220-
param.main_grad = param.grad
221220

222221
# For each bucket, all-reduce and copy all-reduced grads.
223222
for tp in buckets:

0 commit comments

Comments
 (0)