SWA callback blows up memory requirement during the batchnorm update run #9469
Unanswered
harryle95
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment
-
There is issue #9441 regarding this problem |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I was using the previous version of lightning and SWA worked fine. However, when I updated to the latest version of lightning, the batchnorm forward pass run exceeds my GPU memory cap. I suspect this is due to gradient accumulation (I did not use any) in the source code. Can anyone offer any suggestion to address this, I can't revert to prev version due to dependencies.
Beta Was this translation helpful? Give feedback.
All reactions