What is consuming RAM after on_train_epoch_start
but before training_step
#12522
Unanswered
malfonsoarquimea
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment 8 replies
-
it must be the batch that is extracted from your dataloader that is being passed to your |
Beta Was this translation helpful? Give feedback.
8 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi! I am currently facing an issue where a lot of ram is consumed while training. By running the debugger while monitoring ram consumption I managed to find that the ram consumption process happens after on_train_epoch_start callback and before training_step, and I don't have any on_train_batch_start callback defined. I think the issue is related to how I defined my dataset and dataloaders, but I can not find what part of the code is producing the memory consumption since I don't know which piece of the code is called in between on_train_epoch_start and training_stepso I can not debug it.
Do you have some insight on where to look to find where is memory being consumed?
Thanks very much in advance
Beta Was this translation helpful? Give feedback.
All reactions