Interleaving training of several models #14229
Unanswered
turian
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am a long-time lightning fan, on the other hand in the past I have gotten bitten by how opinionated it is about training loops. I have a new project in pure torch, and I'm debating whether I can easily migrate it to lightning. I have a related question here: #14228
I have a pretraining model which trains only training data to minimize a self-supervised loss.
I have a downstream model which uses the pretrained model (either frozen or finetuning it), and trains on dev data and evaluates on test data.
I want to explore multiple training strategies, including pretrain then downstream (simple) and interleaved pretraining and downstream training. What are best practices for using lightning in this setting?
Beta Was this translation helpful? Give feedback.
All reactions