T5 FineTuning freezes on 0% Validation Sanity Check #8543
Unanswered
prikmm
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment 1 reply
-
Dear @prikmm, TPU have a weird behaviour with sharing parameters. Basically, it works only after the weights are moved on TPU. If you do it before, it won't work. This warning is telling you that it detected that some parameters are being tied on You need to add the logic too Best, |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I am currently trying to finetune T5 model for Summarization task. I refactored the code present in this T5 Fine Tuning Notebook according to latest API, and my data. But, I am experiencing freezing of model training on the begin of Validation Sanity Check.
I get this warning -
I have created a basic reproducible notebook on XSumm dataset.
The training starts perfectly when using a GPU on both kaggle and colab. The freezing occurs when using TPU only on both Kaggle and colab.
Beta Was this translation helpful? Give feedback.
All reactions