Training stuck at 0% while training with bfloat16 precision #14203
Unanswered
y199387
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment 2 replies
-
@y199387 Would you mind sharing your environment detail here to reproduce the behaviour? wget https://raw.githubusercontent.com/Lightning-AI/lightning/master/requirements/collect_env_details.py
python collect_env_details.py |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I am trying to run the PYTORCH LIGHTNING CIFAR10 ~94% BASELINE TUTORIAL using cpu. When I try to switch to bf16 precision through
Trainer(precision='bf16')
, the training % is stuck at 0% and never progresses.Beta Was this translation helpful? Give feedback.
All reactions