How does the accumulate_grad_batches affect the lr_scheduler update frequency? #9570
Unanswered
rentainhe
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment 5 replies
-
what do you mean by total steps here? ideally in the codebase schedulers follow their corresponding optimizers. can you share a minimal example? https://colab.research.google.com/drive/1HvWVVTK8j2Nj52qU4Q4YCyzOm0_aLQF3?usp=sharing#scrollTo=4Dk6Ykv8lI7X |
Beta Was this translation helpful? Give feedback.
5 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
It seems when I set accumulate_grad_batches larger than 1, it will affect my lr_scheduler update frequency
I have some problems with
accumulate_grad_batches
, it will auto scale the total steps in training.example
When I set
accumulate_grad_batches=16
, the total steps equals to 4310 and when I changeaccumulate_grad_batches
to 32, the total steps changed to 8620Beta Was this translation helpful? Give feedback.
All reactions