Replies: 1 comment
-
I don't have much experience with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I was wondering how you can resume training from a checkpoint with different hyperparameter config when training with
transformers
library. Given the example below, no matter what you change in thetraining_args
, these will be overridden by whatever training args are saved in the checkpoint. Some things like eval, batch_size and save_steps are overridable if you amend the checkpoint's JSON config, but other hyperparameters are not.Given a non-PEFT model, you could just save the entire model from the checkpoitn, load it up and call
trainer.train()
on it to achieve this behaviour, but given a PEFT setup I'm not sure how you can do this?Beta Was this translation helpful? Give feedback.
All reactions