Right way to load Lora checkpoint to further training #1968
Unanswered
BigDataMLexplorer
asked this question in
Q&A
Replies: 1 comment
-
Hmm, I'm not very familiar with the |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
@BenjaminBossan
Hello, can I ask you plase, how should I specify the Trainer if I want to proceed in training from lora checkpoint with same training arguments the model was trained before checkpoint?
I give the example here:
First training:
Now I load model from checkpoint:
PeftModel.from_pretrained(model, checkpoint_path, is_trainable=True)
I need to specify the Trainer again, if I want to continue from checkpoint and use
trainer.train(resume_from_checkpoint)
What I have to write in Trainer? All the arguemnts like in first training?
I am asking because I want the Trainer to continue with exaxtcly same arguments where the Trainer was stopped in training. For example I want to continue with value of learning rate from checkpoint, which decreased linearly during training. Will it overwrite the learning rate, when I write LR again in Trainer? When I put again the same values of learning rate and other things, what will happen?
And on the other hand, if I want to change
Number of eval and save steps, is it ok to change?
Conlusion:
1)I want just to continue the training where it stopped and change eval and save steps.
2)Is not eneough to write again only the model, training and eval dataset in Trainer ?
If so, what will happen to other stable arguments like fp16, weight decay, batch size?
Thank you very much
Beta Was this translation helpful? Give feedback.
All reactions