sftrainer #757
thistleknot
started this conversation in
General
sftrainer
#757
Replies: 1 comment
-
Axolotl is similar to sfttrainer as it is a wrapper around the hf trainer. As for NEFTune, ymmv. We've done some experiments using the neft-v3 branch and while.ot performs better than the mt bench score that trl reported, it still performed worse than a basic fine tune. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I'm reading that sftrainer can be used to train an llm rather than simply trainer
https://huggingface.co/docs/trl/main/en/sft_trainer
there is also this concept about NEFTune which sounds like masking and/or lora dropout
However, I did also want to ask, does this setup support masked models? I know I was able to specify when manually training using my own code, I could set masked llm for llama.
Beta Was this translation helpful? Give feedback.
All reactions