Unable to get good results with knowledge distillation after the update of online distillation #2957
GrassHeadd
started this conversation in
General
Replies: 1 comment 1 reply
-
Hey, could you clarify what you mean by "I came into many issues with regards to the compatibility of the example dataset and also the file."? What was step 1 for? Was there a specific error that led to that change?
Thanks for the feedback. Did you mean this is a new issue for online distillation only? If you were to go with offline distillation, how was it? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Issue
While trying to replicate the author's knowledge distillation, I came into many issues with regards to the compatibility of the example dataset and also the file. After fixing the issue, I was able to train the model, yet, i was not able to achieve satisfiable results, in fact, the result are very very bad. The model just repeatedly output the last generated token like a broken record. E.g. When prompted with "how are you?", it just outputs "? ? ? ? ? ..."
Expected behaviour:
Model able to train a model that is able to work well
Actual behaviour:
Model only repeatedly outputs the last token
Steps to replicate
axolotl/src/axolotl/prompt_strategies/__init__.py
file'sinto
axolotl train kd_test_config.yml
in the terminalBeta Was this translation helpful? Give feedback.
All reactions