This repository was archived by the owner on Sep 9, 2025. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Skill Knowhow: Refining skills #172
Copy link
Copy link
Open
Labels
good-first-issueGood for newcomersGood for newcomers
Description
We need better documentation on ensuring that the skills you contribute will actually improve the model.
People have reported their generate_.jsonl and train_.jsonl files (as a result of running lab generate) contain relatively inaccurate alternate instruction suggestions. We'll likely get a better skill tree if people learn to look at these outputs, and refine the verbiage of their skills until they're getting strong answers from the teacher model (the one you load up with lab serve before you run lab generate).
@xukai92's commentary:
- how easy/hard to get good synthetic data depends on the q/a themselves. perhaps rewriting or rephrasing could help
- there are some improvements on the method itself we will deploy in the next few days
- potentially switch the model used by lab generate could be helpful, but it requires a better machine. all it needs is an endpoint but instructions to be added (instructions to use a stronger teacher model instructlab#398)
- if you can load
mixtral-8x7b-v0.1.Q4_K_M.gguffrom https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF in lab serve and document how to do this it would be very helpful for others
Tasks
- Add Kai's commentary to the skill info docs
- Suggest in both skill development and triage info docs that changing your wording and rerunning the skill until your
.jsonloutput has improved will be more beneficial to the model
Metadata
Metadata
Assignees
Labels
good-first-issueGood for newcomersGood for newcomers