You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/lab-4/README.md
+5-12Lines changed: 5 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,15 +4,8 @@ Now that you've set up InstructLab, lets get tuning the Granite Model.
4
4
5
5
## Sanity check
6
6
7
-
First thing you should do is verify you can talk to the Granite model, go ahead and run
8
-
the following commands to verify you can.
9
-
10
-
```bash
11
-
cd instructlab
12
-
source venv/bin/activate
13
-
ilab model chat
14
-
/q
15
-
```
7
+
Take a moment to verify that you are not running `ilab model chat` or `ilab model serve` anywhere,
8
+
it will clash with the following commands with training and tuning the model.
16
9
17
10
The Granite family of foundation models span an increasing variety of modalities, including language, code, time series, and science (e.g., materials) - with much more to come. We're building them with transparency and with focus on fulfilling rigorous enterprise requirements that are emerging for AI. If you'd like to learn more about the models themselves and how we build them, check out Granite Models.
18
11
@@ -39,7 +32,7 @@ Knowledge in the taxonomy tree consists of a few more elements than skills:
39
32
40
33
Format of the `qna.yaml`:
41
34
42
-
-`version`: The chache verion of the qna.yaml file, this is the format of the file used for SDG. The value must be the number 3.
35
+
-`version`: The cache version of the qna.yaml file, this is the format of the file used for SDG. The value must be the number 3.
43
36
-`created_by`: Your GitHub username.
44
37
-`domain`: Specify the category of the knowledge.
45
38
-`seed_examples`: A collection of key/value entries.
@@ -271,7 +264,7 @@ ilab model download
271
264
2) Next we need to generate the data, this is done with the following command:
272
265
273
266
```bash
274
-
ilab data generate --pipeline full --model ~/.cache/instructlab/models/merlinite-7b-lab-Q4_K_M.gguf --model-family mixtral
267
+
ilab data generate
275
268
```
276
269
277
270
This can take some time, take note of the time in the right hand corner, this is building 1000 questions off of your initial 15.
@@ -281,7 +274,7 @@ This takes the granite model, leverages the tokenized version of it, and runs th
281
274
hopefully you can take a lunch break or something while this is running.
282
275
283
276
```bash
284
-
ilab model train --pipeline full --effective-batch-size 64 --is-padding-free false --device mps --max-batch-len 4000 --model-dir instructlab/granite-7b-lab --tokenizer-dir models/granite-7b-lab --model-name instructlab/granite-7b-lab
277
+
ilab model train
285
278
```
286
279
287
280
4) When this is completed, you'll need to test this model, which is the following command:
0 commit comments