Skip to content

Commit c29de13

Browse files
cccclaifacebook-github-bot
authored andcommitted
Format llama read me page (#12782)
Summary: Current rendering is a bit off, it's supposed to be text but rendered as code before: {F1980589435} after: {F1980589702} Differential Revision: D78847177
1 parent ef10a35 commit c29de13

File tree

1 file changed

+4
-3
lines changed

1 file changed

+4
-3
lines changed

examples/models/llama/README.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -220,14 +220,15 @@ You can export and run the original Llama 3 8B instruct model.
220220
```
221221
python -m extension.llm.export.export_llm \
222222
--config examples/models/llama/config/llama_q8da4w.yaml
223-
+base.model_clas="llama3"
223+
+base.model_class="llama3"
224224
+base.checkpoint=<consolidated.00.pth.pth> \
225225
+base.params=<params.json>
226226
```
227-
Due to the larger vocabulary size of Llama 3, we recommend quantizing the embeddings with `quantization.embedding_quantize=\'4,32\'` as shown above to further reduce the model size.
228227

228+
Due to the larger vocabulary size of Llama 3, we recommend quantizing the embeddings with `quantization.embedding_quantize=\'4,32\'` as shown above to further reduce the model size.
229229

230-
If you're interested in deploying on non-CPU backends, [please refer the non-cpu-backend section](non_cpu_backends.md)
230+
231+
If you're interested in deploying on non-CPU backends, [please refer the non-cpu-backend section](non_cpu_backends.md)
231232

232233
## Step 3: Run on your computer to validate
233234

0 commit comments

Comments
 (0)