Skip to content

Commit f17199e

Browse files
authored
Update README.md
1 parent bb3d0cc commit f17199e

File tree

1 file changed

+5
-0
lines changed

1 file changed

+5
-0
lines changed

examples/models/llama/README.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -207,6 +207,9 @@ python -m examples.models.llama.export_llama \
207207
--metadata '{"get_bos_id":128000, "get_eos_ids":[128009, 128001]}'
208208
```
209209

210+
For convenience, here's an already ExecuTorch [exported model](https://huggingface.co/executorch-community/Llama-3.2-1B-Instruct-SpinQuant_INT4_EO8-ET/blob/main/Llama-3.2-1B-Instruct-SpinQuant_INT4_EO8.pte) using [this recipe](https://huggingface.co/executorch-community/Llama-3.2-1B-Instruct-SpinQuant_INT4_EO8-ET/blob/main/Export_Recipe_Llama_3_2_1B_Instruct_SpinQuant_INT4_EO8.ipynb) on Hugging Face.
211+
212+
210213
- To use **QAT+LoRA**, download directly from [Llama website](https://www.llama.com/llama-downloads). The model weights are prequantized and can be exported to `pte` file directly by:
211214

212215
```
@@ -235,6 +238,8 @@ python -m examples.models.llama.export_llama \
235238
--metadata '{"get_bos_id":128000, "get_eos_ids":[128009, 128001]}'
236239
```
237240

241+
For convenience, here's an already ExecuTorch [exported model](https://huggingface.co/executorch-community/Llama-3.2-1B-Instruct-QLORA_INT4_EO8-ET/blob/main/Llama-3.2-1B-Instruct-QLORA_INT4_EO8.pte) using [this recipe](https://huggingface.co/executorch-community/Llama-3.2-1B-Instruct-QLORA_INT4_EO8-ET/blob/main/Export_Recipe_Llama_3_2_1B_Instruct_QLORA_INT4_EO8.ipynb) on Hugging Face.
242+
238243
### Option B: Download and export Llama 3 8B instruct model
239244

240245
You can export and run the original Llama 3 8B instruct model.

0 commit comments

Comments
 (0)