Skip to content

Commit 12e51c7

Browse files
committed
Fixed ref links
1 parent c97bd09 commit 12e51c7

File tree

2 files changed

+10
-7
lines changed

2 files changed

+10
-7
lines changed

examples/openvino/llama/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -2,13 +2,13 @@
22
# Export Llama with OpenVINO Backend
33

44
## Download the Model
5-
Follow the [instructions](../../examples/models/llama#step-2-prepare-model) to download the required model files. Export Llama with OpenVINO backend is only verified with Llama-3.2-1B variants at this time.
5+
Follow the [instructions](../../../examples/models/llama/README.md#step-2-prepare-model) to download the required model files. Export Llama with OpenVINO backend is only verified with Llama-3.2-1B variants at this time.
66

77
## Environment Setup
8-
Follow the [instructions](../../backends/openvino/README.md) of **Prerequisites** and **Setup** in `backends/openvino/README.md` to set up the OpenVINO backend.
8+
Follow the [instructions](../../../backends/openvino/README.md) of **Prerequisites** and **Setup** in `backends/openvino/README.md` to set up the OpenVINO backend.
99

1010
## Export the model:
11-
Navigate into `<executorch_root>/examples/openvino/llama` and execute the commands below to export the model. Update the model file paths to match the location where your model is downloaded. Replace device with the target hardware you want to compile the model for (`CPU`, `GPU`, or `NPU`). The exported model will be generated in the same directory with the filename `llama3_2.pte`.
11+
Navigate into `<executorch_root>/examples/openvino/llama` and execute the commands below to export the model. Update the model file paths to match the location where your model is downloaded. Replace device with the target hardware you want to compile the model for (`CPU`, `GPU`, or `NPU`). The exported model will be generated in the same directory with the filename `llama3_2_ov.pte`. For modifying the output name, change `output_name` in `llama3_2_ov_4wo.yaml` file under `export`.
1212

1313
```
1414
LLAMA_CHECKPOINT=<path/to/model/folder>/consolidated.00.pth

examples/openvino/llama/llama3_2_ov_4wo.yaml

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2,17 +2,20 @@ base:
22
metadata: '{"get_bos_id":128000, "get_eos_ids":[128009, 128001]}'
33

44
model:
5-
use_kv_cache: True
5+
use_kv_cache: true
66
dtype_override: fp32
7-
enable_dynamic_shape: False
7+
enable_dynamic_shape: false
88

99
quantization:
1010
pt2e_quantize: "openvino_4wo"
1111
group_size: 128
1212

13+
export:
14+
output_name: "llama3_2_ov.pte"
15+
1316
backend:
1417
openvino:
15-
enabled: True
18+
enabled: true
1619

1720
debug:
18-
verbose: True
21+
verbose: false

0 commit comments

Comments
 (0)