Skip to content

Commit b9c151f

Browse files
authored
Fix documentation link (#1443)
1 parent e80c3dd commit b9c151f

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

docs/source/openvino/export.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -193,7 +193,7 @@ Models larger than 1 billion parameters are exported to the OpenVINO format with
193193
</Tip>
194194

195195

196-
Besides weight-only quantization, you can also apply full model quantization including activations by setting `--quant-mode` to preffered precision. This will quantize both weights and activations of Linear, Convolutional and some other layers to selected mode. Please see example below.
196+
Besides weight-only quantization, you can also apply full model quantization including activations by setting `--quant-mode` to preferred precision. This will quantize both weights and activations of Linear, Convolutional and some other layers to selected mode. Please see example below.
197197

198198
```bash
199199
optimum-cli export openvino -m openai/whisper-large-v3-turbo --quant-mode int8 --dataset librispeech --num-samples 32 --smooth-quant-alpha 0.9 ./whisper-large-v3-turbo

docs/source/openvino/inference.mdx

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -171,7 +171,7 @@ model = OVModelForSequenceClassification.from_pretrained(model_id, device="gpu",
171171

172172
You can also apply fp16, 8-bit or 4-bit weight compression on the Linear, Convolutional and Embedding layers when loading your model to reduce the memory footprint and inference latency.
173173

174-
For more information on the quantization parameters checkout the [documentation](optimziation#weight-only-quantization).
174+
For more information on the quantization parameters checkout the [documentation](optimization#weight-only-quantization).
175175

176176
<Tip warning={true}>
177177

0 commit comments

Comments
 (0)