Skip to content

Commit dd47664

Browse files
committed
up
1 parent 78a8a8d commit dd47664

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

docs/source/quantization-overview.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13,11 +13,11 @@ Quantization in ExecuTorch is backend-specific. Each backend defines how models
1313

1414
The PT2E quantization workflow has three main steps:
1515

16-
1. Create a backend-specific quantizer.
16+
1. Configure a backend-specific quantizer.
1717
2. Prepare, calibrate, convert, and evalute the quantized model in PyTorch
1818
3. Lower the model to the target backend
1919

20-
## 1. Create a Backend-Specific Quantizer
20+
## 1. Configure a Backend-Specific Quantizer
2121

2222
Each backend provides its own quantizer (e.g., XNNPACKQuantizer, CoreMLQuantizer) that defines how quantization should be applied to a model in a way that is compatible with the target hardware.
2323
These quantizers usually support configs that allow users to specify quantization options such as:
@@ -69,4 +69,4 @@ Note that numerics on device can differ those in PyTorch even for unquantized mo
6969

7070
## 3. Lower the model
7171

72-
The final step is to lower the quantized_model to the desired backend, as you would an unquantized one. See backend-specific pages for lowering information.
72+
The final step is to lower the quantized_model to the desired backend, as you would an unquantized one. See [backend-specific pages](backends-overview.md) for lowering information.

0 commit comments

Comments
 (0)