Commit ee7a0f7
Enable 16-bit activations in Cadence Quantizer For fully_connected and linear
Summary:
# Context
We currently only support 8-bit for most operators. We would like to add generic ops for 16-bit activations, for the following ops:
- quantized_fully_connected
- quantized_linear
- quantized_conv (all flavors)
- quantized_matmul
# This Diff
Here, we add support for `quantized_linear` and `quantized_fully_connected`. We need to do the following:
1. Allow 16-bit activations in `quantized_fully_connected_out.cpp` and `quantized_linear_out.cpp`.
2. Allow 16-bit activations in `ref_implementations.py`, so tests can run with 16-bit activations to validate the quantization is correct.
3. Add a quantizer(`CadenceWith16BitLinearActivationsQuantizer`) for checking this works and create a unit test.
Differential Revision: D842847941 parent e26670b commit ee7a0f7
File tree
2 files changed
+11
-1
lines changed- backends/cadence/aot
- quantizer
2 files changed
+11
-1
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
338 | 338 | | |
339 | 339 | | |
340 | 340 | | |
| 341 | + | |
| 342 | + | |
| 343 | + | |
| 344 | + | |
| 345 | + | |
| 346 | + | |
| 347 | + | |
| 348 | + | |
| 349 | + | |
| 350 | + | |
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
261 | 261 | | |
262 | 262 | | |
263 | 263 | | |
264 | | - | |
| 264 | + | |
265 | 265 | | |
266 | 266 | | |
267 | 267 | | |
| |||
0 commit comments