Skip to content

Commit 5eb575e

Browse files
committed
Added overview docs.
Change-Id: I8d769d93e9efbe6b64897a1a6d092c567cd54154
1 parent 39d91a2 commit 5eb575e

File tree

1 file changed

+12
-0
lines changed

1 file changed

+12
-0
lines changed

tensorflow_model_optimization/g3doc/guide/combine/collaborative_optimization.md

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -106,6 +106,18 @@ with PQAT and CQAT collaborative optimization paths.
106106
</table>
107107
</figure>
108108

109+
### CQAT and PCQAT results for models clustered per channel
110+
Results below are obtained with the technique [clustering per channel](https://www.tensorflow.org/model_optimization/guide/clustering).
111+
They illustrate that if convolutional layers of the model are clustered per channel, then the model accuracy is higher. If your model has many convolutional layers, then we recommend to cluster per channel. The compression ratio remains the same, but the model accuracy will be higher. The model optimization pipeline is 'clustered -> cluster preserving QAT -> post training quantization, int8' in our experiments.
112+
<figure>
113+
<table class="tableizer-table">
114+
<tr class="tableizer-firstrow"><th>Model</th><th>Clustered -> CQAT, int8 quantized</th><th>Clustered per channel -> CQAT, int8 quantized</th>
115+
<tr><td>DS-CNN-L</td><td>95.949%</td><td> 96.44%</td></tr>
116+
<tr><td>MobileNet-V2</td><td>71.538%</td><td>72.638%</td></tr>
117+
<tr><td>MobileNet-V2 (pruned)</td><td>71.45%</td><td>71.901%</td></tr>
118+
</table>
119+
</figure>
120+
109121
## Examples
110122

111123
For end-to-end examples of the collaborative optimization techniques described

0 commit comments

Comments
 (0)