Skip to content

Commit c6fc392

Browse files
Introduce the the notebook of using prune policy for latency improvements
PiperOrigin-RevId: 374525721
1 parent e879836 commit c6fc392

File tree

3 files changed

+542
-0
lines changed

3 files changed

+542
-0
lines changed

tensorflow_model_optimization/g3doc/_book.yaml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -31,6 +31,8 @@ upper_tabs:
3131
path: /model_optimization/guide/pruning/pruning_with_keras
3232
- title: Pruning comprehensive guide
3333
path: /model_optimization/guide/pruning/comprehensive_guide
34+
- title: Pruning for on on-device inference with XNNPACK
35+
path: /model_optimization/guide/pruning/pruning_for_on_device_inference
3436

3537
- heading: Quantization
3638
- title: Quantization aware training overview

tensorflow_model_optimization/g3doc/guide/pruning/index.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,8 @@ fits with your use case.
99
[Pruning with Keras](pruning_with_keras.ipynb) example.
1010
* To quickly find the APIs you need for your use case, see the
1111
[pruning comprehensive guide](comprehensive_guide.ipynb).
12+
* To explore the application of pruning for on-device inference, see the
13+
[Pruning for on on-device inference with XNNPACK](pruning_for_on_device_inference.ipynb).
1214

1315
## Overview
1416

0 commit comments

Comments
 (0)