Quantize already pretrained english recognition model #15072
Replies: 1 comment 1 reply
-
To quantize your already pretrained English recognition model (inference.pdmodel, inference.pdiparams, inference.pdiparams.info) using PaddlePaddle/PaddleOCR and Python 3.8 without additional training, here’s how you should proceed: 1. Install PaddleSlimFirst, install the PaddleSlim package, which provides quantization utilities: pip install paddleslim==2.3.2 2. Understand Quantization Methods
3. Recommended Workflow (with minimal or no training)A. Prepare a Quantization Config File
B. Quantization Process ExampleNavigate to the PaddleOCR project root. Then: python deploy/slim/quantization/quant.py \
-c configs/rec/rec_mv3_none_bilstm_ctc.yml \
-o Global.pretrained_model='./path_to_your_model_directory/' \
Global.save_model_dir='./output/quant_model/' Note:
C. Export the Quantized Inference ModelAfter quantization calibration (or minimal training if required): python deploy/slim/quantization/export_model.py \
-c configs/rec/rec_mv3_none_bilstm_ctc.yml \
-o Global.checkpoints=./output/quant_model/best_accuracy \
Global.save_inference_dir=./output/quant_inference_model This will generate a quantized inference model you can deploy just like the original. 4. Alternative: PaddleLite's Post-Training QuantizationIf you need fully post-training, no-training quantization for fast inference on mobile, consider using PaddleLite’s opt tool with post-training quantization. Convert the model for mobile deployment with INT8 support. 5. Summary of Steps
Tips & References
In summary: Response generated by 🤖 feifei-bot | gpt-4.1 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I already have a pre trained english recognition model having model files names inference.pdmodel,inference.pdiparams,inference.pdiparams.info. Now i want to quantize them without do training,how to achieve this. Im using python 3.8
Beta Was this translation helpful? Give feedback.
All reactions