Skip to content

Commit 783664c

Browse files
Enhance documentation and fix minor typos in MCTWrapper and associated tests
1 parent 6e68519 commit 783664c

File tree

9 files changed

+16
-16
lines changed

9 files changed

+16
-16
lines changed

model_compression_toolkit/wrapper/mct_wrapper.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -113,7 +113,7 @@ def __init__(self):
113113
"optimizer", "None", "Optimizer for GPTQ training (low priority)"
114114
"distance_weighting_method", "None", "Distance weighting method for GPTQ (low priority)"
115115
"num_of_images", "5", "Number of images for mixed precision"
116-
"use_hessian_based_scores", "False", "Use Hessian-based scores for mixed precision"
116+
"use_hessian_based_scores", "False", "Use Hessian-based scores for mixed precision (low priority)"
117117
"save_model_path", "'./qmodel.keras' / './qmodel.onnx'", "Path to save quantized model (Keras/Pytorch)"
118118
119119
"""

tests_pytest/keras_tests/e2e_tests/wrapper/test_mct_wrapper_keras_e2e.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -174,7 +174,7 @@ def PTQ_Keras_mixed_precision(float_model: keras.Model) -> Tuple[bool, keras.Mod
174174
['distance_weighting_method', None], # Distance weighting method for mixed precision (low priority).
175175
['num_of_images', 5], # Number of images for mixed precision.
176176
['use_hessian_based_scores', False], # Use Hessian-based sensitivity scores for layer importance (low priority).
177-
['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size.
177+
['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size).
178178
['save_model_path', './qmodel_PTQ_Keras_mixed_precision.keras'] # Path to save the quantized model.
179179
]
180180

@@ -244,12 +244,12 @@ def GPTQ_Keras_mixed_precision(float_model: keras.Model) -> Tuple[bool, keras.Mo
244244
['z_threshold', float('inf')], # Threshold for zero-point quantization (low priority).
245245
['linear_collapsing', True], # Enable linear layer collapsing optimization (low priority).
246246
['residual_collapsing', True], # Enable residual connection collapsing (low priority).
247-
['weights_compression_ratio', 0.75], # Compression ratio for weights.
248247
['n_epochs', 5], # Number of epochs for gradient-based fine-tuning.
249248
['optimizer', None], # Optimizer to use during fine-tuning (low priority).
250249
['distance_weighting_method', None], # Distance weighting method for GPTQ (low priority).
251250
['num_of_images', 5], # Number of images to use for calibration.
252-
['use_hessian_based_scores', False], # Whether to use Hessian-based scores for layer importance.
251+
['use_hessian_based_scores', False], # Whether to use Hessian-based scores for layer importance (low priority).
252+
['weights_compression_ratio', 0.75], # Compression ratio for weights.
253253
['save_model_path', './qmodel_GPTQ_Keras_mixed_precision.keras'] # Path to save the quantized model.
254254
]
255255

tests_pytest/keras_tests/integration_tests/wrapper/test_mct_wrapper_keras_integ.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ def test_quantize_and_export_PTQ_flow(
6767
Mocked Components:
6868
- _get_tpc: TPC configuration
6969
- _select_method: Framework-specific method selection
70-
- _Setting_PTQ: PTQ parameter configuration
70+
- _setting_PTQ: PTQ parameter configuration
7171
- _export_model: Model export functionality
7272
- _post_training_quantization: Actual quantization process
7373
@@ -93,7 +93,7 @@ def test_quantize_and_export_PTQ_flow(
9393

9494
mock_setting_ptq.return_value = {'mock': 'params'}
9595

96-
param_items = [('n_epochs', 10)] # Number of epochs
96+
param_items = [('sdsp_version', '3.14')] # SDSP version for TPC
9797

9898
# Call the method
9999
success, result_model = wrapper.quantize_and_export(

tests_pytest/keras_tests/unit_tests/wrapper/test_mct_wrapper_keras_unit.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -269,9 +269,9 @@ def test_setting_PTQ_mixed_precision(
269269
@patch('model_compression_toolkit.core.CoreConfig')
270270
def test_setting_PTQ(self, mock_core_config: Mock, mock_quant_config: Mock) -> None:
271271
"""
272-
Test _Setting_PTQ method for standard Post-Training Quantization.
272+
Test _setting_PTQ method for standard Post-Training Quantization.
273273
274-
This test verifies that the _Setting_PTQ method correctly configures
274+
This test verifies that the _setting_PTQ method correctly configures
275275
standard Post-Training Quantization parameters without mixed precision,
276276
focusing on fixed-precision quantization with comprehensive error
277277
minimization and optimization techniques.

tests_pytest/pytorch_tests/e2e_tests/wrapper/test_mct_wrapper_pytorch_e2e.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -206,7 +206,7 @@ def PTQ_Pytorch_mixed_precision(float_model):
206206
['distance_weighting_method', None], # Distance weighting method for mixed precision (low priority).
207207
['num_of_images', 5], # Number of images for mixed precision.
208208
['use_hessian_based_scores', False], # Use Hessian-based sensitivity scores for layer importance (low priority).
209-
['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size.
209+
['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size).
210210
['save_model_path', './qmodel_PTQ_Pytorch_mixed_precision.onnx'] # Path to save quantized model as ONNX.
211211
]
212212

@@ -297,7 +297,7 @@ def GPTQ_Pytorch_mixed_precision(float_model):
297297
['optimizer', None], # Optimizer to use during fine-tuning (low priority).
298298
['distance_weighting_method', None], # Distance weighting method for GPTQ (low priority).
299299
['num_of_images', 5], # Number of images to use for calibration.
300-
['use_hessian_based_scores', False], # Whether to use Hessian-based scores for layer importance.
300+
['use_hessian_based_scores', False], # Whether to use Hessian-based scores for layer importance (low priority).
301301
['save_model_path', './qmodel_GPTQ_Pytorch_mixed_precision.onnx'] # Path to save quantized model as ONNX.
302302
]
303303

tests_pytest/pytorch_tests/integration_tests/wrapper/test_mct_wrapper_pytorch_integ.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -67,7 +67,7 @@ def test_quantize_and_export_PTQ_flow(
6767
Mocked Components:
6868
- _get_tpc: TPC configuration
6969
- _select_method: Framework-specific method selection
70-
- _Setting_PTQ: PTQ parameter configuration
70+
- _setting_PTQ: PTQ parameter configuration
7171
- _export_model: Model export functionality
7272
- _post_training_quantization: Actual quantization process
7373
@@ -93,7 +93,7 @@ def test_quantize_and_export_PTQ_flow(
9393

9494
mock_setting_ptq.return_value = {'mock': 'params'}
9595

96-
param_items = [('n_epochs', 10)] # Number of epochs
96+
param_items = [('sdsp_version', '3.14')] # SDSP version for TPC
9797

9898
# Call the method
9999
success, result_model = wrapper.quantize_and_export(

tests_pytest/pytorch_tests/unit_tests/wrapper/test_mct_wrapper_pytorch_unit.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -269,9 +269,9 @@ def test_setting_PTQ_mixed_precision(
269269
def test_setting_PTQ(self, mock_core_config: Mock,
270270
mock_quant_config: Mock) -> None:
271271
"""
272-
Test _Setting_PTQ method for standard Post-Training Quantization.
272+
Test _setting_PTQ method for standard Post-Training Quantization.
273273
274-
This test verifies that the _Setting_PTQ method correctly configures
274+
This test verifies that the _setting_PTQ method correctly configures
275275
standard Post-Training Quantization parameters without mixed precision,
276276
focusing on fixed-precision quantization with comprehensive error
277277
minimization and optimization techniques.

tutorials/notebooks/mct_features_notebooks/keras/example_keras_mct_wrapper.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -391,7 +391,7 @@
391391
" ['distance_weighting_method', None], # Distance weighting method for mixed precision (low priority).\n",
392392
" ['num_of_images', 5], # Number of images for mixed precision.\n",
393393
" ['use_hessian_based_scores', False], # Use Hessian-based sensitivity scores for layer importance (low priority).\n",
394-
" ['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size.\n",
394+
" ['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size).\n",
395395
" ['save_model_path', './qmodel_PTQ_Keras_mixed_precision.keras'] # Path to save the quantized model.\n",
396396
" ]\n",
397397
"\n",

tutorials/notebooks/mct_features_notebooks/pytorch/example_pytorch_mct_wrapper.ipynb

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -381,7 +381,7 @@
381381
" ['distance_weighting_method', None], # Distance weighting method for mixed precision (low priority).\n",
382382
" ['num_of_images', 5], # Number of images for mixed precision.\n",
383383
" ['use_hessian_based_scores', False], # Use Hessian-based sensitivity scores for layer importance (low priority).\n",
384-
" ['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size.\n",
384+
" ['weights_compression_ratio', 0.75], # Target compression ratio for model weights (75% of original size).\n",
385385
" ['save_model_path', './qmodel_PTQ_Pytorch_mixed_precision.onnx'] # Path to save quantized model as ONNX.\n",
386386
" ]\n",
387387
"\n",

0 commit comments

Comments
 (0)