|
41 | 41 | from nncf.quantization.advanced_parameters import AdvancedCompressionParameters
|
42 | 42 | from nncf.quantization.advanced_parameters import AdvancedQuantizationParameters
|
43 | 43 | from nncf.quantization.advanced_parameters import convert_to_dict_recursively
|
| 44 | +from nncf.quantization.advanced_parameters import is_weight_compression_needed |
44 | 45 | from nncf.quantization.algorithms.accuracy_control.algorithm import QuantizationAccuracyRestorer
|
45 | 46 | from nncf.quantization.algorithms.accuracy_control.algorithm import calculate_accuracy_drop
|
46 | 47 | from nncf.quantization.algorithms.accuracy_control.evaluator import Evaluator
|
@@ -121,7 +122,7 @@ def _extract_all_subgraphs(model: ov.Model, current_id: str) -> None:
|
121 | 122 | quantization_algorithm, model, graphs, main_model_graph_id, calibration_dataset, subset_size, 1
|
122 | 123 | )
|
123 | 124 |
|
124 |
| - if advanced_parameters is None or advanced_parameters.is_weight_compression_needed(): |
| 125 | + if is_weight_compression_needed(advanced_parameters): |
125 | 126 | compress_quantize_weights_transformation(quantized_model)
|
126 | 127 |
|
127 | 128 | dump_parameters(
|
@@ -168,7 +169,7 @@ def native_quantize_impl(
|
168 | 169 | warning_model_no_batchwise_support(graph, advanced_parameters, model_type, OPERATIONS_OUTPUT_HAS_NO_BATCH_AXIS)
|
169 | 170 | quantized_model = quantization_algorithm.apply(model, graph, dataset=calibration_dataset)
|
170 | 171 |
|
171 |
| - if advanced_parameters is None or advanced_parameters.is_weight_compression_needed(): |
| 172 | + if is_weight_compression_needed(advanced_parameters): |
172 | 173 | compress_quantize_weights_transformation(quantized_model)
|
173 | 174 |
|
174 | 175 | dump_parameters(
|
@@ -214,7 +215,7 @@ def quantize_with_accuracy_control_impl(
|
214 | 215 | else:
|
215 | 216 | copied_parameters = deepcopy(advanced_quantization_parameters)
|
216 | 217 |
|
217 |
| - compress_weights = copied_parameters.is_weight_compression_needed() |
| 218 | + compress_weights = is_weight_compression_needed(copied_parameters) |
218 | 219 | copied_parameters.backend_params[AdvancedQuantizationParameters.COMPRESS_WEIGHTS] = False
|
219 | 220 |
|
220 | 221 | quantized_model = quantize_impl(
|
|
0 commit comments