File tree Expand file tree Collapse file tree 1 file changed +2
-5
lines changed
src/compressed_tensors/quantization/lifecycle Expand file tree Collapse file tree 1 file changed +2
-5
lines changed Original file line number Diff line number Diff line change 35
35
from compressed_tensors .quantization .quant_scheme import QuantizationScheme
36
36
from compressed_tensors .quantization .utils import (
37
37
KV_CACHE_TARGETS ,
38
- infer_quantization_status ,
39
38
is_kv_cache_quant_scheme ,
40
39
)
41
40
from compressed_tensors .utils .helpers import deprecated , replace_module
@@ -215,9 +214,7 @@ def apply_quantization_status(model: Module, status: QuantizationStatus):
215
214
:param status: status to update the module to
216
215
"""
217
216
218
- current_status = infer_quantization_status (model )
219
-
220
- if status >= QuantizationStatus .INITIALIZED > current_status :
217
+ if status >= QuantizationStatus .INITIALIZED :
221
218
force_zero_point_init = status != QuantizationStatus .COMPRESSED
222
219
223
220
# When decompressing, we set the scale_dtype as the model's dtype
@@ -235,7 +232,7 @@ def apply_quantization_status(model: Module, status: QuantizationStatus):
235
232
)
236
233
)
237
234
238
- if status >= QuantizationStatus .COMPRESSED > current_status :
235
+ if status >= QuantizationStatus .COMPRESSED :
239
236
model .apply (compress_quantized_weights )
240
237
241
238
You can’t perform that action at this time.
0 commit comments