Skip to content

Commit cc27568

Browse files
remove infer_quantization_status
Signed-off-by: Brian Dellabetta <[email protected]>
1 parent 360a9fb commit cc27568

File tree

1 file changed

+2
-5
lines changed
  • src/compressed_tensors/quantization/lifecycle

1 file changed

+2
-5
lines changed

src/compressed_tensors/quantization/lifecycle/apply.py

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,6 @@
3535
from compressed_tensors.quantization.quant_scheme import QuantizationScheme
3636
from compressed_tensors.quantization.utils import (
3737
KV_CACHE_TARGETS,
38-
infer_quantization_status,
3938
is_kv_cache_quant_scheme,
4039
)
4140
from compressed_tensors.utils.helpers import deprecated, replace_module
@@ -215,9 +214,7 @@ def apply_quantization_status(model: Module, status: QuantizationStatus):
215214
:param status: status to update the module to
216215
"""
217216

218-
current_status = infer_quantization_status(model)
219-
220-
if status >= QuantizationStatus.INITIALIZED > current_status:
217+
if status >= QuantizationStatus.INITIALIZED:
221218
force_zero_point_init = status != QuantizationStatus.COMPRESSED
222219

223220
# When decompressing, we set the scale_dtype as the model's dtype
@@ -235,7 +232,7 @@ def apply_quantization_status(model: Module, status: QuantizationStatus):
235232
)
236233
)
237234

238-
if status >= QuantizationStatus.COMPRESSED > current_status:
235+
if status >= QuantizationStatus.COMPRESSED:
239236
model.apply(compress_quantized_weights)
240237

241238

0 commit comments

Comments
 (0)