diff --git a/src/llmcompressor/core/lifecycle.py b/src/llmcompressor/core/lifecycle.py index ff91c70c8e..bd485e01e1 100644 --- a/src/llmcompressor/core/lifecycle.py +++ b/src/llmcompressor/core/lifecycle.py @@ -91,11 +91,8 @@ def initialize( :return: List of data returned from initialization of modifiers :rtype: List[Any] """ - self.state.update(**kwargs) - if self.initialized_: # TODO: do not initialize twice - return - logger.debug("Initializing compression lifecycle") + self.state.update(**kwargs) self.recipe_container.append(recipe, recipe_stage, recipe_args) self.modifiers = self.recipe_container.get_modifiers() self._set_model_layer_prefix() diff --git a/src/llmcompressor/core/state.py b/src/llmcompressor/core/state.py index 42e43e1345..c089987d34 100644 --- a/src/llmcompressor/core/state.py +++ b/src/llmcompressor/core/state.py @@ -112,18 +112,6 @@ class State: model_log_cadence: Optional[float] = None _last_log_step: Union[float, int, None] = None - @property - def compression_ready(self) -> bool: - """ - Check if the model and optimizer are set for compression. - - :return: True if model and optimizer are set, False otherwise - :rtype: bool - """ - ready = self.model is not None and self.optimizer is not None - logger.debug("Compression ready: {}", ready) - return ready - def update( self, model: Any = None, diff --git a/tests/unit/core/test_state.py b/tests/unit/core/test_state.py index 203e2a6647..0cdf7fcb95 100644 --- a/tests/unit/core/test_state.py +++ b/tests/unit/core/test_state.py @@ -63,16 +63,6 @@ def test_state_update(): assert state.model_log_cadence == 2 -@pytest.mark.regression -def test_state_sparsification_ready(): - state = State() - assert not state.compression_ready - - state.model = "model" - state.optimizer = "optimizer" - assert state.compression_ready - - @pytest.mark.regression def test_state_update_loggers(): state = State()