Skip to content

Commit aad4e06

Browse files
committed
update comment
Signed-off-by: Kyle Sayers <kylesayrs@gmail.com>
1 parent f0f6392 commit aad4e06

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

tests/llmcompressor/transformers/kv_cache/test_kv_cache.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -237,7 +237,9 @@ def test_kv_cache_gptq_model_state_dict_attr(kv_cache_fixture, tmp_path):
237237
output_dir, _ = next(kv_cache_fixture(recipe, tmp_path))
238238

239239
with init_empty_weights():
240-
# There is a bug in `apply_quantization_config`
240+
# TODO: There is a bug in `apply_quantization_config` which means that, if using
241+
# CompressedLinears, the compression status is inferred to `compressed` and
242+
# therefore the attention kvcache parameters never undergo initializations
241243
model = AutoModelForCausalLM.from_pretrained(
242244
output_dir,
243245
quantization_config=CompressedTensorsConfig(run_compressed=False),

0 commit comments

Comments
 (0)