Skip to content

Commit 5aa3586

Browse files
committed
use get_compressed_attention_impl
Signed-off-by: Kyle Sayers <[email protected]>
1 parent 0e4e002 commit 5aa3586

File tree

1 file changed

+1
-1
lines changed
  • src/llmcompressor/modifiers/quantization/quantization

1 file changed

+1
-1
lines changed

src/llmcompressor/modifiers/quantization/quantization/mixin.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -243,7 +243,7 @@ def _initialize_hooks(self, model: torch.nn.Module) -> Set[RemovableHandle]:
243243
hooks = set()
244244

245245
# TODO: attnq
246-
# attention_impl = enable_compressed_attention(model)
246+
# attention_impl = get_compressed_attention_impl()
247247
# hooks |= register_calibrate_attn_hooks(self, attention_impl)
248248

249249
for module in model.modules():

0 commit comments

Comments
 (0)