We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent dc2e9b0 commit 4bde0feCopy full SHA for 4bde0fe
examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py
@@ -90,8 +90,8 @@
90
tokenizer.save_pretrained(f"{output_dir}/quantization_stage")
91
92
logger.info(
93
- "llmcompressor does not currently support running ",
+ "llmcompressor does not currently support running "
94
"compressed models in the marlin24 format. "
95
- "The model produced from this example can be ",
96
- "run on vLLM with dtype=torch.float16.",
+ "The model produced from this example can be "
+ "run on vLLM with dtype=torch.float16."
97
)
0 commit comments