We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent ecb7d93 commit 27e6f07Copy full SHA for 27e6f07
examples/quantization_w8a8_fp8/qwen3_vl_moe_fp8_block_example.py
@@ -5,8 +5,12 @@
5
from llmcompressor.modeling import replace_modules_for_calibration
6
from llmcompressor.modifiers.quantization import QuantizationModifier
7
8
+# NOTE: Qwen3-VL-MoE support is not in transformers<=4.56.2
9
+# you may need to install transformes from source
10
+
11
MODEL_ID = "Qwen/Qwen3-VL-235B-A22B-Instruct"
12
13
14
# Load model.
15
model = Qwen3VLMoeForConditionalGeneration.from_pretrained(MODEL_ID, torch_dtype="auto")
16
model = replace_modules_for_calibration(model)
0 commit comments