We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 18eff29 commit 96ace59Copy full SHA for 96ace59
examples/quantization_w8a8_fp8/fp8_block_example.py
@@ -7,9 +7,7 @@
7
MODEL_ID = "Qwen/Qwen3-30B-A3B"
8
9
# Load model.
10
-model = AutoModelForCausalLM.from_pretrained(
11
- MODEL_ID, device_map="auto", torch_dtype="auto"
12
-)
+model = AutoModelForCausalLM.from_pretrained(MODEL_ID, torch_dtype="auto")
13
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
14
15
# Configure the quantization algorithm and scheme.
0 commit comments