We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent a64a777 commit 8f71004Copy full SHA for 8f71004
examples/quantization_2of4_sparse_w4a16/README.md
@@ -49,9 +49,7 @@ import torch
49
from transformers import AutoModelForCausalLM
50
51
model_stub = "neuralmagic/Llama-2-7b-ultrachat200k"
52
-model = AutoModelForCausalLM.from_pretrained(
53
- model_stub, torch_dtype=torch.bfloat16, device_map="auto"
54
-)
+model = AutoModelForCausalLM.from_pretrained(model_stub, torch_dtype=torch.bfloat16)
55
56
dataset = "ultrachat-200k"
57
splits = {"calibration": "train_gen[:5%]", "train": "train_gen"}
0 commit comments