Skip to content

Commit 42575fd

Browse files
authored
default W4A16 alias to use group_size=128 (#94) (#95)
1 parent b77b1e6 commit 42575fd

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

src/compressed_tensors/quantization/quant_scheme.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ def is_preset_scheme(name: str) -> bool:
111111
weights=QuantizationArgs(), input_activations=QuantizationArgs(symmetric=True)
112112
)
113113

114-
W4A16 = dict(weights=QuantizationArgs(num_bits=4, symmetric=True))
114+
W4A16 = dict(weights=QuantizationArgs(num_bits=4, group_size=128))
115115

116116
PRESET_SCHEMES = {
117117
"W8A8": W8A8,

0 commit comments

Comments
 (0)