Skip to content

Commit a173a0c

Browse files
dsikkarahul-tuli
andauthored
Update example to not fail hessian inversion (#904)
* update Signed-off-by: Dipika <[email protected]> * quality --------- Signed-off-by: Dipika <[email protected]> Co-authored-by: Rahul Tuli <[email protected]>
1 parent 644a500 commit a173a0c

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

examples/big_models_with_accelerate/multi_gpu_int8.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,9 @@ def tokenize(sample):
5959
# * quantize the weights to int8 with GPTQ (static per channel)
6060
# * quantize the activations to int8 (dynamic per token)
6161
recipe = [
62-
GPTQModifier(targets="Linear", scheme="W8A8", ignore=["lm_head"]),
62+
GPTQModifier(
63+
targets="Linear", scheme="W8A8", ignore=["lm_head"], dampening_frac=0.1
64+
),
6365
]
6466

6567
# 4) Apply algorithms and save in `compressed-tensors` format.

0 commit comments

Comments
 (0)