You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: examples/GPTQ/README.md
+17-23Lines changed: 17 additions & 23 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -62,10 +62,10 @@ This end-to-end example utilizes the common set of interfaces provided by `fms_m
62
62
63
63
```
64
64
layer mem (MB)
65
-
dtype
66
-
torch.float16224 109.051904
67
-
torch.float3267 4203.757568
68
-
torch.int32 672 3521.904640
65
+
dtype
66
+
torch.bfloat16 67 2101.878784
67
+
torch.float16224 109.051904
68
+
torch.int32 672 3521.904640
69
69
```
70
70
71
71
4. **Evaluate the quantized model**'s performance on a selected task using `lm-eval` library, the command below will run evaluation on [`lambada_openai`](https://huggingface.co/datasets/EleutherAI/lambada_openai) task and show the perplexity/accuracy at the end.
@@ -82,29 +82,23 @@ This end-to-end example utilizes the common set of interfaces provided by `fms_m
0 commit comments