Skip to content

Commit 4f98fe0

Browse files
committed
Update perf number
1 parent ca10839 commit 4f98fe0

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

mixtral-moe/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,8 +14,8 @@ Benchmarks run on an 8xA100-80GB, power limited to 330W with a hybrid cube mesh
1414

1515
| | 1 GPU | 2 GPU | 4 GPU | 8 GPU |
1616
|------------------|---------|-----------|--------|------------|
17-
|baseline(bfloat16)| OOM | 96.67 | 125.70 | 227.82 |
18-
| int8 | 97.92 | 155.03 | 162.58 | 279.35 |
17+
|baseline(bfloat16)| OOM | 96.67 | 155.35 | 227.82 |
18+
| int8 | 97.92 | 155.03 | 216.87 | 279.35 |
1919

2020

2121
## Generate Text

mixtral-moe/generate.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ def device_sync(device):
2424

2525
torch._inductor.config.coordinate_descent_tuning = True
2626
torch._inductor.config.triton.unique_kernel_names = True
27-
# torch._inductor.config.fx_graph_cache = True # Experimental feature to reduce compilation times, will be on by default in future
27+
torch._inductor.config.fx_graph_cache = True # Experimental feature to reduce compilation times, will be on by default in future
2828

2929

3030
# support running without installing as a package

0 commit comments

Comments
 (0)