Skip to content

Commit 0fc29b7

Browse files
Update index.html
1 parent c5f7380 commit 0fc29b7

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

index.html

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -483,9 +483,9 @@ <h4 style="margin-bottom: 16px;">
483483
</div>
484484
<br>
485485
<p>
486-
<strong>Figure 5: Figures 5a and 5d</strong> show that sparse attention significantly improves performance and scalability, enabling higher accuracy at lower compute budgets.
486+
<strong>Figure 5: Figures 5a and 5d</strong> show that sparse attention significantly improves performance and scalability, enabling higher accuracy at lower cost budgets.
487487
<strong>Figures 5b, 5c, 5e, and 5f</strong> highlight that both 8B and 32B models benefit from sparsity, with <em>50–60 percentage point</em> gains in low-cost settings and consistent <em>~5 point</em> improvements even in high-cost regimes.
488-
Notably, sparse models reach these performance levels at much lower costs —for context, 10<sup>5</sup> Tera-eFLOPs is just 22 seconds of B200 usage.
488+
Notably, sparse models reach these performance levels at much lower costs. For reference, 10<sup>5</sup> Tera-eFLOPs is 22 seconds of B200 usage at 100% utilization.
489489
</p>
490490
</div>
491491
<br>

0 commit comments

Comments
 (0)