Skip to content

Commit 4b74e6e

Browse files
Update model name to LLaDA2.0-flash-CAP (#286)
Updated the model name from LLaDA2.0-flash to LLaDA2.0-flash-CAP in the streaming output section and adjusted the corresponding text.
1 parent 7bff73e commit 4b74e6e

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

blog/2025-12-17-diffusion-llm.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -94,12 +94,12 @@ Simply put, if we visualize the attention mask as a geometric shape for the `Q_c
9494

9595
## Streaming output animation
9696

97-
Here is an animation comparing the streaming output of LLaDA2.0-flash (100B / BF16) and gpt-oss-120B (117B / MXFP4). LLaDA2.0-flash is served using SGLang dLLM with TP8 on 8 × H20, while gpt-oss-120B is served using SGLang's standard AR process on the same hardware.
97+
Here is an animation comparing the streaming output of LLaDA2.0-flash-CAP (100B / BF16) and gpt-oss-120B (117B / MXFP4). LLaDA2.0-flash-CAP is served using SGLang dLLM with TP8 on 8 × H20, while gpt-oss-120B is served using SGLang's standard AR process on the same hardware.
9898

99-
Both models are asked to implement the quicksort algorithm in 10 programming languages — a task particularly well-suited for diffusion LLMs. As shown, LLaDA2.0-flash achieves significantly higher throughput at 935 tokens/s, compared to gpt-oss-120B (263 tokens/s) in this scenario.
99+
Both models are asked to implement the quicksort algorithm in 10 programming languages — a task particularly well-suited for diffusion LLMs. As shown, LLaDA2.0-flash-CAP achieves significantly higher throughput at 935 tokens/s, compared to gpt-oss-120B (263 tokens/s) in this scenario.
100100

101101
<p align="center">
102-
<img src="/images/blog/dllm/llada2-vs-gpt-oss.gif" alt="LLaDA2.0-flash vs gpt-oss-120B animation">
102+
<img src="/images/blog/dllm/llada2-vs-gpt-oss.gif" alt="LLaDA2.0-flash-CAP vs gpt-oss-120B animation">
103103
<br>
104104
</p>
105105

0 commit comments

Comments
 (0)