From 955d761e90b0f0e17724168e8038dc7e69bfa0b2 Mon Sep 17 00:00:00 2001 From: youkaichao Date: Mon, 29 Sep 2025 21:58:56 +0800 Subject: [PATCH] add mlu Signed-off-by: youkaichao --- _posts/2025-09-29-deepseek-v3-2.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/_posts/2025-09-29-deepseek-v3-2.md b/_posts/2025-09-29-deepseek-v3-2.md index 9b6aab8..9ef9fee 100644 --- a/_posts/2025-09-29-deepseek-v3-2.md +++ b/_posts/2025-09-29-deepseek-v3-2.md @@ -102,7 +102,7 @@ We are also excited to announce out-of-the-box Blackwell support for this model. We are barely touching the surface of the optimization for DSA and related sparse attention in vLLM. In the coming weeks: * We plan to expand the architectures supported beyond Hopper and Blackwell. -* We will expand the support to other hardwares such as AMD and TPU. With vLLM's extensible systems, developers can add support for models directly. For example, [vllm-ascend](https://github.com/vllm-project/vllm-ascend/tree/v0.9.1-dev/examples/deepseek.md) already supports DeepSeek V3.2! +* We will expand the support to other hardwares such as AMD and TPU. With vLLM's extensible systems, developers can add support for models directly. For example, [vllm-ascend](https://github.com/vllm-project/vllm-ascend/tree/v0.9.1-dev/examples/deepseek.md) and [vllm-mlu](https://github.com/Cambricon/vllm-mlu) already support DeepSeek V3.2! * We continuously test large-scale wide EP serving and disaggregation. * You will soon be able to run an end-to-end RL loop with this model. * We will explore the "masked MHA mode for short sequence prefilling" from DeepSeek.