Skip to content

Commit 01b52f9

Browse files
authored
update news and citation (#3889)
* update news and citation * update
1 parent d46103a commit 01b52f9

File tree

2 files changed

+29
-0
lines changed

2 files changed

+29
-0
lines changed

README.md

Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -25,6 +25,12 @@ ______________________________________________________________________
2525

2626
<details open>
2727
<summary><b>2025</b></summary>
28+
29+
- \[2025/06\] Comprehensive inference optimization for FP8 MoE Models
30+
- \[2025/06\] DeepSeek PD Disaggregation deployment is now supported through integration with [DLSlime](https://github.com/DeepLink-org/DLSlime) and [Mooncake](https://github.com/kvcache-ai/Mooncake). Huge thanks to both teams!
31+
- \[2025/04\] Enhance DeepSeek inference performance by integration deepseek-ai techniques: FlashMLA, DeepGemm, DeepEP, MicroBatch and eplb
32+
- \[2025/01\] Support DeepSeek V3 and R1
33+
2834
</details>
2935

3036
<details close>
@@ -270,6 +276,15 @@ We appreciate all contributions to LMDeploy. Please refer to [CONTRIBUTING.md](.
270276
}
271277
```
272278

279+
```bibtex
280+
@article{zhang2025efficient,
281+
title={Efficient Mixed-Precision Large Language Model Inference with TurboMind},
282+
author={Zhang, Li and Jiang, Youhe and He, Guoliang and Chen, Xin and Lv, Han and Yao, Qian and Fu, Fangcheng and Chen, Kai},
283+
journal={arXiv preprint arXiv:2508.15601},
284+
year={2025}
285+
}
286+
```
287+
273288
# License
274289

275290
This project is released under the [Apache 2.0 license](LICENSE).

README_zh-CN.md

Lines changed: 14 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27,6 +27,11 @@ ______________________________________________________________________
2727
<summary><b>2025</b></summary>
2828
</details>
2929

30+
- 【2025年6月】深度优化 FP8 MoE 模型推理
31+
- 【2025年6月】集成[DLSlime](https://github.com/DeepLink-org/DLSlime)[Mooncake](https://github.com/kvcache-ai/Mooncake),实现DeepSeek PD分离部署,向两个团队表示诚挚的感谢!
32+
- 【2025年4月】集成deepseek-ai组件FlashMLA、DeepGemm、DeepEP、MicroBatch、eplb等,提升DeepSeek推理性能
33+
- 【2025年1月】新增对DeepSeek V3及R1的支持
34+
3035
<details close>
3136
<summary><b>2024</b></summary>
3237

@@ -270,6 +275,15 @@ with lmdeploy.pipeline("internlm/internlm3-8b-instruct") as pipe:
270275
}
271276
```
272277

278+
```bibtex
279+
@article{zhang2025efficient,
280+
title={Efficient Mixed-Precision Large Language Model Inference with TurboMind},
281+
author={Zhang, Li and Jiang, Youhe and He, Guoliang and Chen, Xin and Lv, Han and Yao, Qian and Fu, Fangcheng and Chen, Kai},
282+
journal={arXiv preprint arXiv:2508.15601},
283+
year={2025}
284+
}
285+
```
286+
273287
# 开源许可证
274288

275289
该项目采用 [Apache 2.0 开源许可证](LICENSE)

0 commit comments

Comments
 (0)