Skip to content

Commit 82f8128

Browse files
authored
Update Multi-GPUs/Multi-Nodes Parallelism (#141)
1 parent 1101356 commit 82f8128

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ python3 download_pdfs.py # The code is generated by Doubao AI
4545
## 📖Contents
4646
* 📖[Trending LLM/VLM Topics](#Trending-LLM-VLM-Topics)🔥🔥🔥
4747
* 📖[DeepSeek/MLA Topics](#mla)🔥🔥🔥
48-
* 📖[DP/MP/PP/TP/SP/CP Parallelism](#DP-MP-PP-TP-SP-CP)🔥🔥🔥
48+
* 📖[Multi-GPUs/Multi-Nodes Parallelism](#DP-MP-PP-TP-SP-CP)🔥🔥🔥
4949
* 📖[Disaggregating Prefill and Decoding](#P-D-Disaggregating)🔥🔥🔥
5050
* 📖[LLM Algorithmic/Eval Survey](#LLM-Algorithmic-Eval-Survey)
5151
* 📖[LLM Train/Inference Framework/Design](#LLM-Train-Inference-Framework)
@@ -103,7 +103,7 @@ python3 download_pdfs.py # The code is generated by Doubao AI
103103
|2025.02|🔥🔥[**TransMLA**] TransMLA: Multi-head Latent Attention Is All You Need(@PKU)|[[pdf]](https://arxiv.org/pdf/2502.07864)|[[TransMLA]](https://github.com/fxmeng/TransMLA) ![](https://img.shields.io/github/stars/fxmeng/TransMLA.svg?style=social) | ⭐️⭐️ |
104104
|2025.03|🔥🔥[**X-EcoMLA**] X-EcoMLA: Upcycling Pre-Trained Attention into MLA for Efficient and Extreme KV Compression(@AMD)| [[pdf]](https://arxiv.org/pdf/2503.11132) |⚠️|⭐️⭐️ |
105105

106-
### 📖DP/MP/PP/TP/SP/CP Parallelism ([©️back👆🏻](#paperlist))
106+
### 📖Multi-GPUs/Multi-Nodes Parallelism ([©️back👆🏻](#paperlist))
107107
<div id="DP-MP-PP-TP-SP-CP"></div>
108108

109109
|Date|Title|Paper|Code|Recom|

0 commit comments

Comments
 (0)