You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+2-2Lines changed: 2 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -292,8 +292,8 @@ python3 download_pdfs.py # The code is generated by Doubao AI
292
292
|2025.02| 🔥🔥🔥[**SeerAttention**] SeerAttention: Learning Intrinsic Sparse Attention in Your LLMs(@microsoft) |[[pdf]](https://arxiv.org/abs/2410.13276)|[[SeerAttention]](https://github.com/microsoft/SeerAttention)| ⭐️⭐️⭐️ |
293
293
|2025.03|[**Slim attention**] Slim attention: cut your context memory in half without loss of accuracy, K-cache is all you need for MHA(@OpenMachine.ai) |[[pdf]](https://arxiv.org/pdf/2503.05840)|[[OpenMchine]](https://github.com/OpenMachine-ai/transformer-tricks)| ⭐️⭐️⭐️ |
294
294
|2025.05|🔥🔥[**SageAttention-3**] SageAttention3: Microscaling FP4 Attention for Inference and An Exploration of 8-bit Training(@thu-ml)|[[pdf]](https://arxiv.org/pdf/2505.11594)|[[SageAttention]](https://github.com/thu-ml/SageAttention)| ⭐️⭐️ |
295
-
296
-
295
+
|2025.04|🔥🔥[**Parallel Encoding**] APE: Faster and Longer Context-Augmented Generation via Adaptive Parallel Encoding(@cmu.edu&NVIDIA)|[[pdf]](https://arxiv.org/pdf/2502.05431)|[[APE]](https://github.com/Infini-AI-Lab/APE)| ⭐️⭐️ |
296
+
|2025.04|🔥🔥[**Parallel Encoding**] Block-Attention for Efficient Prefilling(@Tencent etc)|[[pdf]](https://arxiv.org/pdf/2409.15355)|[[Block-attention]](https://github.com/TemporaryLoRA/Block-attention)| ⭐️⭐️ |
0 commit comments