You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: blog/2025-09-10-sglang-hicache.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -55,7 +55,7 @@ This design has already enabled us to integrate three performant backends—[Moo
55
55
56
56
## **Benchmark**
57
57
58
-
Try experience the performance gain on your own! You can find various benchmark about HiCache [here](https://github.com/sgl-project/sglang/tree/main/benchmark/hicache). Following we highlight two benchmark results using the provided benchmark scripts and you can find the config instructions of backends [here](https://github.com/sgl-project/sglang/tree/main/python/sglang/srt/mem_cache/storage). If you have any questions about benchmarking or deployment, feel free to open an issue on GitHub or post in our [slack channel](https://sgl-fru7574.slack.com/archives/C095B2L7UEB).
58
+
Try experience the performance gain on your own! You can find various benchmark about HiCache [here](https://github.com/sgl-project/sglang/tree/main/benchmark/hicache). Following we highlight two benchmark results using the provided benchmark scripts and you can find the config instructions of backends [here](https://github.com/sgl-project/sglang/tree/main/python/sglang/srt/mem_cache/storage). If you have any questions about benchmarking or deployment, feel free to open an issue on GitHub or post in our [slack channel](https://slack.sglang.ai/).
We also want to highlight [NIXL](https://github.com/ai-dynamo/nixl) as a special backend, which is a communication library designed to bridge storage backends such as GPU-direct storage and cloud object storage. You can find more details [here](https://github.com/sgl-project/sglang/tree/main/python/sglang/srt/mem_cache/storage/nixl) and stay tuned for upcoming integration with the [Dynamo](https://github.com/ai-dynamo/dynamo) ecosystems.
89
89
90
90
## Acknowledgement:
91
-
92
-
We would like to express our sincere gratitude for the tremendous support and feedback from the community. In particular, we thank Moein Khazraee, Vishwanath Venkatesan and the Dynamo team from NVIDIA for the NIXL integration;
93
-
Sicheng Pan, Zhangheng Huang, Yi Zhang, Jianxing Zhu and Yifei Kang from the AlibabaCloud TairKVCache Team for contributing the 3FS backend integration;
94
-
Tingwei Huang, Yongke Zhao from Ant Group, Teng Ma, Shangming Cai, Xingyu Liu from AlibabaCloud,
95
-
Jinyang Su, Ke Yang from Approaching.AI, and Zuoyuan Zhang and Mingxing Zhang from the Mooncake community for the Mooncake integration.
96
-
We also appreciate the ongoing efforts from the LMCache, AIBrix and ByteDance EIC teams to bring their products into the ecosystem.
91
+
We would like to express our sincere gratitude for the tremendous support and feedback from the community.
92
+
We are grateful to Sicheng Pan, Zhangheng Huang, Yi Zhang, Jianxing Zhu, and Yifei Kang from the Alibaba Cloud TairKVCache team for the 3FS backend integration;
93
+
Tingwei Huang and Yongke Zhao from Ant Group; Teng Ma, Shangming Cai, and Xingyu Liu from Alibaba Cloud; Jinyang Su and Ke Yang from Approaching.AI; and Zuoyuan Zhang and Mingxing Zhang from the Mooncake community for their efforts on Mooncake integration;
94
+
Moein Khazraee, Vishwanath Venkatesan, and the Dynamo team from NVIDIA for enabling the NIXL integration.
95
+
Special thanks go to Ziyi Xu from the SGLang team, Yuwei An from LMCache, Vikram Sharma Mailthody, Scott Mahlke, and Michael Garland from NVIDIA, as well as Mark Zhao and Christos Kozyrakis from Stanford for their contributions to the HiCache design and implementation.
96
+
Finally, we appreciate the ongoing contributions from the LMCache, AIBrix, and ByteDance EIC teams in bringing their products into the ecosystem.
0 commit comments