v2.6.1
What's Changed
- [From Author] Link CacheGen and CacheBlend to LMCache by @KuntaiDu in https://github.com/DefTruth/Awesome-LLM-Inference/pull/80
- 🔥[LORC] Low-Rank Compression for LLMs KV Cache with a Progressive Compression Strategy by @DefTruth in https://github.com/DefTruth/Awesome-LLM-Inference/pull/81
- Large Language Model Performance Benchmarking on Mobile Platforms: A Thorough Evaluation by @DefTruth in https://github.com/DefTruth/Awesome-LLM-Inference/pull/82
- [LLM Inference] LARGE LANGUAGE MODEL INFERENCE ACCELERATION: A COMPREHENSIVE HARDWARE PERSPECTIVE by @DefTruth in https://github.com/DefTruth/Awesome-LLM-Inference/pull/83
- 🔥[PARALLELSPEC] PARALLELSPEC: PARALLEL DRAFTER FOR EFFICIENT SPECULATIVE DECODING by @DefTruth in https://github.com/DefTruth/Awesome-LLM-Inference/pull/84
New Contributors
- @KuntaiDu made their first contribution in https://github.com/DefTruth/Awesome-LLM-Inference/pull/80
Full Changelog: DefTruth/Awesome-LLM-Inference@v2.6...v2.6.1