Skip to content

😎 A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyond

License

Notifications You must be signed in to change notification settings

XiaoYee/Awesome_Efficient_LRM_Reasoning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

106 Commits
 
 
 
 
 
 

Repository files navigation

A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyond

Paper Github Twitter

Awesome License: MIT


🔔 News


Author

Scope

Taxonomy

Citation

If you find our survey useful for your research, please consider citing:

@article{qu2025survey,
  title={A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond},
  author={Qu, Xiaoye and Li, Yafu and Su, Zhaochen and Sun, Weigao and Yan, Jianhao and Liu, Dongrui and Cui, Ganqu and Liu, Daizong and Liang, Shuxian and He, Junxian and others},
  journal={arXiv preprint arXiv:2503.21614},
  year={2025}
}


Category

🔥 Table of Contents


📜Content

👀 Introduction

In the age of LRMs, we propose that "Efficiency is the essence of intelligence." Just as a wise human knows when to stop thinking and start deciding, a wise model should know when to halt unnecessary deliberation. An intelligent model should manipulate the token economy, i.e., allocating tokens purposefully, skipping redundancy, and optimizing the path to a solution. Rather than naively traversing every possible reasoning path, it should emulate a master strategist, balancing cost and performance with elegant precision.

To summarize, this survey makes the following key contributions to the literature:

  • Instead of offering a general overview of LRMs, we focus on the emerging and critical topic of efficient reasoning in LRMs, providing an in-depth and targeted analysis.
  • We identify and characterize common patterns of reasoning inefficiency, and outline the current challenges that are unique to improving reasoning efficiency in large models.
  • We provide a comprehensive review of recent advancements aimed at enhancing reasoning efficiency, structured across the end-to-end LRM development pipeline, from pretraining and supervised fine-tuning to reinforcement learning and inference.

🚀 Papers

💭 Efficient Reasoning during Inference

Length Budgeting

System Switch

Model Switch

Model Merge

Parallel Search

💫 Efficient Reasoning with SFT

Reasoning Chain Compression

Latent-Space SFT

🧩 Efficient Reasoning with Reinforcement Learning

Efficient Reinforcement Learning with Length Reward

Efficient Reinforcement Learning without Length Reward

💬 Efficient Reasoning during Pre-training

Pretraining with Latent Space

Subquadratic Attention

Linearization

Efficient Reasoning with Subquadratic Attention

🔖 Future Directions

Efficient Multimodal Reasoning and Video Reasoning

Efficient Test-time Scaling and Infinity Thinking

Efficient and Trustworthy Reasoning

Building Efficient Reasoning Applications (RAG, Tool, Agent)

Evaluation and Benchmark


🎉 Contribution

Contributing to this paper list

⭐" Join us in improving this repository! If you know of any important works we've missed, please contribute. Your efforts are highly valued! "

Contributors


About

😎 A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyond

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 6