Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

XiaoYee/Awesome_Efficient_LRM_Reasoning

Open more actions menu

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 

Repository files navigation

A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, Agent, and Beyond

Paper Github Twitter

Awesome License: MIT


🔔 News


Author

Taxonomy

Citation

If you find our survey useful for your research, please consider citing:

@article{qu2025survey,
  title={A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond},
  author={Qu, Xiaoye and Li, Yafu and Su, Zhaochen and Sun, Weigao and Yan, Jianhao and Liu, Dongrui and Cui, Ganqu and Liu, Daizong and Liang, Shuxian and He, Junxian and others},
  journal={arXiv preprint arXiv:2503.21614},
  year={2025}
}


Category

🔥 Table of Contents


📜Content

👀 Introduction

In the age of LRMs, we propose that "Efficiency is the essence of intelligence." Just as a wise human knows when to stop thinking and start deciding, a wise model should know when to halt unnecessary deliberation. An intelligent model should manipulate the token economy, i.e., allocating tokens purposefully, skipping redundancy, and optimizing the path to a solution. Rather than naively traversing every possible reasoning path, it should emulate a master strategist, balancing cost and performance with elegant precision.

To summarize, this survey makes the following key contributions to the literature:

  • Instead of offering a general overview of LRMs, we focus on the emerging and critical topic of efficient reasoning in LRMs, providing an in-depth and targeted analysis.
  • We identify and characterize common patterns of reasoning inefficiency, and outline the current challenges that are unique to improving reasoning efficiency in large models.
  • We provide a comprehensive review of recent advancements aimed at enhancing reasoning efficiency, structured across the end-to-end LRM development pipeline, from pretraining and supervised fine-tuning to reinforcement learning and inference.

🚀 Papers

💭 Efficient Reasoning during Inference

Length Budgeting

System Switch

Model Switch

Model Merge

Parallel Search

💫 Efficient Reasoning with SFT

Reasoning Chain Compression

Latent-Space SFT

🧩 Efficient Reasoning with Reinforcement Learning

Efficient Reinforcement Learning with Length Reward

Efficient Reinforcement Learning without Length Reward

💬 Efficient Reasoning during Pre-training

Pretraining with Latent Space

Subquadratic Attention

Linearization

Efficient Reasoning with Subquadratic Attention

🔖 Future Directions

Efficient Multimodal Reasoning and Video Reasoning

Efficient Test-time Scaling and Infinity Thinking

Efficient and Trustworthy Reasoning

Building Efficient Reasoning Applications (RAG, Tool, Agent)

Evaluation and Benchmark


🎉 Contribution

Contributing to this paper list

⭐" Join us in improving this repository! If you know of any important works we've missed, please contribute. Your efforts are highly valued! "

Contributors


Releases

No releases published

Packages

No packages published

Contributors 6

Morty Proxy This is a proxified and sanitized view of the page, visit original site.