Breaking the reasoning barrier: A survey on LLM complex reasoning through the lens of self-evolution
Publication Type
Conference Proceeding Article
Version
publishedVersion
Publication Date
7-2025
Abstract
The release of OpenAI’s O1 and subsequent projects like DeepSeek R1 has significantly advanced research on complex reasoning in LLMs. This paper systematically analyzes existing reasoning studies from the perspective of self-evolution, structured into three components: data evolution, model evolution, and self-evolution. Data evolution explores methods to generate higher-quality reasoning training data. Model evolution focuses on training strategies to boost reasoning capabilities. Self-evolution research autonomous system evolution via iterating cycles of data and model evolution. We further discuss the scaling law of self-evolution and analyze representative O1-like works through this lens. By summarizing advanced methods and outlining future directions, this paper aims to drive advancements in LLMs’ reasoning abilities.
Discipline
Artificial Intelligence and Robotics
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (ACL 2025), Vienna, Austria, July 27 - August 1
First Page
7377
Last Page
7417
Identifier
10.18653/v1/2025.findings-acl.386
Publisher
ACL
City or Country
Austria
Citation
HE, Tao; LI, Hao; CHEN, Jingchang; LIU, Runxuan; CAO, Yixin; LIAO, Lizi; ZHENG, Zihao; CHU, Zheng; LIANG, Jiafeng; LIU, Ming; and QIN, Bing.
Breaking the reasoning barrier: A survey on LLM complex reasoning through the lens of self-evolution. (2025). Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (ACL 2025), Vienna, Austria, July 27 - August 1. 7377-7417.
Available at: https://ink.library.smu.edu.sg/sis_research/10759
Creative Commons License

This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.18653/v1/2025.findings-acl.386