ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2412.21187
46
90

Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs

30 December 2024
Xingyu Chen
Jiahao Xu
Tian Liang
Zhiwei He
Jianhui Pang
Dian Yu
Linfeng Song
Qiuzhi Liu
M. Zhou
Z. Zhang
Rui Wang
Zhaopeng Tu
Haitao Mi
Dong Yu
    LRM
    ReLM
ArXivPDFHTML
Abstract

The remarkable performance of models like the OpenAI o1 can be attributed to their ability to emulate human-like long-time thinking during inference. These models employ extended chain-of-thought (CoT) processes, exploring multiple strategies to enhance problem-solving capabilities. However, a critical question remains: How to intelligently and efficiently scale computational resources during testing. This paper presents the first comprehensive study on the prevalent issue of overthinking in these models, where excessive computational resources are allocated for simple problems with minimal benefit. We introduce novel efficiency metrics from both outcome and process perspectives to evaluate the rational use of computational resources by o1-like models. Using a self-training paradigm, we propose strategies to mitigate overthinking, streamlining reasoning processes without compromising accuracy. Experimental results show that our approach successfully reduces computational overhead while preserving model performance across a range of testsets with varying difficulty levels, such as GSM8K, MATH500, GPQA, and AIME.

View on arXiv
@article{chen2025_2412.21187,
  title={ Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs },
  author={ Xingyu Chen and Jiahao Xu and Tian Liang and Zhiwei He and Jianhui Pang and Dian Yu and Linfeng Song and Qiuzhi Liu and Mengfei Zhou and Zhuosheng Zhang and Rui Wang and Zhaopeng Tu and Haitao Mi and Dong Yu },
  journal={arXiv preprint arXiv:2412.21187},
  year={ 2025 }
}
Comments on this paper