ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.04472
64
12

DAST: Difficulty-Adaptive Slow-Thinking for Large Reasoning Models

6 March 2025
Yi Shen
J. Zhang
Jieyun Huang
Shuming Shi
Wenjing Zhang
Jiangze Yan
Ning Wang
Kai Wang
Shiguo Lian
    LRM
ArXivPDFHTML
Abstract

Recent advancements in slow-thinking reasoning models have shown exceptional performance in complex reasoning tasks. However, these models often exhibit overthinking-generating redundant reasoning steps for simple problems, leading to excessive computational resource usage. While current mitigation strategies uniformly reduce reasoning tokens, they risk degrading performance on challenging tasks that require extended reasoning. This paper introduces Difficulty-Adaptive Slow-Thinking (DAST), a novel framework that enables models to autonomously adjust the length of Chain-of-Thought(CoT) based on problem difficulty. We first propose a Token Length Budget (TLB) metric to quantify difficulty, then leveraging length-aware reward shaping and length preference optimization to implement DAST. DAST penalizes overlong responses for simple tasks while incentivizing sufficient reasoning for complex problems. Experiments on diverse datasets and model scales demonstrate that DAST effectively mitigates overthinking (reducing token usage by over 30\% on average) while preserving reasoning accuracy on complex problems.

View on arXiv
@article{shen2025_2503.04472,
  title={ DAST: Difficulty-Adaptive Slow-Thinking for Large Reasoning Models },
  author={ Yi Shen and Jian Zhang and Jieyun Huang and Shuming Shi and Wenjing Zhang and Jiangze Yan and Ning Wang and Kai Wang and Shiguo Lian },
  journal={arXiv preprint arXiv:2503.04472},
  year={ 2025 }
}
Comments on this paper