Jieyun Huang
2025
DAST: Difficulty-Adaptive Slow-Thinking for Large Reasoning Models
Yi Shen
|
Jian Zhang
|
Jieyun Huang
|
Shuming Shi
|
Wenjing Zhang
|
Jiangze Yan
|
Ning Wang
|
Kai Wang
|
Zhaoxiang Liu
|
Shiguo Lian
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Recent advancements in slow-thinking reasoning models have shown exceptional performance in complex reasoning tasks. However, their tendency for “overthinking” on simple problems leads to excessive computational resource usage and increased inference latency, which hinders their widespread industrial adoption. While current mitigation strategies uniformly reduce reasoning tokens, they risk degrading performance on challenging tasks that require extended reasoning. This paper introduces Difficulty-Adaptive Slow-Thinking (DAST), a novel framework that enables models to autonomously adjust Chain-of-Thought (CoT) length based on problem difficulty. We propose a Token Length Budget (TLB) metric and leverage budget-aware preference optimization to implement DAST, which penalizes inefficiency on simple problems while incentivizing deep reasoning for complex ones. Experiments demonstrate DAST’s significant value for practical application: it effectively mitigates overthinking, substantially lowering costs and latency—while crucially preserving high accuracy on complex problems, paving the way for the efficient application of advanced reasoning models.
Search
Fix author
Co-authors
- Shiguo Lian 1
- Zhaoxiang Liu 1
- Yi Shen 1
- Shuming Shi 1
- Ning Wang 1
- show all...