Peihao Li


2025

pdf bib
AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time
Junyu Zhang | Runpei Dong | Han Wang | Xuying Ning | Haoran Geng | Peihao Li | Xialin He | Yutong Bai | Jitendra Malik | Saurabh Gupta | Huan Zhang
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing

This paper presents AlphaOne (𝛼1), a universal framework for modulating reasoning progress in large reasoning models (LRMs) at test time. 𝛼1 first introduces 𝛼 moment, which represents the scaled thinking phase with a universal parameter 𝛼.Within this scaled pre-𝛼 moment phase, it dynamically schedules slow thinking transitions by modeling the insertion of reasoning transition tokens as a Bernoulli stochastic process. After the 𝛼 moment, 𝛼1 deterministically terminates slow thinking with the end-of-thinking token, thereby fostering fast reasoning and efficient answer generation. This approach unifies and generalizes existing monotonic scaling methods by enabling flexible and dense slow-to-fast reasoning modulation. Extensive empirical studies on various challenging benchmarks across mathematical, coding, and scientific domains demonstrate 𝛼1‘s superior reasoning capability and efficiency. Project page: https://alphaone-project.github.io/.