Xialin He
2025
AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time
Junyu Zhang
|
Runpei Dong
|
Han Wang
|
Xuying Ning
|
Haoran Geng
|
Peihao Li
|
Xialin He
|
Yutong Bai
|
Jitendra Malik
|
Saurabh Gupta
|
Huan Zhang
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
This paper presents AlphaOne (𝛼1), a universal framework for modulating reasoning progress in large reasoning models (LRMs) at test time. 𝛼1 first introduces 𝛼 moment, which represents the scaled thinking phase with a universal parameter 𝛼.Within this scaled pre-𝛼 moment phase, it dynamically schedules slow thinking transitions by modeling the insertion of reasoning transition tokens as a Bernoulli stochastic process. After the 𝛼 moment, 𝛼1 deterministically terminates slow thinking with the end-of-thinking token, thereby fostering fast reasoning and efficient answer generation. This approach unifies and generalizes existing monotonic scaling methods by enabling flexible and dense slow-to-fast reasoning modulation. Extensive empirical studies on various challenging benchmarks across mathematical, coding, and scientific domains demonstrate 𝛼1‘s superior reasoning capability and efficiency. Project page: https://alphaone-project.github.io/.
Search
Fix author
Co-authors
- Yutong Bai 1
- Runpei Dong 1
- Haoran Geng 1
- Saurabh Gupta 1
- Peihao Li 1
- show all...