Jiajun Luo
2025
Beyond A Single AI Cluster: A Survey of Decentralized LLM Training
Haotian Dong
|
Jingyan Jiang
|
Rongwei Lu
|
Jiajun Luo
|
Jiajun Song
|
Bowen Li
|
Ying Shen
|
Zhi Wang
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
The emergence of large language models (LLMs) has revolutionized AI development, yet their resource demands beyond a single cluster or even datacenter, limiting accessibility to well-resourced organizations. Decentralized training has emerged as a promising paradigm to leverage dispersed resources across clusters, datacenters and even regions, offering the potential to democratize LLM development for broader communities. As the first comprehensive exploration of this emerging field, we present decentralized LLM training as a resource-driven paradigm and categorize existing efforts into community-driven and organizational approaches. We further clarify this through: (1) a comparison with related paradigms, (2) characterization of decentralized resources, and (3) a taxonomy of recent advancements. We also provide up-to-date case studies and outline future directions to advance research in decentralized LLM training.
Search
Fix author
Co-authors
- Haotian Dong 1
- Jingyan Jiang 1
- Bowen Li 1
- Rongwei Lu 1
- Ying Shen 1
- show all...