Demystifying Mixed Outcomes of Self-Training: Pre-training Analyses on Non-Toy LLMs

Yusuke Nakamura, Hirokazu Kiyomaru, Chaoran Liu, Shuhei Kurita, Daisuke Kawahara


Abstract
We investigate whether large language models (LLMs) can improve through recursive training on self-generated text, a topic where prior studies report conflicting outcomes: some find evidence of performance gains (i.e., self-improvement), while others observe performance degradation (i.e., model collapse). To clarify this discrepancy, we use the OLMo-2 models as non-toy LLMs and perform multiple rounds of continual pre-training using self-generated text with different prompting strategies and data filtering. Our experiments show that naive recursive self-training does not improve either perplexity or downstream task performance, regardless of model size. These results suggest that model collapse observed in naive recursive training is inherent to the training procedure itself, while self-improvement likely owes its success not to the model’s autonomous refinement but to human-designed, strategic synthetic pipelines that inject external intelligence.
Anthology ID:
2026.findings-eacl.213
Volume:
Findings of the Association for Computational Linguistics: EACL 2026
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4107–4113
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.213/
DOI:
Bibkey:
Cite (ACL):
Yusuke Nakamura, Hirokazu Kiyomaru, Chaoran Liu, Shuhei Kurita, and Daisuke Kawahara. 2026. Demystifying Mixed Outcomes of Self-Training: Pre-training Analyses on Non-Toy LLMs. In Findings of the Association for Computational Linguistics: EACL 2026, pages 4107–4113, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Demystifying Mixed Outcomes of Self-Training: Pre-training Analyses on Non-Toy LLMs (Nakamura et al., Findings 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.213.pdf
Checklist:
 2026.findings-eacl.213.checklist.pdf