@inproceedings{kadlcik-stefanik-2024-self,
    title = "Self-training Language Models for Arithmetic Reasoning",
    author = "Kadl{\v{c}}{\'i}k, Marek  and
      {\v{S}}tef{\'a}nik, Michal",
    editor = "Al-Onaizan, Yaser  and
      Bansal, Mohit  and
      Chen, Yun-Nung",
    booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
    month = nov,
    year = "2024",
    address = "Miami, Florida, USA",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2024.findings-emnlp.721/",
    doi = "10.18653/v1/2024.findings-emnlp.721",
    pages = "12378--12386",
    abstract = "Recent language models achieve impressive results in tasks involving complex multistep reasoning, but scaling these capabilities further traditionally requires expensive collection of more annotated data.In this work, we explore the potential of improving models' reasoning capabilities without new data, merely using automated feedback to the validity of their predictions in arithmetic reasoning (self-training).In systematic experimentation across six different arithmetic reasoning datasets, we find that models can substantially improve in both single-round (offline) and online self-training, reaching a correct result in +13.9{\%} and +25.9{\%} more cases, respectively, underlining the importance of actuality of self-training feedback. We further find that in the single-round, offline self-training, traditional supervised training can deliver gains comparable to preference optimization, but in online self-training, preference optimization methods largely outperform supervised training thanks to their superior stability and robustness on unseen types of problems."
}Markdown (Informal)
[Self-training Language Models for Arithmetic Reasoning](https://preview.aclanthology.org/ingest-emnlp/2024.findings-emnlp.721/) (Kadlčík & Štefánik, Findings 2024)
ACL