Archaeology at TSAR 2025 Shared Task Teaching Small Models to do CEFR Simplifications

Rares-Alexandru Roscan, Sergiu Nisioi


Abstract
Large language models (LLMs) have demonstrated strong performance in text simplification tasks, but their high computational cost and proprietary nature often limit practical use, especially in education. We explore open-source LLMs for CEFR-level text simplification. By reducing model size and computational requirements, our approach enables greater accessibility and deployment in educational environments. Our results show some of the lowest error rates in producing CEFR-compliant texts at TSAR 2025, using models with 8 billion and 1 billion parameters. Such approaches have the potential to democratize NLP technologies for real-world applications.
Anthology ID:
2025.tsar-1.22
Volume:
Proceedings of the Fourth Workshop on Text Simplification, Accessibility and Readability (TSAR 2025)
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Matthew Shardlow, Fernando Alva-Manchego, Kai North, Regina Stodden, Horacio Saggion, Nouran Khallaf, Akio Hayakawa
Venues:
TSAR | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
251–260
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.tsar-1.22/
DOI:
Bibkey:
Cite (ACL):
Rares-Alexandru Roscan and Sergiu Nisioi. 2025. Archaeology at TSAR 2025 Shared Task Teaching Small Models to do CEFR Simplifications. In Proceedings of the Fourth Workshop on Text Simplification, Accessibility and Readability (TSAR 2025), pages 251–260, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Archaeology at TSAR 2025 Shared Task Teaching Small Models to do CEFR Simplifications (Roscan & Nisioi, TSAR 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.tsar-1.22.pdf