Abstract
Zelda Rose is a command line interface for pretraining transformer-based models. Its purpose is to enable an easy start for users interested in training these ubiquitous models, but unable or unwilling to engage with more comprehensive — but more complex — frameworks and the complex interactions between libraries for managing models, datasets and computations. Training a model requires no code on the user’s part and produce models directly compatible with the HuggingFace ecosystem, allowing quick and easy distribution and reuse. A particular care is given to lowering the cost of maintainability and future-proofing, by making the code as modular as possible and taking advantage of third-party libraries to limit ad-hoc code to the strict minimum.- Anthology ID:
- 2023.nlposs-1.6
- Volume:
- Proceedings of the 3rd Workshop for Natural Language Processing Open Source Software (NLP-OSS 2023)
- Month:
- December
- Year:
- 2023
- Address:
- Singapore
- Editors:
- Liling Tan, Dmitrijs Milajevs, Geeticka Chauhan, Jeremy Gwinnup, Elijah Rippeth
- Venues:
- NLPOSS | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 54–58
- Language:
- URL:
- https://aclanthology.org/2023.nlposs-1.6
- DOI:
- 10.18653/v1/2023.nlposs-1.6
- Cite (ACL):
- Loïc Grobol. 2023. Zelda Rose: a tool for hassle-free training of transformer models. In Proceedings of the 3rd Workshop for Natural Language Processing Open Source Software (NLP-OSS 2023), pages 54–58, Singapore. Association for Computational Linguistics.
- Cite (Informal):
- Zelda Rose: a tool for hassle-free training of transformer models (Grobol, NLPOSS-WS 2023)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-1/2023.nlposs-1.6.pdf