@inproceedings{nawrot-2023-nanot5,
title = "nano{T}5: Fast {\&} Simple Pre-training and Fine-tuning of T5 Models with Limited Resources",
author = "Nawrot, Piotr",
editor = "Tan, Liling and
Milajevs, Dmitrijs and
Chauhan, Geeticka and
Gwinnup, Jeremy and
Rippeth, Elijah",
booktitle = "Proceedings of the 3rd Workshop for Natural Language Processing Open Source Software (NLP-OSS 2023)",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.nlposs-1.11/",
doi = "10.18653/v1/2023.nlposs-1.11",
pages = "95--101",
abstract = "State-of-the-art language models like T5 have revolutionized the NLP landscape, but their computational demands hinder a large portion of the research community. To address this challenge, we present nanoT5, a specially-optimized PyTorch framework for efficient pre-training and fine-tuning of T5 models. Drawing on insights from optimizer differences and prioritizing efficiency, nanoT5 allows a T5-Base model to be pre-trained on a single GPU in just 16 hours, without any loss in performance. With the introduction of this open-source framework, we hope to widen the accessibility to language modelling research and cater to the community`s demand for more user-friendly T5 (Encoder-Decoder) implementations. We make our contributions, including configurations, codebase, pre-training insights, and pre-trained models, available to the public."
}
Markdown (Informal)
[nanoT5: Fast & Simple Pre-training and Fine-tuning of T5 Models with Limited Resources](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.nlposs-1.11/) (Nawrot, NLPOSS 2023)
ACL