@inproceedings{howard-ruder-2018-universal,
title = "Universal Language Model Fine-tuning for Text Classification",
author = "Howard, Jeremy and
Ruder, Sebastian",
editor = "Gurevych, Iryna and
Miyao, Yusuke",
booktitle = "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2018",
address = "Melbourne, Australia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/P18-1031/",
doi = "10.18653/v1/P18-1031",
pages = "328--339",
abstract = "Inductive transfer learning has greatly impacted computer vision, but existing approaches in NLP still require task-specific modifications and training from scratch. We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP, and introduce techniques that are key for fine-tuning a language model. Our method significantly outperforms the state-of-the-art on six text classification tasks, reducing the error by 18-24{\%} on the majority of datasets. Furthermore, with only 100 labeled examples, it matches the performance of training from scratch on 100 times more data. We open-source our pretrained models and code."
}
Markdown (Informal)
[Universal Language Model Fine-tuning for Text Classification](https://preview.aclanthology.org/jlcl-multiple-ingestion/P18-1031/) (Howard & Ruder, ACL 2018)
ACL