@inproceedings{kirstain-etal-2022-examples,
    title = "A Few More Examples May Be Worth Billions of Parameters",
    author = "Kirstain, Yuval  and
      Lewis, Patrick  and
      Riedel, Sebastian  and
      Levy, Omer",
    editor = "Goldberg, Yoav  and
      Kozareva, Zornitsa  and
      Zhang, Yue",
    booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022",
    month = dec,
    year = "2022",
    address = "Abu Dhabi, United Arab Emirates",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2022.findings-emnlp.72/",
    doi = "10.18653/v1/2022.findings-emnlp.72",
    pages = "1017--1029",
    abstract = "We investigate the dynamics of increasing the number of model parameters versus the number of labeled examples across a wide variety of tasks. Our exploration reveals that while scaling parameters consistently yields performance improvements, the contribution of additional examples highly depends on the task{'}s format. Specifically, in open question answering tasks, enlarging the training set does not improve performance. In contrast, classification, extractive question answering, and multiple choice tasks benefit so much from additional examples that collecting a few hundred examples is often ``worth'' billions of parameters. We hypothesize that unlike open question answering, which involves recalling specific information, solving strategies for tasks with a more restricted output space transfer across examples, and can therefore be learned with small amounts of labeled data."
}Markdown (Informal)
[A Few More Examples May Be Worth Billions of Parameters](https://preview.aclanthology.org/ingest-emnlp/2022.findings-emnlp.72/) (Kirstain et al., Findings 2022)
ACL
- Yuval Kirstain, Patrick Lewis, Sebastian Riedel, and Omer Levy. 2022. A Few More Examples May Be Worth Billions of Parameters. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 1017–1029, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.