@inproceedings{siddhant-lipton-2018-deep,
title = "Deep {B}ayesian Active Learning for Natural Language Processing: Results of a Large-Scale Empirical Study",
author = "Siddhant, Aditya and
Lipton, Zachary C.",
editor = "Riloff, Ellen and
Chiang, David and
Hockenmaier, Julia and
Tsujii, Jun{'}ichi",
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
month = oct # "-" # nov,
year = "2018",
address = "Brussels, Belgium",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/D18-1318/",
doi = "10.18653/v1/D18-1318",
pages = "2904--2909",
abstract = "Several recent papers investigate Active Learning (AL) for mitigating the data dependence of deep learning for natural language processing. However, the applicability of AL to real-world problems remains an open question. While in supervised learning, practitioners can try many different methods, evaluating each against a validation set before selecting a model, AL affords no such luxury. Over the course of one AL run, an agent annotates its dataset exhausting its labeling budget. Thus, given a new task, we have no opportunity to compare models and acquisition functions. This paper provides a large-scale empirical study of deep active learning, addressing multiple tasks and, for each, multiple datasets, multiple models, and a full suite of acquisition functions. We find that across all settings, Bayesian active learning by disagreement, using uncertainty estimates provided either by Dropout or Bayes-by-Backprop significantly improves over i.i.d. baselines and usually outperforms classic uncertainty sampling."
}
Markdown (Informal)
[Deep Bayesian Active Learning for Natural Language Processing: Results of a Large-Scale Empirical Study](https://preview.aclanthology.org/jlcl-multiple-ingestion/D18-1318/) (Siddhant & Lipton, EMNLP 2018)
ACL