@inproceedings{wang-etal-2021-calibrate-listeners,
title = "Calibrate your listeners! Robust communication-based training for pragmatic speakers",
author = "Wang, Rose and
White, Julia and
Mu, Jesse and
Goodman, Noah",
editor = "Moens, Marie-Francine and
Huang, Xuanjing and
Specia, Lucia and
Yih, Scott Wen-tau",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
month = nov,
year = "2021",
address = "Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2021.findings-emnlp.83/",
doi = "10.18653/v1/2021.findings-emnlp.83",
pages = "977--984",
abstract = "To be good conversational partners, natural language processing (NLP) systems should be trained to produce contextually useful utterances. Prior work has investigated training NLP systems with communication-based objectives, where a neural listener stands in as a communication partner. However, these systems commonly suffer from semantic drift where the learned language diverges radically from natural language. We propose a method that uses a population of neural listeners to regularize speaker training. We first show that language drift originates from the poor uncertainty calibration of a neural listener, which makes high-certainty predictions on novel sentences. We explore ensemble- and dropout-based populations of listeners and find that the former results in better uncertainty quantification. We evaluate both population-based objectives on reference games, and show that the ensemble method with better calibration enables the speaker to generate pragmatic utterances while scaling to a large vocabulary and generalizing to new games and listeners."
}
Markdown (Informal)
[Calibrate your listeners! Robust communication-based training for pragmatic speakers](https://preview.aclanthology.org/fix-sig-urls/2021.findings-emnlp.83/) (Wang et al., Findings 2021)
ACL