@inproceedings{tian-etal-2021-bert,
title = "How does {BERT} process disfluency?",
author = "Tian, Ye and
Nieradzik, Tim and
Jalali, Sepehr and
Shiu, Da-shan",
editor = "Li, Haizhou and
Levow, Gina-Anne and
Yu, Zhou and
Gupta, Chitralekha and
Sisman, Berrak and
Cai, Siqi and
Vandyke, David and
Dethlefs, Nina and
Wu, Yan and
Li, Junyi Jessy",
booktitle = "Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue",
month = jul,
year = "2021",
address = "Singapore and Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2021.sigdial-1.22/",
doi = "10.18653/v1/2021.sigdial-1.22",
pages = "208--217",
abstract = "Natural conversations are filled with disfluencies. This study investigates if and how BERT understands disfluency with three experiments: (1) a behavioural study using a downstream task, (2) an analysis of sentence embeddings and (3) an analysis of the attention mechanism on disfluency. The behavioural study shows that without fine-tuning on disfluent data, BERT does not suffer significant performance loss when presented disfluent compared to fluent inputs (exp1). Analysis on sentence embeddings of disfluent and fluent sentence pairs reveals that the deeper the layer, the more similar their representation (exp2). This indicates that deep layers of BERT become relatively invariant to disfluency. We pinpoint attention as a potential mechanism that could explain this phenomenon (exp3). Overall, the study suggests that BERT has knowledge of disfluency structure. We emphasise the potential of using BERT to understand natural utterances without disfluency removal."
}
Markdown (Informal)
[How does BERT process disfluency?](https://preview.aclanthology.org/jlcl-multiple-ingestion/2021.sigdial-1.22/) (Tian et al., SIGDIAL 2021)
ACL
- Ye Tian, Tim Nieradzik, Sepehr Jalali, and Da-shan Shiu. 2021. How does BERT process disfluency?. In Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 208–217, Singapore and Online. Association for Computational Linguistics.