@inproceedings{ravfogel-etal-2021-counterfactual,
title = "Counterfactual Interventions Reveal the Causal Effect of Relative Clause Representations on Agreement Prediction",
author = "Ravfogel, Shauli and
Prasad, Grusha and
Linzen, Tal and
Goldberg, Yoav",
editor = "Bisazza, Arianna and
Abend, Omri",
booktitle = "Proceedings of the 25th Conference on Computational Natural Language Learning",
month = nov,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2021.conll-1.15/",
doi = "10.18653/v1/2021.conll-1.15",
pages = "194--209",
abstract = "When language models process syntactically complex sentences, do they use their representations of syntax in a manner that is consistent with the grammar of the language? We propose AlterRep, an intervention-based method to address this question. For any linguistic feature of a given sentence, AlterRep generates counterfactual representations by altering how the feature is encoded, while leaving in- tact all other aspects of the original representation. By measuring the change in a model{'}s word prediction behavior when these counterfactual representations are substituted for the original ones, we can draw conclusions about the causal effect of the linguistic feature in question on the model{'}s behavior. We apply this method to study how BERT models of different sizes process relative clauses (RCs). We find that BERT variants use RC boundary information during word prediction in a manner that is consistent with the rules of English grammar; this RC boundary information generalizes to a considerable extent across different RC types, suggesting that BERT represents RCs as an abstract linguistic category."
}
Markdown (Informal)
[Counterfactual Interventions Reveal the Causal Effect of Relative Clause Representations on Agreement Prediction](https://preview.aclanthology.org/fix-sig-urls/2021.conll-1.15/) (Ravfogel et al., CoNLL 2021)
ACL