@inproceedings{marecek-rosa-2019-balustrades,
title = "From Balustrades to Pierre Vinken: Looking for Syntax in Transformer Self-Attentions",
author = "Mare{\v{c}}ek, David and
Rosa, Rudolf",
editor = "Linzen, Tal and
Chrupa{\l}a, Grzegorz and
Belinkov, Yonatan and
Hupkes, Dieuwke",
booktitle = "Proceedings of the 2019 ACL Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP",
month = aug,
year = "2019",
address = "Florence, Italy",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/W19-4827/",
doi = "10.18653/v1/W19-4827",
pages = "263--275",
abstract = "We inspect the multi-head self-attention in Transformer NMT encoders for three source languages, looking for patterns that could have a syntactic interpretation. In many of the attention heads, we frequently find sequences of consecutive states attending to the same position, which resemble syntactic phrases. We propose a transparent deterministic method of quantifying the amount of syntactic information present in the self-attentions, based on automatically building and evaluating phrase-structure trees from the phrase-like sequences. We compare the resulting trees to existing constituency treebanks, both manually and by computing precision and recall."
}
Markdown (Informal)
[From Balustrades to Pierre Vinken: Looking for Syntax in Transformer Self-Attentions](https://preview.aclanthology.org/jlcl-multiple-ingestion/W19-4827/) (Mareček & Rosa, BlackboxNLP 2019)
ACL