@inproceedings{harbecke-alt-2020-considering,
title = "Considering Likelihood in {NLP} Classification Explanations with Occlusion and Language Modeling",
author = "Harbecke, David and
Alt, Christoph",
editor = "Rijhwani, Shruti and
Liu, Jiangming and
Wang, Yizhong and
Dror, Rotem",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2020.acl-srw.16/",
doi = "10.18653/v1/2020.acl-srw.16",
pages = "111--117",
abstract = "Recently, state-of-the-art NLP models gained an increasing syntactic and semantic understanding of language, and explanation methods are crucial to understand their decisions. Occlusion is a well established method that provides explanations on discrete language data, e.g. by removing a language unit from an input and measuring the impact on a model`s decision. We argue that current occlusion-based methods often produce invalid or syntactically incorrect language data, neglecting the improved abilities of recent NLP models. Furthermore, gradient-based explanation methods disregard the discrete distribution of data in NLP. Thus, we propose OLM: a novel explanation method that combines occlusion and language models to sample valid and syntactically correct replacements with high likelihood, given the context of the original input. We lay out a theoretical foundation that alleviates these weaknesses of other explanation methods in NLP and provide results that underline the importance of considering data likelihood in occlusion-based explanation."
}
Markdown (Informal)
[Considering Likelihood in NLP Classification Explanations with Occlusion and Language Modeling](https://preview.aclanthology.org/jlcl-multiple-ingestion/2020.acl-srw.16/) (Harbecke & Alt, ACL 2020)
ACL