@inproceedings{lai-etal-2019-many,
title = "Many Faces of Feature Importance: Comparing Built-in and Post-hoc Feature Importance in Text Classification",
author = "Lai, Vivian and
Cai, Zheng and
Tan, Chenhao",
editor = "Inui, Kentaro and
Jiang, Jing and
Ng, Vincent and
Wan, Xiaojun",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/D19-1046/",
doi = "10.18653/v1/D19-1046",
pages = "486--495",
abstract = "Feature importance is commonly used to explain machine predictions. While feature importance can be derived from a machine learning model with a variety of methods, the consistency of feature importance via different methods remains understudied. In this work, we systematically compare feature importance from built-in mechanisms in a model such as attention values and post-hoc methods that approximate model behavior such as LIME. Using text classification as a testbed, we find that 1) no matter which method we use, important features from traditional models such as SVM and XGBoost are more similar with each other, than with deep learning models; 2) post-hoc methods tend to generate more similar important features for two models than built-in methods. We further demonstrate how such similarity varies across instances. Notably, important features do not always resemble each other better when two models agree on the predicted label than when they disagree."
}
Markdown (Informal)
[Many Faces of Feature Importance: Comparing Built-in and Post-hoc Feature Importance in Text Classification](https://preview.aclanthology.org/landing_page/D19-1046/) (Lai et al., EMNLP-IJCNLP 2019)
ACL