@inproceedings{feng-etal-2024-promptexplainer,
title = "{P}rompt{E}xplainer: Explaining Language Models through Prompt-based Learning",
author = "Feng, Zijian and
Zhou, Hanzhang and
Zhu, Zixiao and
Mao, Kezhi",
editor = "Graham, Yvette and
Purver, Matthew",
booktitle = "Findings of the Association for Computational Linguistics: EACL 2024",
month = mar,
year = "2024",
address = "St. Julian{'}s, Malta",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.findings-eacl.60/",
pages = "882--895",
abstract = "Pretrained language models have become workhorses for various natural language processing (NLP) tasks, sparking a growing demand for enhanced interpretability and transparency. However, prevailing explanation methods, such as attention-based and gradient-based strategies, largely rely on linear approximations, potentially causing inaccuracies such as accentuating irrelevant input tokens. To mitigate the issue, we develop PromptExplainer, a novel method for explaining language models through prompt-based learning. PromptExplainer aligns the explanation process with the masked language modeling (MLM) task of pretrained language models and leverages the prompt-based learning framework for explanation generation. It disentangles token representations into the explainable embedding space using the MLM head and extracts discriminative features with a verbalizer to generate class-dependent explanations. Extensive experiments demonstrate that PromptExplainer significantly outperforms state-of-the-art explanation methods."
}
Markdown (Informal)
[PromptExplainer: Explaining Language Models through Prompt-based Learning](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.findings-eacl.60/) (Feng et al., Findings 2024)
ACL