Robustness to Modification with Shared Words in Paraphrase Identification

Zhouxing Shi, Minlie Huang


Abstract
Revealing the robustness issues of natural language processing models and improving their robustness is important to their performance under difficult situations. In this paper, we study the robustness of paraphrase identification models from a new perspective – via modification with shared words, and we show that the models have significant robustness issues when facing such modifications. To modify an example consisting of a sentence pair, we either replace some words shared by both sentences or introduce new shared words. We aim to construct a valid new example such that a target model makes a wrong prediction. To find a modification solution, we use beam search constrained by heuristic rules, and we leverage a BERT masked language model for generating substitution words compatible with the context. Experiments show that the performance of the target models has a dramatic drop on the modified examples, thereby revealing the robustness issue. We also show that adversarial training can mitigate this issue.
Anthology ID:
2020.findings-emnlp.16
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2020
Month:
November
Year:
2020
Address:
Online
Editors:
Trevor Cohn, Yulan He, Yang Liu
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
164–171
Language:
URL:
https://aclanthology.org/2020.findings-emnlp.16
DOI:
10.18653/v1/2020.findings-emnlp.16
Bibkey:
Cite (ACL):
Zhouxing Shi and Minlie Huang. 2020. Robustness to Modification with Shared Words in Paraphrase Identification. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 164–171, Online. Association for Computational Linguistics.
Cite (Informal):
Robustness to Modification with Shared Words in Paraphrase Identification (Shi & Huang, Findings 2020)
Copy Citation:
PDF:
https://preview.aclanthology.org/proper-vol2-ingestion/2020.findings-emnlp.16.pdf
Data
MRPCPAWS