@inproceedings{zhu-etal-2022-improving,
title = "Improving Robustness of Language Models from a Geometry-aware Perspective",
author = "Zhu, Bin and
Gu, Zhaoquan and
Wang, Le and
Chen, Jinyin and
Xuan, Qi",
editor = "Muresan, Smaranda and
Nakov, Preslav and
Villavicencio, Aline",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2022.findings-acl.246/",
doi = "10.18653/v1/2022.findings-acl.246",
pages = "3115--3125",
abstract = "Recent studies have found that removing the norm-bounded projection and increasing search steps in adversarial training can significantly improve robustness. However, we observe that a too large number of search steps can hurt accuracy. We aim to obtain strong robustness efficiently using fewer steps. Through a toy experiment, we find that perturbing the clean data to the decision boundary but not crossing it does not degrade the test accuracy. Inspired by this, we propose friendly adversarial data augmentation (FADA) to generate friendly adversarial data. On top of FADA, we propose geometry-aware adversarial training (GAT) to perform adversarial training on friendly adversarial data so that we can save a large number of search steps. Comprehensive experiments across two widely used datasets and three pre-trained language models demonstrate that GAT can obtain stronger robustness via fewer steps. In addition, we provide extensive empirical results and in-depth analyses on robustness to facilitate future studies."
}
Markdown (Informal)
[Improving Robustness of Language Models from a Geometry-aware Perspective](https://preview.aclanthology.org/add-emnlp-2024-awards/2022.findings-acl.246/) (Zhu et al., Findings 2022)
ACL