ART: Attention-Regularized Transformers for Multi-Modal Robustness

Mohammed Bouri, Mohammed Erradi, Adnane Saoud


Abstract
Transformers have become the standard in Natural Language Processing (NLP) and Computer Vision (CV) due to their strong performance, yet they remain highly sensitive to small input changes, often referred to as adversarial attacks, such as synonym swaps in text or pixel-level perturbations in images. These adversarial attacks can mislead predictions, while existing defenses are often domain-specific or lack formal robustness guarantees. We propose the Attention-Regularized Transformer (ART), a framework that enhances robustness across modalities. ART builds on the Attention Sensitivity Tensor (AST), which quantifies the effect of input perturbations on attention outputs. By incorporating an AST-based regularizer into training, ART encourages stable attention maps under adversarial perturbations in both text and image tasks. We evaluate ART on IMDB, QNLI, CIFAR-10, CIFAR-100, and Imagenette. Results show consistent robustness gains over strong baselines such as FreeLB and DSRM: up to +36.9% robust accuracy on IMDB and QNLI, and +5–25% on image benchmarks across multiple Vision Transformer (ViT) architectures, while maintaining or improving clean accuracy. ART is also highly efficient, training over 10× faster than adversarial methods on text and requiring only 1.25× the cost of standard training on images, compared to 1.5–5.5× for recent robust ViTs. Codes are available at [https://github.com/cliclab-um6p/ART](https://github.com/cliclab-um6p/ART)
Anthology ID:
2026.findings-eacl.292
Volume:
Findings of the Association for Computational Linguistics: EACL 2026
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5510–5535
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.292/
DOI:
Bibkey:
Cite (ACL):
Mohammed Bouri, Mohammed Erradi, and Adnane Saoud. 2026. ART: Attention-Regularized Transformers for Multi-Modal Robustness. In Findings of the Association for Computational Linguistics: EACL 2026, pages 5510–5535, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
ART: Attention-Regularized Transformers for Multi-Modal Robustness (Bouri et al., Findings 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.292.pdf
Checklist:
 2026.findings-eacl.292.checklist.pdf