Abstract
Recent improvements in the predictive quality of natural language processing systems are often dependent on a substantial increase in the number of model parameters. This has led to various attempts of compressing such models, but existing methods have not considered the differences in the predictive power of various model components or in the generalizability of the compressed models. To understand the connection between model compression and out-of-distribution generalization, we define the task of compressing language representation models such that they perform best in a domain adaptation setting. We choose to address this problem from a causal perspective, attempting to estimate the average treatment effect (ATE) of a model component, such as a single layer, on the model’s predictions. Our proposed ATE-guided Model Compression scheme (AMoC), generates many model candidates, differing by the model components that were removed. Then, we select the best candidate through a stepwise regression model that utilizes the ATE to predict the expected performance on the target domain. AMoC outperforms strong baselines on dozens of domain pairs across three text classification and sequence tagging tasks.1- Anthology ID:
- 2021.tacl-1.80
- Volume:
- Transactions of the Association for Computational Linguistics, Volume 9
- Month:
- Year:
- 2021
- Address:
- Cambridge, MA
- Editors:
- Brian Roark, Ani Nenkova
- Venue:
- TACL
- SIG:
- Publisher:
- MIT Press
- Note:
- Pages:
- 1355–1373
- Language:
- URL:
- https://aclanthology.org/2021.tacl-1.80
- DOI:
- 10.1162/tacl_a_00431
- Cite (ACL):
- Guy Rotman, Amir Feder, and Roi Reichart. 2021. Model Compression for Domain Adaptation through Causal Effect Estimation. Transactions of the Association for Computational Linguistics, 9:1355–1373.
- Cite (Informal):
- Model Compression for Domain Adaptation through Causal Effect Estimation (Rotman et al., TACL 2021)
- PDF:
- https://preview.aclanthology.org/ingest-acl-2023-videos/2021.tacl-1.80.pdf