@inproceedings{chakraborty-etal-2023-systematic,
title = "Systematic Generalization by Finetuning? Analyzing Pretrained Language Models Using Constituency Tests",
author = "Chakraborty, Aishik and
Cheung, Jackie CK and
O{'}Donnell, Timothy J.",
editor = "Belinkov, Yonatan and
Hao, Sophie and
Jumelet, Jaap and
Kim, Najoung and
McCarthy, Arya and
Mohebbi, Hosein",
booktitle = "Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.blackboxnlp-1.27/",
doi = "10.18653/v1/2023.blackboxnlp-1.27",
pages = "357--366",
abstract = "Constituents are groups of words that behave as a syntactic unit. Many linguistic phenomena (e.g., question formation, diathesis alternations) require the manipulation and rearrangement of constituents in a sentence. In this paper, we investigate how different finetuning setups affect the ability of pretrained sequence-to-sequence language models such as BART and T5 to replicate constituency tests {---} transformations that involve manipulating constituents in a sentence. We design multiple evaluation settings by varying the combinations of constituency tests and sentence types that a model is exposed to during finetuning. We show that models can replicate a linguistic transformation on a specific type of sentence that they saw during finetuning, but performance degrades substantially in other settings, showing a lack of systematic generalization. These results suggest that models often learn to manipulate sentences at a surface level unrelated to the constituent-level syntactic structure, for example by copying the first word of a sentence. These results may partially explain the brittleness of pretrained language models in downstream tasks."
}
Markdown (Informal)
[Systematic Generalization by Finetuning? Analyzing Pretrained Language Models Using Constituency Tests](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.blackboxnlp-1.27/) (Chakraborty et al., BlackboxNLP 2023)
ACL