@inproceedings{kissane-etal-2025-probing,
title = "Probing Internal Representations of Multi-Word Verbs in Large Language Models",
author = "Kissane, Hassane and
Schilling, Achim and
Krauss, Patrick",
editor = {Ojha, Atul Kr. and
Giouli, Voula and
Mititelu, Verginica Barbu and
Constant, Mathieu and
Korvel, Gra{\v{z}}ina and
Do{\u{g}}ru{\"o}z, A. Seza and
Rademaker, Alexandre},
booktitle = "Proceedings of the 21st Workshop on Multiword Expressions (MWE 2025)",
month = may,
year = "2025",
address = "Albuquerque, New Mexico, U.S.A.",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.mwe-1.2/",
pages = "7--13",
ISBN = "979-8-89176-243-5",
abstract = "This study investigates the internal representations of verb-particle combinations, called multi-word verbs, within transformer-based large language models (LLMs), specifically examining how these models capture lexical and syntactic properties at different neural network layers. Using the BERT architecture, we analyze the representations of its layers for two different verb-particle constructions: phrasal verbs like ``give up'' and prepositional verbs like ``look at''. Our methodology includes training probing classifiers on the model output to classify these categories at both word and sentence levels. The results indicate that the model{'}s middle layers achieve the highest classification accuracies. To further analyze the nature of these distinctions, we conduct a data separability test using the Generalized Discrimination Value (GDV). While GDV results show weak linear separability between the two verb types, probing classifiers still achieve high accuracy, suggesting that representations of these linguistic categories may be ``non-linearly separable''. This aligns with previous research indicating that linguistic distinctions in neural networks are not always encoded in a linearly separable manner. These findings computationally support usage-based claims on the representation of verb-particle constructions and highlight the complex interaction between neural network architectures and linguistic structures."
}
Markdown (Informal)
[Probing Internal Representations of Multi-Word Verbs in Large Language Models](https://preview.aclanthology.org/fix-sig-urls/2025.mwe-1.2/) (Kissane et al., MWE 2025)
ACL