@inproceedings{ju-etal-2025-domain,
title = "Domain Regeneration: How well do {LLM}s match syntactic properties of text domains?",
author = "Ju, Da and
Blix, Hagen and
Williams, Adina",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.120/",
pages = "2367--2388",
ISBN = "979-8-89176-256-5",
abstract = "Recent improvement in large language model performance have, in all likelihood, been accompanied by improvement in how well they can approximate the distribution of their training data. In this work, we explore the following question: which properties of text domains do LLMs faithfully approximate, and how well do they do so? Applying observational approaches familiar from corpus linguistics, we prompt a commonly used, opensource LLM to regenerate text from two domains of permissively licensed English text which are often contained in LLM training data{---}Wikipedia and news text. This regeneration paradigm allows us to investigate whether LLMs can faithfully match the original human text domains in a fairly semantically-controlled setting. We investigate varying levels of syntactic abstraction, from more simple properties like sentence length, and article readability, to more complex and higher order properties such as dependency tag distribution, parse depth, and parse complexity. We find that the majority of the regenerated distributions show a shifted mean, a lower standard deviation, and a reduction of the long tail, as compared to the human originals."
}
Markdown (Informal)
[Domain Regeneration: How well do LLMs match syntactic properties of text domains?](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.120/) (Ju et al., Findings 2025)
ACL