@inproceedings{mishra-etal-2026-models,
title = "The Model{'}s Language Matters: A Comparative Privacy Analysis of {LLM}s",
author = "Mishra, Abhishek Kumar and
Boutet, Antoine and
Magnana, Lucas",
editor = "Demberg, Vera and
Inui, Kentaro and
Marquez, Llu{\'i}s",
booktitle = "Findings of the {A}ssociation for {C}omputational {L}inguistics: {EACL} 2026",
month = mar,
year = "2026",
address = "Rabat, Morocco",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.54/",
pages = "1038--1048",
ISBN = "979-8-89176-386-9",
abstract = "Large Language Models (LLMs) are increasingly deployed in multilingual settings that process sensitive data, yet their scale and linguistic variability can amplify privacy risks. While prior privacy evaluations focus predominantly on English, we investigate how language structure shapes privacy leakage in LLMs trained on English, Spanish, French, and Italian medical corpora. We quantify six corpus-level linguistic indicators and evaluate vulnerability under three attack families: extraction, counterfactual memorization, and membership inference. Across languages, we find that leakage systematically tracks structural properties: Italian exhibits the strongest exposure, consistent with its highest redundancy and longer lexical units, whereas English shows the clearest membership separability, aligning with its higher syntactic entropy and stronger surface-identifiable cues. In contrast, French and Spanish remain comparatively more resilient overall, aided by higher morphological complexity. These results provide quantitative evidence that language matters for privacy leakage, motivating language-aware and structure-adaptive privacy-preserving mechanisms for multilingual LLM deployments."
}Markdown (Informal)
[The Model’s Language Matters: A Comparative Privacy Analysis of LLMs](https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.54/) (Mishra et al., Findings 2026)
ACL