Context-Aware Membership Inference Attacks against Pre-trained Large Language Models

Hongyan Chang, Ali Shahin Shamsabadi, Kleomenis Katevas, Hamed Haddadi, Reza Shokri


Abstract
Membership Inference Attacks (MIAs) on pre-trained Large Language Models (LLMs) aim at determining if a data point was part of the model’s training set. Prior MIAs that are built for classification models fail at LLMs, due to ignoring the generative nature of LLMs across token sequences. In this paper, we present a novel attack on pre-trained LLMs that adapts MIA statistical tests to the perplexity dynamics of subsequences within a data point. Our method significantly outperforms prior approaches, revealing context-dependent memorization patterns in pre-trained LLMs.
Anthology ID:
2025.emnlp-main.370
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7299–7321
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.370/
DOI:
Bibkey:
Cite (ACL):
Hongyan Chang, Ali Shahin Shamsabadi, Kleomenis Katevas, Hamed Haddadi, and Reza Shokri. 2025. Context-Aware Membership Inference Attacks against Pre-trained Large Language Models. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 7299–7321, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Context-Aware Membership Inference Attacks against Pre-trained Large Language Models (Chang et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.370.pdf
Checklist:
 2025.emnlp-main.370.checklist.pdf