@inproceedings{antebi-etal-2025-tag,
title = "Tag{\&}Tab: Pretraining Data Detection in Large Language Models Using Keyword-Based Membership Inference Attack",
author = "Antebi, Sagiv and
Habler, Edan and
Shabtai, Asaf and
Elovici, Yuval",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.283/",
doi = "10.18653/v1/2025.findings-emnlp.283",
pages = "5273--5286",
ISBN = "979-8-89176-335-7",
abstract = "Large language models (LLMs) have become essential tools for digital task assistance. Their training relies heavily on the collection of vast amounts of data, which may include copyright-protected or sensitive information. Recent studies on detecting pretraining data in LLMs have primarily focused on sentence- or paragraph-level membership inference attacks (MIAs), usually involving probability analysis of the target model{'}s predicted tokens. However, these methods often exhibit poor accuracy, failing to account for the semantic importance of textual content and word significance. To address these shortcomings, we propose Tag{\&}Tab, a novel approach for detecting data used in LLM pretraining. Our method leverages established natural language processing (NLP) techniques to tag keywords in the input text, a process we term Tagging. Then, the LLM is used to obtain probabilities for these keywords and calculate their average log-likelihood to determine input text membership, a process we refer to as Tabbing. Our experiments on four benchmark datasets (BookMIA, MIMIR, PatentMIA, and the Pile) and several open-source LLMs of varying sizes demonstrate an average increase in AUC scores ranging from 5.3{\%} to 17.6{\%} over state-of-the-art methods. Tag{\&}Tab not only sets a new standard for data leakage detection in LLMs, but its outstanding performance is a testament to the importance of words in MIAs on LLMs."
}Markdown (Informal)
[Tag&Tab: Pretraining Data Detection in Large Language Models Using Keyword-Based Membership Inference Attack](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.283/) (Antebi et al., Findings 2025)
ACL