FiMMIA: scaling semantic perturbation-based membership inference across modalities

Anton Emelyanov, Sergei Kudriashov, Alena Fenogenova


Abstract
Membership Inference Attacks (MIAs) aim to determine whether a specific data point was included in the training set of a target model. Although there are have been numerous methods developed for detecting data contamination in large language models (LLMs), their performance on multimodal LLMs (MLLMs) falls short due to the instabilities introduced through multimodal component adaptation and possible distribution shifts across multiple inputs. In this work, we investigate multimodal membership inference and address two issues: first, by identifying distribution shifts in the existing datasets, and second, by releasing an extended baseline pipeline to detect them. We also generalize the perturbation-based membership inference methods to MLLMs and release FiMMIA — a modular Framework for Multimodal MIA. We propose to train a neural networks to analyze the target model’s behavior on perturbed inputs, capturing interactions between semantic domains and loss values on members and non-members in the local neighborhood of each sample. Comprehensive evaluations on various fine-tuned multimodal models demonstrate the effectiveness of our perturbation-based membership inference attacks in multimodal settings.
Anthology ID:
2026.eacl-demo.11
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 3: System Demonstrations)
Month:
March
Year:
2026
Address:
Rabat, Marocco
Editors:
Danilo Croce, Jochen Leidner, Nafise Sadat Moosavi
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
139–153
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-demo.11/
DOI:
Bibkey:
Cite (ACL):
Anton Emelyanov, Sergei Kudriashov, and Alena Fenogenova. 2026. FiMMIA: scaling semantic perturbation-based membership inference across modalities. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 3: System Demonstrations), pages 139–153, Rabat, Marocco. Association for Computational Linguistics.
Cite (Informal):
FiMMIA: scaling semantic perturbation-based membership inference across modalities (Emelyanov et al., EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-demo.11.pdf