@inproceedings{pavlova-makhlouf-2026-mosaic,
title = "{MOSAIC}: Masked Objective with Selective Adaptation for In-domain Contrastive Learning",
author = "Pavlova, Vera and
Makhlouf, Mohammed",
editor = "Demberg, Vera and
Inui, Kentaro and
Marquez, Llu{\'i}s",
booktitle = "Findings of the {A}ssociation for {C}omputational {L}inguistics: {EACL} 2026",
month = mar,
year = "2026",
address = "Rabat, Morocco",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.197/",
pages = "3792--3807",
ISBN = "979-8-89176-386-9",
abstract = "We introduce MOSAIC (Masked Objective with Selective Adaptation for In-domain Contrastive learning), a multi-stage framework for domain adaptation of text embedding models that incorporates joint domain-specific masked supervision. Our approach addresses the challenges of adapting large-scale general-domain text embedding models to specialized domains. By jointly optimizing masked language modeling (MLM) and contrastive objectives within a unified training pipeline, our method enables effective learning of domain-relevant representations while preserving the robust semantic discrimination properties of the original model. We empirically validate our approach on both high-resource and low-resource domains, achieving improvements up to 13.4{\%} in NDCG@10 (Normalized Discounted Cumulative Gain) over strong general-domain baselines. Comprehensive ablation studies further demonstrate the effectiveness of each component, highlighting the importance of balanced joint supervision and staged adaptation."
}Markdown (Informal)
[MOSAIC: Masked Objective with Selective Adaptation for In-domain Contrastive Learning](https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.197/) (Pavlova & Makhlouf, Findings 2026)
ACL