@inproceedings{li-etal-2025-conan,
title = "Conan-Embedding-v2: Training an {LLM} from Scratch for Text Embeddings",
author = "Li, Shiyu and
Tang, Yang and
Liu, Ruijie and
Chen, Shi-Zhe and
Chen, Xi",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-luhme/2025.emnlp-main.758/",
doi = "10.18653/v1/2025.emnlp-main.758",
pages = "15011--15027",
ISBN = "979-8-89176-332-6",
abstract = "Large language models (LLMs) have recently demonstrated excellent performance in text embedding tasks. Previous work usually use LoRA to fine-tune existing LLMs, which are limited by the data and training gap between LLMs and embedding models. In this work, we introduce Conan-embedding-v2, a new 1.4B-parameter LLM trained from scratch and fine-tuned as a text embedder. First, we add news data and multilingual pairs for LLM pretraining to bridge the data gap. Based on this, we propose a cross-lingual retrieval dataset that enables the LLM to better integrate embeddings across different languages. Second, whereas LLMs use a causal mask with token-level loss, embedding models use a bidirectional mask with sentence-level loss. This training gap makes full fine-tuning less effective than LoRA. We introduce a soft-masking mechanism to gradually transition between these two types of masks, enabling the model to learn more comprehensive representations. Based on this, we propose a dynamic hard negative mining method that exposes the model to more difficult negative examples throughout the training process. Being intuitive and effective, with only approximately 1.4B parameters, Conan-embedding-v2 achieves SOTA performance on both the Massive Text Embedding Benchmark (MTEB) and Chinese MTEB (May 19, 2025)."
}Markdown (Informal)
[Conan-Embedding-v2: Training an LLM from Scratch for Text Embeddings](https://preview.aclanthology.org/ingest-luhme/2025.emnlp-main.758/) (Li et al., EMNLP 2025)
ACL