SMEC:Rethinking Matryoshka Representation Learning for Retrieval Embedding Compression

Biao Zhang, Lixin Chen, Tong Liu, Bo Zheng


Abstract
Large language models (LLMs) generate high-dimensional embeddings that capture rich semantic and syntactic information. However, high-dimensional embeddings exacerbate computational complexity and storage requirements, thereby hindering practical deployment. To address these challenges, we propose a novel training framework named Sequential Matryoshka Embedding Compression (SMEC). This framework introduces the Sequential Matryoshka Representation Learning(SMRL) method to mitigate gradient variance during training, the Adaptive Dimension Selection (ADS) module to reduce information degradation during dimension pruning, and the Selectable Cross-batch Memory (S-XBM) module to enhance unsupervised learning between high- and low-dimensional embeddings. Experiments on image, text, and multimodal datasets demonstrate that SMEC achieves significant dimensionality reduction while maintaining performance. For instance, on the BEIR dataset, our approach improves the performance of compressed LLM2Vec embeddings (256 dimensions) by 1.1 points and 2.7 points compared to the Matryoshka-Adaptor and Search-Adaptor models, respectively.
Anthology ID:
2025.emnlp-main.1332
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
26220–26233
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1332/
DOI:
Bibkey:
Cite (ACL):
Biao Zhang, Lixin Chen, Tong Liu, and Bo Zheng. 2025. SMEC:Rethinking Matryoshka Representation Learning for Retrieval Embedding Compression. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 26220–26233, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
SMEC:Rethinking Matryoshka Representation Learning for Retrieval Embedding Compression (Zhang et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1332.pdf
Checklist:
 2025.emnlp-main.1332.checklist.pdf