Yuqing Liu
2025
LLMInit: A Free Lunch from Large Language Models for Selective Initialization of Recommendation
Weizhi Zhang
|
Liangwei Yang
|
Wooseong Yang
|
Henry Peng Zou
|
Yuqing Liu
|
Ke Xu
|
Sourav Medya
|
Philip S. Yu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Collaborative filtering (CF) is widely adopted in industrial recommender systems (RecSys) for modeling user-item interactions across numerous applications, but often struggles with cold-start and data-sparse scenarios. Recent advancements in pre-trained large language models (LLMs) with rich semantic knowledge, offer promising solutions to these challenges. However, deploying LLMs at scale is hindered by their significant computational demands and latency. In this paper, we propose a novel and scalable LLM-RecSys framework, LLMInit, designed to integrate pretrained LLM embeddings into CF models through selective initialization strategies. Specifically, we identify the embedding collapse issue observed when CF models scale and match the large embedding sizes in LLMs and avoid the problem by introducing efficient sampling methods, including, random, uniform, and variance-based selections. Comprehensive experiments conducted on multiple real-world datasets demonstrate that LLMInit significantly improves recommendation performance while maintaining low computational costs, offering a practical and scalable solution for industrial applications. To facilitate industry adoption and promote future research, we provide open-source access to our implementation at https://github.com/DavidZWZ/LLMInit.
Search
Fix author
Co-authors
- Sourav Medya 1
- Ke Xu 1
- Liangwei Yang 1
- Wooseong Yang 1
- Philip S. Yu 1
- show all...