Zaifu Zhan
2025
The Efficiency vs. Accuracy Trade-off: Optimizing RAG-Enhanced LLM Recommender Systems Using Multi-Head Early Exit
Huixue Zhou
|
Hengrui Gu
|
Zaifu Zhan
|
Xi Liu
|
Kaixiong Zhou
|
Yongkang Xiao
|
Mingfu Liang
|
Srinivas Prasad Govindan
|
Piyush Chawla
|
Jiyan Yang
|
Xiangfei Meng
|
Huayu Li
|
Buyun Zhang
|
Liang Luo
|
Wen-Yen Chen
|
Yiping Han
|
Bo Long
|
Rui Zhang
|
Tianlong Chen
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
The deployment of Large Language Models (LLMs) in recommender systems for Click-Through Rate (CTR) prediction requires a careful balance between computational efficiency and predictive accuracy. This paper introduces OptiRAG-Rec, a comprehensive framework that integrates Retrieval-Augmented Generation (RAG) with a novel multi-head early exit architecture to address both challenges. By leveraging Graph Convolutional Networks (GCNs) as efficient retrieval mechanisms, the framework significantly reduces data retrieval times while maintaining high model performance. Additionally, the multi-head early exit strategy dynamically terminates inference based on real-time predictive confidence assessments, enhancing responsiveness without sacrificing accuracy. Experimental results demonstrate that OptiRAG-Rec reduces computation time while preserving the precision required for reliable recommendations, establishing a new benchmark for efficient and accurate LLM deployment in recommendation.
Towards Better Multi-task Learning: A Framework for Optimizing Dataset Combinations in Large Language Models
Zaifu Zhan
|
Rui Zhang
Findings of the Association for Computational Linguistics: NAACL 2025
To efficiently select optimal dataset combinations for enhancing multi-task learning (MTL) performance in large language models, we proposed a novel framework that leverages a neural network to predict the best dataset combinations. The framework iteratively refines the selection, greatly improving efficiency, while being model-, dataset-, and domain-independent. Through experiments on 12 biomedical datasets across four tasks—named entity recognition, relation extraction, event extraction, and text classification—we demonstrate that our approach effectively identifies better combinations, even for tasks that may seem unpromising from a human perspective. This verifies that our framework provides a promising solution for maximizing MTL potential.
Search
Fix author
Co-authors
- Rui Zhang 2
- Piyush Chawla 1
- Wen-Yen Chen 1
- Tianlong Chen 1
- Srinivas Prasad Govindan 1
- show all...