2025
pdf
bib
abs
User Feedback Alignment for LLM-powered Exploration in Large-scale Recommendation Systems
Jianling Wang
|
Yifan Liu
|
Yinghao Sun
|
Xuejian Ma
|
Yueqi Wang
|
He Ma
|
Zhengyang Su
|
Minmin Chen
|
Mingyan Gao
|
Onkar Dalal
|
Ed H. Chi
|
Lichan Hong
|
Ningren Han
|
Haokai Lu
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 6: Industry Track)
Exploration, the act of broadening user experiences beyond their established preferences, is challenging in large-scale recommendation systems due to feedback loops and limited signals on user exploration patterns. Large Language Models (LLMs) offer potential solutions by leveraging their world knowledge to recommend novel content outside these loops. A key challenge is aligning LLMs with user preferences while preserving their knowledge and reasoning. To enhance planning for new user interests using LLMs, this paper introduces a novel approach that combines hierarchical planning with LLM inference-time scaling. This method aims to improve recommendation relevancy without compromising novelty. We decouple novelty and user-alignment, training separate LLMs for each objective. We then scale up the novelty-focused LLM’s inference and select the best-of-n predictions using the user-aligned LLM. Live experiments demonstrate efficacy, showing significant gains in both user satisfaction (measured by watch activity and active user counts) and exploration diversity.
2024
pdf
bib
abs
Aligning Large Language Models with Recommendation Knowledge
Yuwei Cao
|
Nikhil Mehta
|
Xinyang Yi
|
Raghunandan Hulikal Keshavan
|
Lukasz Heldt
|
Lichan Hong
|
Ed Chi
|
Maheswaran Sathiamoorthy
Findings of the Association for Computational Linguistics: NAACL 2024
Large language models (LLMs) have recently been used as backbones for recommender systems. However, their performance often lags behind conventional methods in standard tasks like retrieval. We attribute this to a mismatch between LLMs’ knowledge and the knowledge crucial for effective recommendations. While LLMs excel at natural language reasoning, they cannot model complex user-item interactions inherent in recommendation tasks. We propose bridging the knowledge gap and equipping LLMs with recommendation-specific knowledge to address this. Operations such as Masked Item Modeling (MIM) and Bayesian Personalized Ranking (BPR) have found success in conventional recommender systems. Inspired by this, we simulate these operations through natural language to generate auxiliary-task data samples that encode item correlations and user preferences. Fine-tuning LLMs on such auxiliary-task data samples and incorporating more informative recommendation-task data samples facilitates the injection of recommendation-specific knowledge into LLMs. Extensive experiments across retrieval, ranking, and rating prediction tasks on LLMs such as FLAN-T5-Base and FLAN-T5-XL show the effectiveness of our technique in domains such as Amazon Toys & Games, Beauty, and Sports & Outdoors. Notably, our method outperforms conventional and LLM-based baselines, including the current SOTA, by significant margins in retrieval, showcasing its potential for enhancing recommendation quality.
pdf
bib
Leveraging LLM Reasoning Enhances Personalized Recommender Systems
Alicia Tsai
|
Adam Kraft
|
Long Jin
|
Chenwei Cai
|
Anahita Hosseini
|
Taibai Xu
|
Zemin Zhang
|
Lichan Hong
|
Ed H. Chi
|
Xinyang Yi
Findings of the Association for Computational Linguistics: ACL 2024