What is it about?
A Scalable Way to Use Generative AI in Recommendation Systems by Applying Large Language Models Only to the Most Difficult User Cases, Reducing Computational Costs While Preserving Accuracy and Enabling the System to Handle Growth Without Slowing Down or Becoming Too Expensive
Featured Image
Photo by Denise Jans on Unsplash
Why is it important?
What makes this work timely is that platforms increasingly want to use generative AI to enhance recommendations, but full-scale adoption remains impractical. Our approach offers a selective, data-driven method to decide exactly when a large language model should be used and when traditional methods are enough. This provides a unique, responsible way to improve recommendation systems at scale, reducing costs while ensuring better performance for hard-to-serve users.
Perspectives
This research emerged from an honest question: How can we harness generative AI without overwhelming our systems, budgets, or users? LLMs bring remarkable capabilities, but using them indiscriminately is neither scalable nor responsible. In this paper, I tried to bridge that tension by proposing a framework that uses these models only where they matter most. My hope is that this perspective helps push the field toward approaches that value both innovation and practicality, ensuring that advanced AI benefits the users who need it without creating unnecessary costs.
Kirandeep Kaur
University of Washington
Read the Original
This page is a summary of: Efficient and Responsible Adaptation of Large Language Models for Robust Top-k Recommendations, ACM Transactions on Recommender Systems, November 2025, ACM (Association for Computing Machinery),
DOI: 10.1145/3774778.
You can read the full text:
Resources
Contributors
The following have contributed to this page







