Live Session

19 Sep
 
14:00
SGT
Tutorial: On Large Language Models for Recommendation
Add Session to Calendar 2023-09-19 02:00 pm 2023-09-19 03:20 pm UTC+8:00 Tutorial: On Large Language Models for Recommendation Tutorial: On Large Language Models for Recommendation is taking place on the RecSys Hub. Https://recsyshub.org

Tutorial on Large Language Models for Recommendation

View on ACM Digital Library
Speakers

Wenyue Hua (Rutgers University), Lei Li (Hong Kong Baptist University), Shuyuan Xu (Rutgers University), Li Chen (Hong Kong Baptist University), Yongfeng Zhang (Rutgers University)

View PDF
Abstract

Foundation Models such as Large Language Models (LLMs) have significantly advanced many research areas. In particular, LLMs offer significant advantages for recommender systems, making them valuable tools for personalized recommendations. For example, by formulating various recommendation tasks such as rating prediction, sequential recommendation, straightforward recommendation, and explanation generation into language instructions, LLMs make it possible to build universal recommendation engines that can handle different recommendation tasks. Additionally, LLMs have a remarkable capacity for understanding natural language, enabling them to comprehend user preferences, item descriptions, and contextual information to generate more accurate and relevant recommendations, leading to improved user satisfaction and engagement. This tutorial introduces Foundation Models such as LLMs for recommendation. We will introduce how recommender system advanced from shallow models to deep models and to large models, how LLMs enable generative recommendation in contrast to traditional discriminative recommendation, and how to build LLM-based recommender systems. We will cover multiple perspectives of LLM-based recommendation, including data preparation, model design, model pre-training, fine-tuning and prompting, multi-modality and multi-task learning, as well as trustworthy perspectives of LLM-based recommender systems such as fairness and transparency.

Join the Conversation

Tag questions @LiveContent to add to live session Q&A