337

MSL: Not All Tokens Are What You Need for Tuning LLM as a Recommender

Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), 2025
Main:10 Pages
8 Figures
Bibliography:1 Pages
8 Tables
Appendix:1 Pages
Abstract

Large language models (LLMs), known for their comprehension capabilities and extensive knowledge, have been increasingly applied to recommendation systems (RS). Given the fundamental gap between the mechanism of LLMs and the requirement of RS, researchers have focused on fine-tuning LLMs with recommendation-specific data to enhance their performance. Language Modeling Loss (LML), originally designed for language generation tasks, is commonly adopted. However, we identify two critical limitations of LML: 1) it exhibits significant divergence from the recommendation objective; 2) it erroneously treats all fictitious item descriptions as negative samples, introducing misleading training signals.

View on arXiv
Comments on this paper