Live Session
Teatro Petruzzelli
Paper
16 Oct
 
12:00
CEST
Session 9: Sequential Recommendation 2
Add Session to Calendar 2024-10-16 12:00 pm 2024-10-16 01:10 pm Europe/Rome Session 9: Sequential Recommendation 2 Session 9: Sequential Recommendation 2 is taking place on the RecSys Hub. Https://recsyshub.org
Main Track

Distillation Matters: Empowering Sequential Recommenders to Match the Performance of Large Language Models

View on ACM Digital Library

Yu Cui (Zhejiang University), Feng Liu (OPPO Co Ltd), Pengbo Wang (University of Electronic Science and Technology of China), Bohao Wang (Zhejiang University), Heng Tang (Zhejiang University), Yi Wan (OPPO Co Ltd), Jun Wang (OPPO Co Ltd) and Jiawei Chen (Zhejiang University)

View Paper PDFView Poster
Abstract

Owing to their powerful semantic reasoning capabilities, Large Language Models (LLMs) have been effectively utilized as recommenders, achieving impressive performance. However, the high inference latency of LLMs significantly restricts their practical deployment. To address this issue, this work investigates knowledge distillation from cumbersome LLM-based recommendation models to lightweight conventional sequential models. It encounters three challenges: 1) the teacher’s knowledge may not always be reliable; 2) the capacity gap between the teacher and student makes it difficult for the student to assimilate the teacher’s knowledge; 3) divergence in semantic space poses a challenge to distill the knowledge from embeddings.To tackle these challenges, this work proposes a novel distillation strategy, DLLM2Rec, specifically tailored for knowledge distillation from LLM-based recommendation models to conventional sequential models. DLLM2Rec comprises: 1) Importance-aware ranking distillation, which filters reliable and student-friendly knowledge by weighting instances according to teacher confidence and student-teacher consistency; 2) Collaborative embedding distillation integrates knowledge from teacher embeddings with collaborative signals mined from the data. Extensive experiments demonstrate the effectiveness of the proposed DLLM2Rec, boosting three typical sequential models with an average improvement of 47.97%, even enabling them to surpass LLM-based recommenders in some cases.

Join the Conversation

Head to Slido and select the paper's assigned session to join the live discussion.

Conference Agenda

View Full Agenda →
No items found.