HyMiRec: A Hybrid Multi-interest Learning Framework for LLM-based Sequential Recommendation

📅 2025-10-15
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing LLM-based sequential recommendation methods face two key bottlenecks: (1) truncating long user interaction sequences due to prohibitive inference overhead, thereby losing long-term preferences; and (2) relying on a single prediction embedding, which limits modeling of users’ multifaceted interests. To address these, we propose HyMiRec—a hybrid multi-interest learning framework. It first extracts coarse-grained long-term interests via a lightweight recommendation module, then leverages an LLM to model fine-grained, dynamic preferences. A cosine-similarity-driven residual codebook enables efficient compression and reuse of historical embeddings. Furthermore, a decoupled multi-interest query mechanism adaptively separates and represents heterogeneous interest signals. Extensive experiments on multiple public and industrial datasets demonstrate significant improvements over state-of-the-art methods. Online A/B tests confirm consistent gains in both click-through rate and recommendation diversity.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have recently demonstrated strong potential for sequential recommendation. However, current LLM-based approaches face critical limitations in modeling users'long-term and diverse interests. First, due to inference latency and feature fetching bandwidth constraints, existing methods typically truncate user behavior sequences to include only the most recent interactions, resulting in the loss of valuable long-range preference signals. Second, most current methods rely on next-item prediction with a single predicted embedding, overlooking the multifaceted nature of user interests and limiting recommendation diversity. To address these challenges, we propose HyMiRec, a hybrid multi-interest sequential recommendation framework, which leverages a lightweight recommender to extracts coarse interest embeddings from long user sequences and an LLM-based recommender to captures refined interest embeddings. To alleviate the overhead of fetching features, we introduce a residual codebook based on cosine similarity, enabling efficient compression and reuse of user history embeddings. To model the diverse preferences of users, we design a disentangled multi-interest learning module, which leverages multiple interest queries to learn disentangles multiple interest signals adaptively, allowing the model to capture different facets of user intent. Extensive experiments are conducted on both benchmark datasets and a collected industrial dataset, demonstrating our effectiveness over existing state-of-the-art methods. Furthermore, online A/B testing shows that HyMiRec brings consistent improvements in real-world recommendation systems.
Problem

Research questions and friction points this paper is trying to address.

Addresses truncation of user sequences losing long-term preferences
Solves single-embedding limitation in capturing diverse user interests
Overcomes feature fetching constraints through efficient embedding compression
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hybrid framework combines lightweight and LLM recommenders
Residual codebook compresses user history via cosine similarity
Disentangled multi-interest module adaptively captures diverse preferences
🔎 Similar Papers
No similar papers found.
J
Jingyi Zhou
Xiaohongshu Inc., Beijing, China; Fudan University, Shanghai, China
C
Cheng Chen
Xiaohongshu Inc., Beijing, China
K
Kai Zuo
Xiaohongshu Inc., Beijing, China
Manjie Xu
Manjie Xu
Peking University
Cognitive Reasoning
Z
Zhendong Fu
Xiaohongshu Inc., Beijing, China
Y
Yibo Chen
Xiaohongshu Inc., Beijing, China
X
Xu Tang
Xiaohongshu Inc., Beijing, China
Yao Hu
Yao Hu
浙江大学
Machine Learning