Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Catalog-Native LLM: Speaking Item-ID Dialect with Less Entanglement for Recommendation

About

While collaborative filtering delivers predictive accuracy and efficiency, and Large Language Models (LLMs) enable expressive and generalizable reasoning, modern recommendation systems must bring these strengths together. Growing user expectations, such as natural-language queries and transparent explanations, further highlight the need for a unified approach. However, doing so is nontrivial. Collaborative signals are often token-efficient but semantically opaque, while LLMs are semantically rich but struggle to model implicit user preferences when trained only on textual inputs. This paper introduces Item-ID + Oral-language Mixture-of-Experts Language Model (IDIOMoE), which treats item interaction histories as a native dialect within the language space, enabling collaborative signals to be understood in the same way as natural language. By splitting the Feed Forward Network of each block of a pretrained LLM into a separate text expert and an item expert with token-type gating, our method avoids destructive interference between text and catalog modalities. IDIOMoE demonstrates strong recommendation performance across both public and proprietary datasets, while preserving the text understanding of the pretrained model.

Reza Shirkavand, Xiaokai Wei, Chen Wang, Zheng Hui, Heng Huang, Michelle Gong• 2025

Related benchmarks

TaskDatasetResultRank
Sequential RecommendationAmazon Beauty
NDCG@106.65
84
Sequential RecommendationAmazon Toys
R@100.0927
51
Sequential RecommendationAmazon Sports
HR@106.74
22
Sequential RecommendationAmazon Instruments
NDCG@100.1054
16
Sequential RecommendationAmazon Games
NDCG@106.05
16
Sequential RecommendationAmazon Books
NDCG@102.24
13
Sequential RecommendationIndustrial Dataset
NDCG@1027.1
6
Showing 7 of 7 rows

Other info

Follow for update