Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ADER: Adaptively Distilled Exemplar Replay Towards Continual Learning for Session-based Recommendation

About

Session-based recommendation has received growing attention recently due to the increasing privacy concern. Despite the recent success of neural session-based recommenders, they are typically developed in an offline manner using a static dataset. However, recommendation requires continual adaptation to take into account new and obsolete items and users, and requires "continual learning" in real-life applications. In this case, the recommender is updated continually and periodically with new data that arrives in each update cycle, and the updated model needs to provide recommendations for user activities before the next model update. A major challenge for continual learning with neural models is catastrophic forgetting, in which a continually trained model forgets user preference patterns it has learned before. To deal with this challenge, we propose a method called Adaptively Distilled Exemplar Replay (ADER) by periodically replaying previous training samples (i.e., exemplars) to the current model with an adaptive distillation loss. Experiments are conducted based on the state-of-the-art SASRec model using two widely used datasets to benchmark ADER with several well-known continual learning techniques. We empirically demonstrate that ADER consistently outperforms other baselines, and it even outperforms the method using all historical data at every update cycle. This result reveals that ADER is a promising solution to mitigate the catastrophic forgetting issue towards building more realistic and scalable session-based recommenders.

Fei Mi, Xiaoyu Lin, Boi Faltings• 2020

Related benchmarks

TaskDatasetResultRank
RecommendationGowalla
Recall@200.7135
100
RecommendationGowalla
Runtime68
12
RecommendationDIGINETICA
MRR@1015.71
7
RecommendationYoochoose
MRR@1035.25
7
RecommendationAmazon Sports
MRR@100.1661
7
Incremental Sequential RecommendationYoochoose
Latency (minutes)178
6
Incremental Sequential RecommendationAMAZON
Latency (min)216
6
Incremental Sequential RecommendationDIGINETICA
Time (minutes)180
6
Incremental Sequential RecommendationAverage (DIGINETICA, YOOCHOOSE, Gowalla, Amazon)
Delta Time-46.9
5
Showing 9 of 9 rows

Other info

Follow for update