Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

On the Efficiency of Sequentially Aware Recommender Systems: Cotten4Rec

About

Sequential recommendation (SR) models predict a user's next interaction by modeling their historical behaviors. Transformer-based SR methods, notably BERT4Rec, effectively capture these patterns but incur significant computational overhead due to extensive intermediate computations associated with Softmax-based attention. We propose Cotten4Rec, a novel SR model utilizing linear-time cosine similarity attention, implemented through a single optimized compute unified device architecture (CUDA) kernel. By minimizing intermediate buffers and kernel-launch overhead, Cotten4Rec substantially reduces resource usage compared to BERT4Rec and the linear-attention baseline, LinRec, especially for datasets with moderate sequence lengths and vocabulary sizes. Evaluations across three benchmark datasets confirm that Cotten4Rec achieves considerable reductions in memory and runtime with minimal compromise in recommendation accuracy, demonstrating Cotten4Rec's viability as an efficient alternative for practical, large-scale sequential recommendation scenarios where computational resources are critical.

Shankar Veludandi, Gulrukh Kurdistan, Uzma Mushtaque• 2026

Related benchmarks

TaskDatasetResultRank
Sequential RecommendationML 1M
NDCG@100.38
49
Sequential RecommendationBeauty
Memory (MB)4.09e+3
18
Sequential RecommendationML-20M
Memory (MB)2.32e+3
18
Showing 3 of 3 rows

Other info

Follow for update