Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Masked Graph Transformer for Large-Scale Recommendation

About

Graph Transformers have garnered significant attention for learning graph-structured data, thanks to their superb ability to capture long-range dependencies among nodes. However, the quadratic space and time complexity hinders the scalability of Graph Transformers, particularly for large-scale recommendation. Here we propose an efficient Masked Graph Transformer, named MGFormer, capable of capturing all-pair interactions among nodes with a linear complexity. To achieve this, we treat all user/item nodes as independent tokens, enhance them with positional embeddings, and feed them into a kernelized attention module. Additionally, we incorporate learnable relative degree information to appropriately reweigh the attentions. Experimental results show the superior performance of our MGFormer, even with a single attention layer.

Huiyuan Chen, Zhe Xu, Chin-Chia Michael Yeh, Vivian Lai, Yan Zheng, Minghua Xu, Hanghang Tong• 2024

Related benchmarks

TaskDatasetResultRank
RecommendationGowalla (test)
Recall@200.1306
126
RecommendationEpinions (test)
Recall@208.54
33
RecommendationAli-Display (test)
NDCG@200.0649
17
RecommendationFood (test)
NDCG@200.026
9
Showing 4 of 4 rows

Other info

Follow for update