Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Inductive Representation Learning on Temporal Graphs

About

Inductive representation learning on temporal graphs is an important step toward salable machine learning on real-world dynamic networks. The evolving nature of temporal dynamic graphs requires handling new nodes as well as capturing temporal patterns. The node embeddings, which are now functions of time, should represent both the static node features and the evolving topological structures. Moreover, node and topological features can be temporal as well, whose patterns the node embeddings should also capture. We propose the temporal graph attention (TGAT) layer to efficiently aggregate temporal-topological neighborhood features as well as to learn the time-feature interactions. For TGAT, we use the self-attention mechanism as building block and develop a novel functional time encoding technique based on the classical Bochner's theorem from harmonic analysis. By stacking TGAT layers, the network recognizes the node embeddings as functions of time and is able to inductively infer embeddings for both new and observed nodes as the graph evolves. The proposed approach handles both node classification and link prediction task, and can be naturally extended to include the temporal edge features. We evaluate our method with transductive and inductive tasks under temporal settings with two benchmark and one industrial dataset. Our TGAT model compares favorably to state-of-the-art baselines as well as the previous temporal graph embedding approaches.

Da Xu, Chuanwei Ruan, Evren Korpeoglu, Sushant Kumar, Kannan Achan• 2020

Related benchmarks

TaskDatasetResultRank
Node ClassificationREDDIT--
66
Link PredictionReddit (inductive)
AP97.9
52
Dynamic Link DetectionENRON
AP71.12
44
Dynamic Graph Anomaly DetectionMOOC S2
AUROC64.82
42
Dynamic Graph Anomaly DetectionWikipedia S2
AUROC59.18
42
Dynamic new link predictionSocial Evo.
AP0.9501
37
Link PredictionEnron (inductive)
AP69.3
37
Link PredictionReddit (transductive)
AP98.6
30
Link PredictionEnron (transductive)
AP68.02
28
Link PredictionLastFM (transductive)
AP63.3
28
Showing 10 of 59 rows

Other info

Follow for update