Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Causal Attention for Interpretable and Generalizable Graph Classification

About

In graph classification, attention and pooling-based graph neural networks (GNNs) prevail to extract the critical features from the input graph and support the prediction. They mostly follow the paradigm of learning to attend, which maximizes the mutual information between the attended graph and the ground-truth label. However, this paradigm makes GNN classifiers recklessly absorb all the statistical correlations between input features and labels in the training data, without distinguishing the causal and noncausal effects of features. Instead of underscoring the causal features, the attended graphs are prone to visit the noncausal features as the shortcut to predictions. Such shortcut features might easily change outside the training distribution, thereby making the GNN classifiers suffer from poor generalization. In this work, we take a causal look at the GNN modeling for graph classification. With our causal assumption, the shortcut feature serves as a confounder between the causal feature and prediction. It tricks the classifier to learn spurious correlations that facilitate the prediction in in-distribution (ID) test evaluation, while causing the performance drop in out-of-distribution (OOD) test data. To endow the classifier with better interpretation and generalization, we propose the Causal Attention Learning (CAL) strategy, which discovers the causal patterns and mitigates the confounding effect of shortcuts. Specifically, we employ attention modules to estimate the causal and shortcut features of the input graph. We then parameterize the backdoor adjustment of causal theory -- combine each causal feature with various shortcut features. It encourages the stable relationships between the causal estimation and prediction, regardless of the changes in shortcut parts and distributions. Extensive experiments on synthetic and real-world datasets demonstrate the effectiveness of CAL.

Yongduo Sui, Xiang Wang, Jiancan Wu, Min Lin, Xiangnan He, Tat-Seng Chua• 2021

Related benchmarks

TaskDatasetResultRank
Graph ClassificationPROTEINS
Accuracy76.92
742
Graph ClassificationMUTAG
Accuracy89.91
697
Graph ClassificationNCI1
Accuracy83.89
460
Graph ClassificationCOLLAB
Accuracy82.68
329
Graph ClassificationIMDB-B
Accuracy74.13
322
Graph ClassificationIMDB-M
Accuracy52.6
218
Graph ClassificationMolHIV
ROC AUC57.95
82
Graph ClassificationTwitter
Accuracy55.36
57
Graph ClassificationDrugOOD EC50 (OOD test)
ROC AUC76.45
52
Graph ClassificationDrugOOD EC50 (Scaffold-based OOD shift)
ROC AUC66.1
36
Showing 10 of 68 rows

Other info

Follow for update