Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Watermarking Graph Neural Networks based on Backdoor Attacks

About

Graph Neural Networks (GNNs) have achieved promising performance in various real-world applications. Building a powerful GNN model is not a trivial task, as it requires a large amount of training data, powerful computing resources, and human expertise in fine-tuning the model. Moreover, with the development of adversarial attacks, e.g., model stealing attacks, GNNs raise challenges to model authentication. To avoid copyright infringement on GNNs, verifying the ownership of the GNN models is necessary. This paper presents a watermarking framework for GNNs for both graph and node classification tasks. We 1) design two strategies to generate watermarked data for the graph classification task and one for the node classification task, 2) embed the watermark into the host model through training to obtain the watermarked GNN model, and 3) verify the ownership of the suspicious model in a black-box setting. The experiments show that our framework can verify the ownership of GNN models with a very high probability (up to $99\%$) for both tasks. Finally, we experimentally show that our watermarking approach is robust against a state-of-the-art model extraction technique and four state-of-the-art defenses against backdoor attacks.

Jing Xu, Stefanos Koffas, Oguzhan Ersoy, Stjepan Picek• 2021

Related benchmarks

TaskDatasetResultRank
Graph ClassificationPROTEINS
Accuracy73.99
742
Graph ClassificationENZYMES
Accuracy43.61
305
Node ClassificationPhoto--
165
Node ClassificationComputers--
143
Node ClassificationCora
F1 Score80.64
48
Node ClassificationCiteseer
F1 Score67.31
39
Node ClassificationPhoto
AUC98.44
38
Node ClassificationComputers
AUC97.63
38
Node ClassificationPhysics
Overall F190.33
34
Node ClassificationCS
Overall F188.83
34
Showing 10 of 30 rows

Other info

Follow for update