Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Topological Detection of Trojaned Neural Networks

About

Deep neural networks are known to have security issues. One particular threat is the Trojan attack. It occurs when the attackers stealthily manipulate the model's behavior through Trojaned training samples, which can later be exploited. Guided by basic neuroscientific principles we discover subtle -- yet critical -- structural deviation characterizing Trojaned models. In our analysis we use topological tools. They allow us to model high-order dependencies in the networks, robustly compare different networks, and localize structural abnormalities. One interesting observation is that Trojaned models develop short-cuts from input to output layers. Inspired by these observations, we devise a strategy for robust detection of Trojaned models. Compared to standard baselines it displays better performance on multiple benchmarks.

Songzhu Zheng, Yikai Zhang, Hubert Wagner, Mayank Goswami, Chao Chen• 2021

Related benchmarks

TaskDatasetResultRank
Trojaned Model DetectionMNIST LeNet5 (test)
Accuracy85
5
Trojaned Model DetectionMNIST Resnet18 (test)
Accuracy87
5
Trojaned Model DetectionCIFAR10 Resnet18 (test)
Accuracy93
5
Trojaned Model DetectionCIFAR10 Densenet121 (test)
Accuracy84
5
Trojan DetectionIARPA/NIST TrojAI ResNet Round 1
Accuracy77
4
Trojan DetectionIARPA/NIST TrojAI DenseNet Round 1
ACC62
4
Showing 6 of 6 rows

Other info

Follow for update