Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Backdoors in Neural Models of Source Code

About

Deep neural networks are vulnerable to a range of adversaries. A particularly pernicious class of vulnerabilities are backdoors, where model predictions diverge in the presence of subtle triggers in inputs. An attacker can implant a backdoor by poisoning the training data to yield a desired target prediction on triggered inputs. We study backdoors in the context of deep-learning for source code. (1) We define a range of backdoor classes for source-code tasks and show how to poison a dataset to install such backdoors. (2) We adapt and improve recent algorithms from robust statistics for our setting, showing that backdoors leave a spectral signature in the learned representation of source code, thus enabling detection of poisoned data. (3) We conduct a thorough evaluation on different architectures and languages, showing the ease of injecting backdoors and our ability to eliminate them.

Goutham Ramakrishnan, Aws Albarghouthi• 2020

Related benchmarks

TaskDatasetResultRank
Code SummarizationPY150
Recall100
12
Code SummarizationCSN
Recall100
12
Code SummarizationPyT
Recall100
12
Method Name PredictionPY150
Recall99.99
12
Method Name PredictionCSN
Recall100
12
Method Name PredictionPyT
Recall100
12
Showing 6 of 6 rows

Other info

Follow for update