FRAUD-RLA: A new reinforcement learning adversarial attack against credit card fraud detection
About
Adversarial attacks pose a significant threat to data-driven systems, and researchers have spent considerable resources studying them. Despite its economic relevance, this trend largely overlooked the issue of credit card fraud detection. To address this gap, we propose a new threat model that demonstrates the limitations of existing attacks and highlights the necessity to investigate new approaches. We then design a new adversarial attack for credit card fraud detection, employing reinforcement learning to bypass classifiers. This attack, called FRAUD-RLA, is designed to maximize the attacker's reward by optimizing the exploration-exploitation tradeoff and working with significantly less required knowledge than competitors. Our experiments, conducted on three different heterogeneous datasets and against two fraud detection systems, indicate that FRAUD-RLA is effective, even considering the severe limitations imposed by our threat model.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Adversarial Attack | Kaggle | Average Cumulative Reward0.96 | 32 | |
| Adversarial Attack | Kaggle Credit Card Fraud | Average Cumulative Reward91 | 32 | |
| Fraud Detection | Generator 1000 frauds horizon Neural Network | Average Cumulative Reward0.93 | 18 | |
| Fraud Detection | Generator 4000 frauds horizon Neural Network | Average Cumulative Reward0.97 | 18 | |
| Fraud Detection | Generator Neural Network 300 frauds horizon | Average Cumulative Reward80 | 18 | |
| Fraud Detection | Dataset-1 | -- | 16 | |
| Fraud Detection | Generator Random Forest | Average Cumulative Reward (300 Frauds)0.33 | 9 | |
| Fraud Detection | Dataset 2 | -- | 8 | |
| Fraud Detection | Dataset 3 | -- | 8 | |
| Fraud Detection | Dataset 2 | -- | 8 |