Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RelaxLoss: Defending Membership Inference Attacks without Losing Utility

About

As a long-term threat to the privacy of training data, membership inference attacks (MIAs) emerge ubiquitously in machine learning models. Existing works evidence strong connection between the distinguishability of the training and testing loss distributions and the model's vulnerability to MIAs. Motivated by existing results, we propose a novel training framework based on a relaxed loss with a more achievable learning target, which leads to narrowed generalization gap and reduced privacy leakage. RelaxLoss is applicable to any classification model with added benefits of easy implementation and negligible overhead. Through extensive evaluations on five datasets with diverse modalities (images, medical data, transaction records), our approach consistently outperforms state-of-the-art defense mechanisms in terms of resilience against MIAs as well as model utility. Our defense is the first that can withstand a wide range of attacks while preserving (or even improving) the target model's utility. Source code is available at https://github.com/DingfanChen/RelaxLoss

Dingfan Chen, Ning Yu, Mario Fritz• 2022

Related benchmarks

TaskDatasetResultRank
Membership Inference Attack DefenseCIFAR100 (test)
Loss (Series)0.65
60
Membership Inference Attack DefenseCIFAR10
AUC (Loss-Series)55
26
Membership InferenceTinyImageNet
Loss0.63
23
Membership Inference DefenseTinyImageNet (test)
AUC (Loss-Series)0.67
15
Defense against Membership Inference AttacksCIFAR10
Loss Series Score0.57
15
Membership InferenceCIFAR10 Pair (test)
Loss6.56
8
Membership Inference Attack DefenseCIFAR100 Pair Case
Loss-Series AUC0.63
8
Membership Inference Attack DefenseCIFAR100 Half Case
Loss-Series AUC0.64
8
Computational Efficiency AnalysisGeneral Empirical Evaluation
Latency (ms/sample)0.1414
7
Membership Inference Attack DefenseImageNet-10 (test)
Model Score0.7646
7
Showing 10 of 17 rows

Other info

Follow for update