Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Evaluations of Machine Learning Privacy Defenses are Misleading

About

Empirical defenses for machine learning privacy forgo the provable guarantees of differential privacy in the hope of achieving higher utility while resisting realistic adversaries. We identify severe pitfalls in existing empirical privacy evaluations (based on membership inference attacks) that result in misleading conclusions. In particular, we show that prior evaluations fail to characterize the privacy leakage of the most vulnerable samples, use weak attacks, and avoid comparisons with practical differential privacy baselines. In 5 case studies of empirical privacy defenses, we find that prior evaluations underestimate privacy leakage by an order of magnitude. Under our stronger evaluation, none of the empirical defenses we study are competitive with a properly tuned, high-utility DP-SGD baseline (with vacuous provable guarantees).

Michael Aerni, Jie Zhang, Florian Tram\`er• 2024

Related benchmarks

TaskDatasetResultRank
Membership Inference Attack DefenseCIFAR100 (test)
Loss (Series)0.54
60
Membership Inference Attack DefenseCIFAR10
AUC (Loss-Series)62
26
Membership InferenceTinyImageNet
Loss0.52
23
Defense against Membership Inference AttacksCIFAR10
Loss Series Score0.58
15
Membership Inference DefenseTinyImageNet (test)
AUC (Loss-Series)0.54
15
Membership Inference Attack DefenseCIFAR100 Half Case
Loss-Series AUC0.61
8
Membership Inference Attack DefenseCIFAR100 Pair Case
Loss-Series AUC0.58
8
Membership InferenceCIFAR10 Pair (test)
Loss11.82
8
Membership Inference Attack DefenseTinyImageNet (Half)
Loss0.53
7
Membership InferenceCIFAR10 Half (test)
Loss Series13.55
7
Showing 10 of 11 rows

Other info

Follow for update