Dynamic Probabilistic Noise Injection for Membership Inference Defense
About
Membership Inference Attacks (MIAs) expose privacy risks by determining whether a specific sample was part of a model's training set. These threats are especially serious in sensitive domains such as healthcare and finance. Traditional mitigation techniques, such as static differential privacy, rely on injecting a fixed amount of noise during training or inference. However, this often leads to a detrimental trade-off: the noise may be insufficient to counter sophisticated attacks or, when increased, can substantially degrade model accuracy. To address this limitation, we propose DynaNoise, an adaptive inference-time defense that modulates injected noise based on per-query sensitivity. DynaNoise estimates risk using measures such as Shannon entropy and scales the noise variance accordingly, followed by a smoothing step that re-normalizes the perturbed outputs to preserve predictive utility. We further introduce MIDPUT (Membership Inference Defense Privacy-Utility Trade-off), a scalar metric that captures both privacy gains and accuracy retention. Our evaluation on several benchmark datasets demonstrates that DynaNoise substantially lowers attack success rates while maintaining competitive accuracy, achieving strong overall MIDPUT scores compared to state-of-the-art defenses.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Membership Inference Attack Defense | ImageNet-10 (test) | Model Score0.7962 | 7 | |
| Membership Inference Attack Defense | CIFAR10 (test) | Model Score0.7807 | 7 | |
| Computational Efficiency Analysis | General Empirical Evaluation | Latency (ms/sample)0.1432 | 7 | |
| Membership Inference Attack Defense | SST-2 (test) | Accuracy86.12 | 7 | |
| Membership Inference Defense Evaluation | SST-2 | MIDPUT_conf0.0139 | 6 | |
| Membership Inference Defense Evaluation | CIFAR10 | MIDPUT Confidence0.1043 | 6 | |
| Membership Inference Defense Evaluation | ImageNet-10 | MIDPUT_conf0.0512 | 6 |