Improving the Gaussian Mechanism for Differential Privacy: Analytical Calibration and Optimal Denoising
About
The Gaussian mechanism is an essential building block used in multitude of differentially private data analysis algorithms. In this paper we revisit the Gaussian mechanism and show that the original analysis has several important limitations. Our analysis reveals that the variance formula for the original mechanism is far from tight in the high privacy regime ($\varepsilon \to 0$) and it cannot be extended to the low privacy regime ($\varepsilon \to \infty$). We address these limitations by developing an optimal Gaussian mechanism whose variance is calibrated directly using the Gaussian cumulative density function instead of a tail bound approximation. We also propose to equip the Gaussian mechanism with a post-processing step based on adaptive estimation techniques by leveraging that the distribution of the perturbation is known. Our experiments show that analytical calibration removes at least a third of the variance of the noise compared to the classical Gaussian mechanism, and that denoising dramatically improves the accuracy of the Gaussian mechanism in the high-dimensional regime.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Regression | Communities and Crime 1990 US Census / 1990 US LEMAS / 1995 FBI UCR (test (20%)) | MSE (Mean)0.0182 | 78 | |
| Regression | California Housing Standard (test) | MSE0.5922 | 78 | |
| Regression | Criteo Sponsored Search Conversion Log (test) | MSE3.12e+3 | 78 | |
| Training Process | Criteo Sponsored Search Conversion Log (train) | Training Time3.4281 | 5 | |
| Training Process | California Housing (train) | Training Time0.291 | 5 | |
| Training Process | Communities and Crime (train) | Training Time (s)0.286 | 5 |