Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Differentially Private Neural Tangent Kernels for Privacy-Preserving Data Generation

About

Maximum mean discrepancy (MMD) is a particularly useful distance metric for differentially private data generation: when used with finite-dimensional features it allows us to summarize and privatize the data distribution once, which we can repeatedly use during generator training without further privacy loss. An important question in this framework is, then, what features are useful to distinguish between real and synthetic data distributions, and whether those enable us to generate quality synthetic data. This work considers the using the features of $\textit{neural tangent kernels (NTKs)}$, more precisely $\textit{empirical}$ NTKs (e-NTKs). We find that, perhaps surprisingly, the expressiveness of the untrained e-NTK features is comparable to that of the features taken from pre-trained perceptual features using public data. As a result, our method improves the privacy-accuracy trade-off compared to other state-of-the-art methods, without relying on any public data, as demonstrated on several tabular and image benchmark datasets.

Yilin Yang, Kamil Adamczewski, Danica J. Sutherland, Xiaoxiao Li, Mijung Park• 2023

Related benchmarks

TaskDatasetResultRank
Image GenerationCelebA 64 x 64 (test)
FID299.9
203
Image GenerationCelebA 32x32 (test)
FID227.8
17
Differentially Private Image SynthesisCAMELYON
FID234.5
16
Differentially Private Image SynthesisMNIST
FID69.2
16
Differentially Private Image SynthesisCelebA
FID227.8
16
Differentially Private Image SynthesisF-MNIST
FID120.5
16
Differentially Private Image SynthesisCIFAR-10
FID346.9
16
Image GenerationCelebA 128x128 (test)
FID272.1
14
Showing 8 of 8 rows

Other info

Follow for update