Generalizing Dataset Distillation via Deep Generative Prior
About
Dataset Distillation aims to distill an entire dataset's knowledge into a few synthetic images. The idea is to synthesize a small number of synthetic data points that, when given to a learning algorithm as training data, result in a model approximating one trained on the original data. Despite recent progress in the field, existing dataset distillation methods fail to generalize to new architectures and scale to high-resolution datasets. To overcome the above issues, we propose to use the learned prior from pre-trained deep generative models to synthesize the distilled data. To achieve this, we present a new optimization algorithm that distills a large number of images into a few intermediate feature vectors in the generative model's latent space. Our method augments existing techniques, significantly improving cross-architecture generalization in all settings.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet-A (val) | Accuracy39.3 | 55 | |
| Image Classification | PathMNIST v1 (test) | Accuracy46.68 | 36 | |
| Image Classification | ImageWoof 256x256 (test) | Accuracy33.8 | 26 | |
| Image Classification | ImageWoof full-sized (test) | Top-1 Accuracy33.8 | 23 | |
| Image Classification | CIFAR-10 | AlexNet Accuracy30.1 | 9 | |
| Image Classification | ImageNette 128x128 (val) | Accuracy31 | 9 | |
| Image Classification | ImageWoof 128x128 (val) | Accuracy17.8 | 9 | |
| Image Classification | ImageFruit 128x128 (val) | Accuracy22.3 | 9 | |
| Image Classification | ImNet-A 128x128 (val) | Accuracy41.8 | 6 | |
| Image Classification | ImNet-B 128x128 (val) | Accuracy42.1 | 6 |