Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Generalizing Dataset Distillation via Deep Generative Prior

About

Dataset Distillation aims to distill an entire dataset's knowledge into a few synthetic images. The idea is to synthesize a small number of synthetic data points that, when given to a learning algorithm as training data, result in a model approximating one trained on the original data. Despite recent progress in the field, existing dataset distillation methods fail to generalize to new architectures and scale to high-resolution datasets. To overcome the above issues, we propose to use the learned prior from pre-trained deep generative models to synthesize the distilled data. To achieve this, we present a new optimization algorithm that distills a large number of images into a few intermediate feature vectors in the generative model's latent space. Our method augments existing techniques, significantly improving cross-architecture generalization in all settings.

George Cazenavette, Tongzhou Wang, Antonio Torralba, Alexei A. Efros, Jun-Yan Zhu• 2023

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet-A (val)
Accuracy39.3
55
Image ClassificationPathMNIST v1 (test)
Accuracy46.68
36
Image ClassificationImageWoof 256x256 (test)
Accuracy33.8
26
Image ClassificationImageWoof full-sized (test)
Top-1 Accuracy33.8
23
Image ClassificationCIFAR-10
AlexNet Accuracy30.1
9
Image ClassificationImageNette 128x128 (val)
Accuracy31
9
Image ClassificationImageWoof 128x128 (val)
Accuracy17.8
9
Image ClassificationImageFruit 128x128 (val)
Accuracy22.3
9
Image ClassificationImNet-A 128x128 (val)
Accuracy41.8
6
Image ClassificationImNet-B 128x128 (val)
Accuracy42.1
6
Showing 10 of 29 rows

Other info

Code

Follow for update