Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Dataset Condensation with Gradient Matching

About

As the state-of-the-art machine learning methods in many fields rely on larger datasets, storing datasets and training models on them become significantly more expensive. This paper proposes a training set synthesis technique for data-efficient learning, called Dataset Condensation, that learns to condense large dataset into a small set of informative synthetic samples for training deep neural networks from scratch. We formulate this goal as a gradient matching problem between the gradients of deep neural network weights that are trained on the original and our synthetic data. We rigorously evaluate its performance in several computer vision benchmarks and demonstrate that it significantly outperforms the state-of-the-art methods. Finally we explore the use of our method in continual learning and neural architecture search and report promising gains when limited memory and computations are available.

Bo Zhao, Konda Reddy Mopuri, Hakan Bilen• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy25.2
3518
Image ClassificationCIFAR-10 (test)
Accuracy53.9
3381
Image ClassificationMNIST (test)
Accuracy98.8
894
Image ClassificationCIFAR-100 (val)--
776
Time Series ForecastingETTh1--
729
Image ClassificationCIFAR-100
Accuracy34.8
691
Image ClassificationFashion MNIST (test)
Accuracy83.6
592
Image ClassificationCIFAR10 (test)
Accuracy53.9
585
Image ClassificationCIFAR-10
Accuracy53.9
508
Image ClassificationCIFAR-10
Accuracy53.9
507
Showing 10 of 74 rows
...

Other info

Follow for update