Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dataset Condensation with Distribution Matching

About

Computational cost of training state-of-the-art deep models in many learning problems is rapidly increasing due to more sophisticated models and larger datasets. A recent promising direction for reducing training cost is dataset condensation that aims to replace the original large training set with a significantly smaller learned synthetic set while preserving the original information. While training deep models on the small set of condensed images can be extremely fast, their synthesis remains computationally expensive due to the complex bi-level optimization and second-order derivative computation. In this work, we propose a simple yet effective method that synthesizes condensed images by matching feature distributions of the synthetic and original training images in many sampled embedding spaces. Our method significantly reduces the synthesis cost while achieving comparable or better performance. Thanks to its efficiency, we apply our method to more realistic and larger datasets with sophisticated neural architectures and obtain a significant performance boost. We also show promising practical benefits of our method in continual learning and neural architecture search.

Bo Zhao, Hakan Bilen• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy43.6
3518
Image ClassificationCIFAR-10 (test)
Accuracy63
3381
Image ClassificationMNIST (test)
Accuracy98.6
882
Image ClassificationCIFAR-100 (val)--
661
Image ClassificationCIFAR10 (test)
Accuracy63
585
Image ClassificationCIFAR-10
Accuracy63
507
Image ClassificationCIFAR100
Accuracy43.6
331
Image ClassificationCIFAR-10 (val)
Top-1 Accuracy63
329
Image ClassificationCIFAR-100
Accuracy12.4
302
Image ClassificationCIFAR10 (test)
Test Accuracy63.7
284
Showing 10 of 50 rows

Other info

Follow for update