Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dataset Distillation by Matching Training Trajectories

About

Dataset distillation is the task of synthesizing a small dataset such that a model trained on the synthetic set will match the test accuracy of the model trained on the full dataset. In this paper, we propose a new formulation that optimizes our distilled data to guide networks to a similar state as those trained on real data across many training steps. Given a network, we train it for several iterations on our distilled data and optimize the distilled data with respect to the distance between the synthetically trained parameters and the parameters trained on real data. To efficiently obtain the initial and target network parameters for large-scale datasets, we pre-compute and store training trajectories of expert networks trained on the real dataset. Our method handily outperforms existing methods and also allows us to distill higher-resolution visual data.

George Cazenavette, Tongzhou Wang, Antonio Torralba, Alexei A. Efros, Jun-Yan Zhu• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy47.7
3518
Image ClassificationCIFAR-10 (test)
Accuracy71.6
3381
Image ClassificationCIFAR-100 (val)--
661
Image ClassificationCIFAR10 (test)
Accuracy71.6
585
Image ClassificationFashion MNIST (test)
Accuracy90
568
Image ClassificationCIFAR-10
Accuracy71.6
507
Image ClassificationCIFAR100
Accuracy47.7
331
Image ClassificationCIFAR-10 (val)
Top-1 Accuracy71.6
329
Image ClassificationCIFAR10 (test)
Test Accuracy71.6
284
ClassificationCIFAR10 (test)
Accuracy71.6
266
Showing 10 of 69 rows

Other info

Follow for update