Dataset Distillation by Automatic Training Trajectories
About
Dataset Distillation is used to create a concise, yet informative, synthetic dataset that can replace the original dataset for training purposes. Some leading methods in this domain prioritize long-range matching, involving the unrolling of training trajectories with a fixed number of steps (NS) on the synthetic dataset to align with various expert training trajectories. However, traditional long-range matching methods possess an overfitting-like problem, the fixed step size NS forces synthetic dataset to distortedly conform seen expert training trajectories, resulting in a loss of generality-especially to those from unencountered architecture. We refer to this as the Accumulated Mismatching Problem (AMP), and propose a new approach, Automatic Training Trajectories (ATT), which dynamically and adaptively adjusts trajectory length NS to address the AMP. Our method outperforms existing methods particularly in tests involving cross-architectures. Moreover, owing to its adaptive nature, it exhibits enhanced stability in the face of parameter variations.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 (test) | Accuracy51.2 | 3518 | |
| Image Classification | CIFAR-10 (test) | Accuracy74.5 | 3381 | |
| Classification | CIFAR10 (test) | Accuracy74.5 | 266 | |
| Classification | CIFAR-100 (test) | Accuracy51.2 | 129 | |
| Medical Image Classification | Covid (test) | Accuracy87.62 | 43 | |
| Image Classification | PathMNIST v2 (test) | Accuracy88.41 | 35 | |
| Image Classification | Tiny ImageNet 64x64 (test) | Accuracy25.8 | 27 |