Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Dataset Pruning: Reducing Training Data by Examining Generalization Influence

About

The great success of deep learning heavily relies on increasingly larger training data, which comes at a price of huge computational and infrastructural costs. This poses crucial questions that, do all training data contribute to model's performance? How much does each individual training sample or a sub-training-set affect the model's generalization, and how to construct the smallest subset from the entire training data as a proxy training set without significantly sacrificing the model's performance? To answer these, we propose dataset pruning, an optimization-based sample selection method that can (1) examine the influence of removing a particular set of training samples on model's generalization ability with theoretical guarantee, and (2) construct the smallest subset of training data that yields strictly constrained generalization gap. The empirically observed generalization gap of dataset pruning is substantially consistent with our theoretical expectations. Furthermore, the proposed method prunes 40% training examples on the CIFAR-10 dataset, halves the convergence time with only 1.3% test accuracy decrease, which is superior to previous score-based sample selection methods.

Shuo Yang, Zeke Xie, Hanyu Peng, Min Xu, Mingming Sun, Ping Li• 2022

Related benchmarks

TaskDatasetResultRank
Graph ClassificationMUTAG
Accuracy87.9
697
Graph Classificationogbg-molpcba (test)
AP27.7
206
Graph ClassificationOGBG-MOLHIV v1 (test)
ROC-AUC0.769
88
Graph ClassificationDHFR
Accuracy75.6
80
Graph ClassificationOGBG-MOLPCBA v1 (test)
AP26.4
77
Molecular ClassificationHIV
ROC-AUC77.9
35
Molecular Property RegressionQM9 U0 (test)
MAE16.1
24
Molecular Property RegressionQM9 Zpve (test)
MAE1.68
24
Showing 8 of 8 rows

Other info

Follow for update