Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Performance-Driven Benchmark for Feature Selection in Tabular Deep Learning

About

Academic tabular benchmarks often contain small sets of curated features. In contrast, data scientists typically collect as many features as possible into their datasets, and even engineer new features from existing ones. To prevent overfitting in subsequent downstream modeling, practitioners commonly use automated feature selection methods that identify a reduced subset of informative features. Existing benchmarks for tabular feature selection consider classical downstream models, toy synthetic datasets, or do not evaluate feature selectors on the basis of downstream performance. Motivated by the increasing popularity of tabular deep learning, we construct a challenging feature selection benchmark evaluated on downstream neural networks including transformers, using real datasets and multiple methods for generating extraneous features. We also propose an input-gradient-based analogue of Lasso for neural networks that outperforms classical feature selection methods on challenging problems such as selecting from corrupted or second-order features.

Valeriia Cherepanova, Roman Levin, Gowthami Somepalli, Jonas Geiping, C. Bayan Bruss, Andrew Gordon Wilson, Tom Goldstein, Micah Goldblum• 2023

Related benchmarks

TaskDatasetResultRank
ClassificationHI
Accuracy0.821
45
ClassificationCO
Accuracy0.969
39
ClassificationHE
Accuracy39.3
38
ClassificationGE
Accuracy57.7
37
Aggregate performance evaluationOverall Performance across 12 Datasets
Rank1.42
29
ClassificationEY
Accuracy72.5
29
RegressionYE
Negative RMSE-0.776
29
ClassificationJA
Accuracy73.6
29
RegressionMI
Negative RMSE-0.895
29
ClassificationAL with second-order extra features
Accuracy96.3
29
Showing 10 of 55 rows

Other info

Code

Follow for update