Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training

About

Tabular data underpins numerous high-impact applications of machine learning from fraud detection to genomics and healthcare. Classical approaches to solving tabular problems, such as gradient boosting and random forests, are widely used by practitioners. However, recent deep learning methods have achieved a degree of performance competitive with popular techniques. We devise a hybrid deep learning approach to solving tabular data problems. Our method, SAINT, performs attention over both rows and columns, and it includes an enhanced embedding method. We also study a new contrastive self-supervised pre-training method for use when labels are scarce. SAINT consistently improves performance over previous deep learning methods, and it even outperforms gradient boosting methods, including XGBoost, CatBoost, and LightGBM, on average over a variety of benchmark tasks.

Gowthami Somepalli, Micah Goldblum, Avi Schwarzschild, C. Bayan Bruss, Tom Goldstein• 2021

Related benchmarks

TaskDatasetResultRank
ClassificationCAD 1% labels
AUC78.02
27
ClassificationInfarction 1% labels
AUC75.63
27
ClassificationInfarction 10% labels
AUC0.7993
27
ClassificationCAD 10% labels
AUC83.37
27
ClassificationDVM 10% labels
Accuracy83.36
27
ClassificationDVM 1% labels
Accuracy27.98
27
Tabular ClassificationTabZilla avg across 98 datasets
Mean Accuracy84
20
Showing 7 of 7 rows

Other info

Follow for update