CARTE: Pretraining and Transfer for Tabular Learning
About
Pretrained deep-learning models are the go-to solution for images or text. However, for tabular data the standard is still to train tree-based models. Indeed, transfer learning on tables hits the challenge of data integration: finding correspondences, correspondences in the entries (entity matching) where different words may denote the same entity, correspondences across columns (schema matching), which may come in different orders, names... We propose a neural architecture that does not need such correspondences. As a result, we can pretrain it on background data that has not been matched. The architecture -- CARTE for Context Aware Representation of Table Entries -- uses a graph representation of tabular (or relational) data to process tables with different columns, string embedding of entries and columns names to model an open vocabulary, and a graph-attentional network to contextualize entries with column names and neighboring entries. An extensive benchmark shows that CARTE facilitates learning, outperforming a solid set of baselines including the best tree-based models. CARTE also enables joint learning across tables with unmatched columns, enhancing a small table with bigger ones. CARTE opens the door to large pretrained models for tabular data.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Tuple Matching | Webtable | Recall@1082.5 | 10 | |
| Column matching | Webtable | Recall@1085 | 10 | |
| Column matching | CIUS | Recall@1082.5 | 10 | |
| Tuple Matching | CancerKG | Recall@1075.4 | 10 | |
| Tuple Matching | CIUS | Recall@1082 | 10 | |
| Tuple Matching | SAUS | Recall@1080 | 10 | |
| Column matching | CovidKG | Recall@1073 | 10 | |
| Column matching | SAUS | Recall@1081.7 | 10 | |
| Column matching | CancerKG | Recall@1075 | 10 | |
| Tuple Matching | CovidKG | Recall@1070 | 10 |