TabPFGen -- Tabular Data Generation with TabPFN
About
Advances in deep generative modelling have not translated well to tabular data. We argue that this is caused by a mismatch in structure between popular generative models and discriminative models of tabular data. We thus devise a technique to turn TabPFN -- a highly performant transformer initially designed for in-context discriminative tabular tasks -- into an energy-based generative model, which we dub TabPFGen. This novel framework leverages the pre-trained TabPFN as part of the energy function and does not require any additional training or hyperparameter tuning, thus inheriting TabPFN's in-context learning capability. We can sample from TabPFGen analogously to other energy-based models. We demonstrate strong results on standard generative modelling tasks, including data augmentation, class-balancing, and imputation, unlocking a new frontier of tabular data generation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Tabular Data Synthesis Fidelity | steel | KS Statistic (Mean)0.64 | 90 | |
| Tabular Data Synthesis Fidelity | biodeg | KS Statistic (Mean)0.56 | 90 | |
| Tabular Data Synthesis Fidelity | fourier | KS Fidelity0.76 | 88 | |
| Tabular Data Synthesis Fidelity | PROTEIN | Mean KS Statistic0.77 | 88 | |
| Tabular Data Synthesis | fourier | Chi-squared Result0.01 | 48 | |
| Tabular Data Synthesis | steel | Chi-squared Test Result0.07 | 47 | |
| Tabular Data Synthesis | biodeg | Chi-Squared Test Result0.07 | 47 | |
| Classification | biodeg | Balanced Accuracy81.36 | 45 | |
| Classification | steel | Balanced Accuracy96.31 | 45 | |
| Tabular Data Synthesis | steel | Inverse KL Divergence0.42 | 45 |