Neural Mutual Information Estimation with Vector Copulas
About
Estimating mutual information (MI) is a fundamental task in data science and machine learning. Existing estimators mainly rely on either highly flexible models (e.g., neural networks), which require large amounts of data, or overly simplified models (e.g., Gaussian copula), which fail to capture complex distributions. Drawing upon recent vector copula theory, we propose a principled interpolation between these two extremes to achieve a better trade-off between complexity and capacity. Experiments on state-of-the-art synthetic benchmarks and real-world data with diverse modalities demonstrate the advantages of the proposed estimator.
Yanzhi Chen, Zijing Ou, Adrian Weller, Michael U. Gutmann• 2025
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mutual Information Estimation | In-Meta-Distribution (IMD) (test) | MSE14 | 36 | |
| Mutual Information Estimation | Out-of-Meta-Distribution (OoMD) (test) | MSE24 | 36 | |
| Mutual Information Estimation | M extended (test) | Time / sample (s)3.93 | 12 |
Showing 3 of 3 rows