Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Understanding the Limitations of Variational Mutual Information Estimators

About

Variational approaches based on neural networks are showing promise for estimating mutual information (MI) between high dimensional variables. However, they can be difficult to use in practice due to poorly understood bias/variance tradeoffs. We theoretically show that, under some conditions, estimators such as MINE exhibit variance that could grow exponentially with the true amount of underlying MI. We also empirically demonstrate that existing estimators fail to satisfy basic self-consistency properties of MI, such as data processing and additivity under independence. Based on a unified perspective of variational approaches, we develop a new estimator that focuses on variance reduction. Empirical results on standard benchmark tasks demonstrate that our proposed estimator exhibits improved bias-variance trade-offs on standard benchmark tasks.

Jiaming Song, Stefano Ermon• 2019

Related benchmarks

TaskDatasetResultRank
Mutual Information EstimationGaussian d=20 (test)
Bias0.25
60
Mutual Information EstimationGaussian d=5, N=64
Variance0.05
42
Mutual Information EstimationGaussian setting d=20, N=64
Bias0.18
30
Mutual Information EstimationGaussian (d=10, N=64)
Bias0.24
30
Domain AdaptationMNIST to MNIST-M (test)--
24
Learning label-irrelevant representationsCMU-PIE cropped
Latency (sec./max step)0.621
7
Disentangled Representation LearningDsprite
MSE0.65
7
Mutual Information EstimationGaussian (d=5, N=64, MI=6)
Variance0.32
6
Mutual Information EstimationGaussian d=5, N=64, MI=8
Variance1.4
6
Mutual Information EstimationGaussian (d=5, N=64, MI=10)
Variance8.89
6
Showing 10 of 11 rows

Other info

Follow for update