Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Scalable Infomin Learning

About

The task of infomin learning aims to learn a representation with high utility while being uninformative about a specified target, with the latter achieved by minimising the mutual information between the representation and the target. It has broad applications, ranging from training fair prediction models against protected attributes, to unsupervised learning with disentangled representations. Recent works on infomin learning mainly use adversarial training, which involves training a neural network to estimate mutual information or its proxy and thus is slow and difficult to optimise. Drawing on recent advances in slicing techniques, we propose a new infomin learning approach, which uses a novel proxy metric to mutual information. We further derive an accurate and analytically computable approximation to this proxy metric, thereby removing the need of constructing neural network-based mutual information estimators. Experiments on algorithmic fairness, disentangled representation learning and domain adaptation verify that our method can effectively remove unwanted information with limited time budget.

Yanzhi Chen, Weihao Sun, Yingzhen Li, Adrian Weller• 2023

Related benchmarks

TaskDatasetResultRank
Domain AdaptationMNIST to MNIST-M (test)--
24
Disentangled Representation LearningDsprite
MSE0.5
7
Learning label-irrelevant representationsCMU-PIE cropped
Latency (sec./max step)0.581
7
Domain AdaptationCIFAR10 to STL10 (test)
Inference Time (sec/step)3.146
4
Showing 4 of 4 rows

Other info

Code

Follow for update