Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Robust Representations via Multi-View Information Bottleneck

About

The information bottleneck principle provides an information-theoretic method for representation learning, by training an encoder to retain all information which is relevant for predicting the label while minimizing the amount of other, excess information in the representation. The original formulation, however, requires labeled data to identify the superfluous information. In this work, we extend this ability to the multi-view unsupervised setting, where two views of the same underlying entity are provided but the label is unknown. This enables us to identify superfluous information as that not shared by both views. A theoretical analysis leads to the definition of a new multi-view model that produces state-of-the-art results on the Sketchy dataset and label-limited versions of the MIR-Flickr dataset. We also extend our theory to the single-view setting by taking advantage of standard data augmentation techniques, empirically showing better generalization capabilities when compared to common unsupervised approaches for representation learning.

Marco Federici, Anjan Dutta, Patrick Forr\'e, Nate Kushman, Zeynep Akata• 2020

Related benchmarks

TaskDatasetResultRank
ClusteringCOIL-20
ACC53.6
47
ClusteringCOIL-100
ACC53.27
28
ClusteringE-MNIST
Accuracy53.16
25
ClusteringHandwritten (test)
ACC63.25
23
ClassificationBDGP
Acc89
17
ClusteringE-FMNIST
ACC (Clustering)53.71
13
ClusteringOffice-31
ACC_clu33.08
13
ClassificationOffice-31 (test)
Accuracy (cls)61.48
11
ClassificationE-FMNIST (test)
Accuracy75.33
11
ClassificationE-MNIST (test)
Accuracy90.81
11
Showing 10 of 22 rows

Other info

Follow for update