Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Contrastive Multiview Coding

About

Humans view the world through many sensory channels, e.g., the long-wavelength light channel, viewed by the left eye, or the high-frequency vibrations channel, heard by the right ear. Each view is noisy and incomplete, but important factors, such as physics, geometry, and semantics, tend to be shared between all views (e.g., a "dog" can be seen, heard, and felt). We investigate the classic hypothesis that a powerful representation is one that models view-invariant factors. We study this hypothesis under the framework of multiview contrastive learning, where we learn a representation that aims to maximize mutual information between different views of the same scene but is otherwise compact. Our approach scales to any number of views, and is view-agnostic. We analyze key properties of the approach that make it work, finding that the contrastive loss outperforms a popular alternative based on cross-view prediction, and that the more views we learn from, the better the resulting representation captures underlying scene semantics. Our approach achieves state-of-the-art results on image and video unsupervised learning benchmarks. Code is released at: http://github.com/HobbitLong/CMC/.

Yonglong Tian, Dilip Krishnan, Phillip Isola• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet-1k (val)
Top-1 Accuracy70.6
1453
Image ClassificationImageNet (val)
Top-1 Acc70.6
1206
Image ClassificationImageNet-1k (val)
Top-1 Accuracy64.8
840
Image ClassificationImageNet 1k (test)
Top-1 Accuracy68.4
798
Object DetectionCOCO (val)
mAP63.58
613
Image ClassificationImageNet-1K
Top-1 Acc66.2
524
Action RecognitionUCF101 (mean of 3 splits)
Accuracy59.1
357
Image ClassificationImageNet (val)
Top-1 Accuracy68.4
354
Action RecognitionUCF101 (test)
Accuracy59.1
307
Action RecognitionHMDB51 (test)
Accuracy0.267
249
Showing 10 of 94 rows
...

Other info

Code

Follow for update