Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Representation Learning with Contrastive Predictive Coding

About

While supervised learning has enabled great progress in many applications, unsupervised learning has not seen such widespread adoption, and remains an important and challenging endeavor for artificial intelligence. In this work, we propose a universal unsupervised learning approach to extract useful representations from high-dimensional data, which we call Contrastive Predictive Coding. The key insight of our model is to learn such representations by predicting the future in latent space by using powerful autoregressive models. We use a probabilistic contrastive loss which induces the latent space to capture information that is maximally useful to predict future samples. It also makes the model tractable by using negative sampling. While most prior work has focused on evaluating representations for a particular modality, we demonstrate that our approach is able to learn useful representations achieving strong performance on four distinct domains: speech, images, text and reinforcement learning in 3D environments.

Aaron van den Oord, Yazhe Li, Oriol Vinyals• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)--
3518
Image ClassificationCIFAR-10 (test)--
3381
Image ClassificationImageNet-1k (val)
Top-1 Accuracy48.7
1453
Image ClassificationImageNet (val)
Top-1 Acc48.7
1206
Multivariate ForecastingETTh1
MSE0.687
645
Image ClassificationCIFAR10 (test)
Accuracy80.69
585
Image ClassificationImageNet-1K
Top-1 Acc48.7
524
Semantic Textual SimilaritySTS tasks (STS12, STS13, STS14, STS15, STS16, STS-B, SICK-R) various (test)
STS12 Score70.16
393
Image ClassificationSTL-10 (test)
Accuracy87.9
357
Image ClassificationImageNet (val)
Top-1 Accuracy48.7
354
Showing 10 of 122 rows
...

Other info

Follow for update