Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improving Self-Supervised Learning by Characterizing Idealized Representations

About

Despite the empirical successes of self-supervised learning (SSL) methods, it is unclear what characteristics of their representations lead to high downstream accuracies. In this work, we characterize properties that SSL representations should ideally satisfy. Specifically, we prove necessary and sufficient conditions such that for any task invariant to given data augmentations, desired probes (e.g., linear or MLP) trained on that representation attain perfect accuracy. These requirements lead to a unifying conceptual framework for improving existing SSL methods and deriving new ones. For contrastive learning, our framework prescribes simple but significant improvements to previous methods such as using asymmetric projection heads. For non-contrastive learning, we use our framework to derive a simple and novel objective. Our resulting SSL algorithms outperform baselines on standard benchmarks, including SwAV+multicrops on linear probing of ImageNet.

Yann Dubois, Tatsunori Hashimoto, Stefano Ermon, Percy Liang• 2022

Related benchmarks

TaskDatasetResultRank
Image ClassificationFood-101
Accuracy77.9
494
Image ClassificationDTD
Accuracy73.9
487
Image ClassificationStanford Cars
Accuracy62.2
477
Image ClassificationCIFAR100
Accuracy77.6
331
Image ClassificationOxford-IIIT Pets
Accuracy88
259
Image ClassificationCIFAR10
Accuracy93.6
240
Image ClassificationCaltech-101
Accuracy91.5
198
Image ClassificationFGVC Aircraft--
185
Image ClassificationOxford Flowers 102
Accuracy95.3
172
Image ClassificationImageNet (val)
Top-1 Accuracy68.9
118
Showing 10 of 10 rows

Other info

Code

Follow for update