Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Self-labelling via simultaneous clustering and representation learning

About

Combining clustering and representation learning is one of the most promising approaches for unsupervised learning of deep neural networks. However, doing so naively leads to ill posed learning problems with degenerate solutions. In this paper, we propose a novel and principled learning formulation that addresses these issues. The method is obtained by maximizing the information between labels and input data indices. We show that this criterion extends standard crossentropy minimization to an optimal transport problem, which we solve efficiently for millions of input images and thousands of labels using a fast variant of the Sinkhorn-Knopp algorithm. The resulting method is able to self-label visual data so as to train highly competitive image representations without manual labels. Our method achieves state of the art representation learning performance for AlexNet and ResNet-50 on SVHN, CIFAR-10, CIFAR-100 and ImageNet and yields the first self-supervised AlexNet that outperforms the supervised Pascal VOC detection baseline. Code and models are available.

Yuki Markus Asano, Christian Rupprecht, Andrea Vedaldi• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)--
3518
Semantic segmentationPASCAL VOC 2012 (val)
Mean IoU45.7
2040
Image ClassificationImageNet-1k (val)
Top-1 Accuracy61.5
1453
Image ClassificationImageNet (val)
Top-1 Acc61.5
1206
Image ClassificationCIFAR-10 (test)
Accuracy83.4
906
Object DetectionPASCAL VOC 2007 (test)
mAP59.2
821
Image ClassificationImageNet-1K
Top-1 Acc61.5
524
Image ClassificationSVHN (test)
Accuracy94.5
362
Image ClassificationImageNet (val)
Top-1 Accuracy61.5
354
Image ClassificationGTSRB
Accuracy59
291
Showing 10 of 43 rows

Other info

Code

Follow for update