Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Bootstrap your own latent: A new approach to self-supervised Learning

About

We introduce Bootstrap Your Own Latent (BYOL), a new approach to self-supervised image representation learning. BYOL relies on two neural networks, referred to as online and target networks, that interact and learn from each other. From an augmented view of an image, we train the online network to predict the target network representation of the same image under a different augmented view. At the same time, we update the target network with a slow-moving average of the online network. While state-of-the art methods rely on negative pairs, BYOL achieves a new state of the art without them. BYOL reaches $74.3\%$ top-1 classification accuracy on ImageNet using a linear evaluation with a ResNet-50 architecture and $79.6\%$ with a larger ResNet. We show that BYOL performs on par or better than the current state of the art on both transfer and semi-supervised benchmarks. Our implementation and pretrained models are given on GitHub.

Jean-Bastien Grill, Florian Strub, Florent Altch\'e, Corentin Tallec, Pierre H. Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, Bilal Piot, Koray Kavukcuoglu, R\'emi Munos, Michal Valko• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy62.7
3518
Image ClassificationCIFAR-10 (test)
Accuracy92
3381
Semantic segmentationADE20K (val)
mIoU38.8
2888
Object DetectionCOCO 2017 (val)
AP41.1
2643
Semantic segmentationPASCAL VOC 2012 (val)
Mean IoU76.3
2142
Image ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy80.1
1952
Image ClassificationImageNet-1k (val)
Top-1 Accuracy79.6
1469
Semantic segmentationPASCAL VOC 2012 (test)
mIoU76.3
1415
Image ClassificationImageNet (val)
Top-1 Acc79.6
1206
Instance SegmentationCOCO 2017 (val)
APm0.34
1201
Showing 10 of 462 rows
...

Other info

Code

Follow for update