Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Exploring Simple Siamese Representation Learning

About

Siamese networks have become a common structure in various recent models for unsupervised visual representation learning. These models maximize the similarity between two augmentations of one image, subject to certain conditions for avoiding collapsing solutions. In this paper, we report surprising empirical results that simple Siamese networks can learn meaningful representations even using none of the following: (i) negative sample pairs, (ii) large batches, (iii) momentum encoders. Our experiments show that collapsing solutions do exist for the loss and structure, but a stop-gradient operation plays an essential role in preventing collapsing. We provide a hypothesis on the implication of stop-gradient, and further show proof-of-concept experiments verifying it. Our "SimSiam" method achieves competitive results on ImageNet and downstream tasks. We hope this simple baseline will motivate people to rethink the roles of Siamese architectures for unsupervised representation learning. Code will be made available.

Xinlei Chen, Kaiming He• 2020

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR-100 (test)
Accuracy64.7
3518
Image ClassificationCIFAR-10 (test)
Accuracy91.6
3381
Semantic segmentationADE20K (val)
mIoU47.3
2888
Object DetectionCOCO 2017 (val)
AP40.4
2643
Semantic segmentationPASCAL VOC 2012 (val)
Mean IoU77.09
2142
Image ClassificationImageNet-1k (val)
Top-1 Accuracy83
1469
Image ClassificationImageNet (val)
Top-1 Acc71.3
1206
Instance SegmentationCOCO 2017 (val)
APm0.427
1201
Video Object SegmentationDAVIS 2017 (val)
J mean64.8
1193
ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy (%)70
1163
Showing 10 of 234 rows
...

Other info

Follow for update