Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Contrastive Clustering

About

In this paper, we propose a one-stage online clustering method called Contrastive Clustering (CC) which explicitly performs the instance- and cluster-level contrastive learning. To be specific, for a given dataset, the positive and negative instance pairs are constructed through data augmentations and then projected into a feature space. Therein, the instance- and cluster-level contrastive learning are respectively conducted in the row and column space by maximizing the similarities of positive pairs while minimizing those of negative ones. Our key observation is that the rows of the feature matrix could be regarded as soft labels of instances, and accordingly the columns could be further regarded as cluster representations. By simultaneously optimizing the instance- and cluster-level contrastive loss, the model jointly learns representations and cluster assignments in an end-to-end manner. Extensive experimental results show that CC remarkably outperforms 17 competitive clustering methods on six challenging image benchmarks. In particular, CC achieves an NMI of 0.705 (0.431) on the CIFAR-10 (CIFAR-100) dataset, which is an up to 19\% (39\%) performance improvement compared with the best baseline.

Yunfan Li, Peng Hu, Zitao Liu, Dezhong Peng, Joey Tianyi Zhou, Xi Peng• 2020

Related benchmarks

TaskDatasetResultRank
Image ClusteringCIFAR-10
NMI0.705
243
Image ClusteringSTL-10
ACC85
229
ClusteringCIFAR-10 (test)
Accuracy79
184
Image ClusteringImageNet-10
NMI0.859
166
ClusteringSTL-10 (test)
Accuracy85
146
ClusteringCIFAR-100 (test)
ACC43
110
Image ClusteringCIFAR-100
ACC42.9
101
ClusteringFashion MNIST
NMI67.5
95
ClusteringCIFAR100 20
ACC42.9
93
Image ClusteringImagenet dog-15
NMI44.5
90
Showing 10 of 56 rows

Other info

Follow for update