You Never Cluster Alone
About
Recent advances in self-supervised learning with instance-level contrastive objectives facilitate unsupervised clustering. However, a standalone datum is not perceiving the context of the holistic cluster, and may undergo sub-optimal assignment. In this paper, we extend the mainstream contrastive learning paradigm to a cluster-level scheme, where all the data subjected to the same cluster contribute to a unified representation that encodes the context of each data group. Contrastive learning with this representation then rewards the assignment of each datum. To implement this vision, we propose twin-contrast clustering (TCC). We define a set of categorical variables as clustering assignment confidence, which links the instance-level learning track with the cluster-level one. On one hand, with the corresponding assignment variables being the weight, a weighted aggregation along the data points implements the set representation of a cluster. We further propose heuristic cluster augmentation equivalents to enable cluster-level contrastive learning. On the other hand, we derive the evidence lower-bound of the instance-level contrastive objective with the assignments. By reparametrizing the assignment variables, TCC is trained end-to-end, requiring no alternating steps. Extensive experiments show that TCC outperforms the state-of-the-art on challenging benchmarks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Clustering | CIFAR-10 | NMI0.79 | 243 | |
| Image Clustering | STL-10 | ACC81.4 | 229 | |
| Clustering | CIFAR-10 (test) | Accuracy90.6 | 184 | |
| Image Clustering | ImageNet-10 | NMI0.848 | 166 | |
| Clustering | STL-10 (test) | Accuracy81.4 | 146 | |
| Clustering | CIFAR-100 (test) | ACC49.1 | 110 | |
| Clustering | CIFAR100 20 | ACC49.1 | 93 | |
| Clustering | ImageNet-10 (test) | ACC89.7 | 69 | |
| Grouping | Imagenet Dogs | ACC59.5 | 59 | |
| Clustering | Imagenet Dogs | NMI5.54e+3 | 46 |