End-to-end Differentiable Clustering with Associative Memories
About
Clustering is a widely used unsupervised learning technique involving an intensive discrete optimization problem. Associative Memory models or AMs are differentiable neural networks defining a recursive dynamical system, which have been integrated with various deep learning architectures. We uncover a novel connection between the AM dynamics and the inherent discrete assignment necessary in clustering to propose a novel unconstrained continuous relaxation of the discrete clustering problem, enabling end-to-end differentiable clustering with AM, dubbed ClAM. Leveraging the pattern completion ability of AMs, we further develop a novel self-supervised clustering loss. Our evaluations on varied datasets demonstrate that ClAM benefits from the self-supervision, and significantly improves upon both the traditional Lloyd's k-means algorithm, and more recent continuous clustering relaxations (by upto 60% in terms of the Silhouette Coefficient).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Clustering | CIFAR-10 (test) | -- | 184 | |
| Clustering | STL-10 (test) | -- | 146 | |
| Clustering | CIFAR-100 (test) | -- | 110 | |
| Deep Clustering | Fashion MNIST (test) | SC0.279 | 28 | |
| Deep Clustering | CIFAR-100 | SC0.053 | 21 | |
| Deep Clustering | STL-10 | SC0.108 | 19 | |
| Clustering | USPS (test) | -- | 19 | |
| Deep Clustering | CIFAR-10 | SC0.208 | 18 | |
| Clustering | Fashion-MNIST standard (test) | ARI47.2 | 17 | |
| Deep Clustering | 20 Newsgroups (20NG) | SC-0.008 | 16 |