Deep clustering with concrete k-means
About
We address the problem of simultaneously learning a k-means clustering and deep feature representation from unlabelled data, which is of interest due to the potential of deep k-means to outperform traditional two-step feature extraction and shallow-clustering strategies. We achieve this by developing a gradient-estimator for the non-differentiable k-means objective via the Gumbel-Softmax reparameterisation trick. In contrast to previous attempts at deep clustering, our concrete k-means model can be optimised with respect to the canonical k-means objective and is easily trained end-to-end without resorting to alternating optimisation. We demonstrate the efficacy of our method on standard clustering benchmarks.
Boyan Gao, Yongxin Yang, Henry Gouk, Timothy M. Hospedales• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Clustering | MNIST (test) | NMI0.814 | 122 | |
| Clustering | MNIST | NMI0.814 | 92 | |
| Image Clustering | USPS | NMI0.707 | 43 | |
| Clustering | USPS (test) | ACC72.1 | 19 | |
| Clustering | 20NEWSGROUP (test) | ACC47.3 | 10 |
Showing 5 of 5 rows