Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Deep Embedded K-Means Clustering

About

Recently, deep clustering methods have gained momentum because of the high representational power of deep neural networks (DNNs) such as autoencoder. The key idea is that representation learning and clustering can reinforce each other: Good representations lead to good clustering while good clustering provides good supervisory signals to representation learning. Critical questions include: 1) How to optimize representation learning and clustering? 2) Should the reconstruction loss of autoencoder be considered always? In this paper, we propose DEKM (for Deep Embedded K-Means) to answer these two questions. Since the embedding space generated by autoencoder may have no obvious cluster structures, we propose to further transform the embedding space to a new space that reveals the cluster-structure information. This is achieved by an orthonormal transformation matrix, which contains the eigenvectors of the within-class scatter matrix of K-means. The eigenvalues indicate the importance of the eigenvectors' contributions to the cluster-structure information in the new space. Our goal is to increase the cluster-structure information. To this end, we discard the decoder and propose a greedy method to optimize the representation. Representation learning and clustering are alternately optimized by DEKM. Experimental results on the real-world datasets demonstrate that DEKM achieves state-of-the-art performance.

Wengang Guo, Kaiyan Lin, Wei Ye• 2021

Related benchmarks

TaskDatasetResultRank
ClusteringMNIST
NMI0.9106
92
ClusteringUSPS
NMI82.23
82
ClusteringCOIL-20
ACC72.62
47
Deep ClusteringFashion MNIST (test)
SC0.819
28
ClusteringREUTERS 10K
ACC76.42
23
ClusteringFRGC
NMI0.5078
22
Deep ClusteringCIFAR-100
SC0.047
21
Deep ClusteringSTL-10
SC0.804
19
Deep ClusteringCIFAR-10
SC0.622
18
Deep ClusteringUSPS
SC0.843
14
Showing 10 of 15 rows

Other info

Code

Follow for update