Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

How to Achieve the Intended Aim of Deep Clustering Now, without Deep Learning

About

Deep clustering (DC) is often quoted to have a key advantage over $k$-means clustering. Yet, this advantage is often demonstrated using image datasets only, and it is unclear whether it addresses the fundamental limitations of $k$-means clustering. Deep Embedded Clustering (DEC) learns a latent representation via an autoencoder and performs clustering based on a $k$-means-like procedure, while the optimization is conducted in an end-to-end manner. This paper investigates whether the deep-learned representation has enabled DEC to overcome the known fundamental limitations of $k$-means clustering, i.e., its inability to discover clusters of arbitrary shapes, varied sizes and densities. Our investigations on DEC have a wider implication on deep clustering methods in general. Notably, none of these methods exploit the underlying data distribution. We uncover that a non-deep learning approach achieves the intended aim of deep clustering by making use of distributional information of clusters in a dataset to effectively address these fundamental limitations.

Kai Ming Ting, Wei-Jie Xu, Hang Zhang• 2026

Related benchmarks

TaskDatasetResultRank
Image ClusteringCIFAR-10
NMI0.74
243
Image ClusteringSTL-10--
229
Image ClusteringImageNet-10
NMI0.88
166
ClusteringCOIL-20--
47
ClusteringImagenet Dogs
NMI51
46
ClusteringDLPFC
ARI54
30
ClusteringMNIST
NMI82
24
ClusteringMNIST
ARI0.77
19
Clustering2Crescents
ARI1
4
ClusteringDiff-Sizes
ARI97
4
Showing 10 of 28 rows

Other info

Follow for update