Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

S2WTM: Spherical Sliced-Wasserstein Autoencoder for Topic Modeling

About

Modeling latent representations in a hyperspherical space has proven effective for capturing directional similarities in high-dimensional text data, benefiting topic modeling. Variational autoencoder-based neural topic models (VAE-NTMs) commonly adopt the von Mises-Fisher prior to encode hyperspherical structure. However, VAE-NTMs often suffer from posterior collapse, where the KL divergence term in the objective function highly diminishes, leading to ineffective latent representations. To mitigate this issue while modeling hyperspherical structure in the latent space, we propose the Spherical Sliced Wasserstein Autoencoder for Topic Modeling (S2WTM). S2WTM employs a prior distribution supported on the unit hypersphere and leverages the Spherical Sliced-Wasserstein distance to align the aggregated posterior distribution with the prior. Experimental results demonstrate that S2WTM outperforms state-of-the-art topic models, generating more coherent and diverse topics while improving performance on downstream tasks.

Suman Adhya, Debarshi Kumar Sanyal• 2025

Related benchmarks

TaskDatasetResultRank
Topic Modeling20NG
NPMI0.167
23
Topic ModelingBBC
NPMI0.252
17
Document Clustering20NG (test)
NMI0.437
13
Document ClusteringBBC (test)
NMI0.729
13
Document ClusteringM10 (test)
NMI0.464
13
Document ClusteringSS (test)
NMI0.547
13
Document ClusteringPascal (test)
NMI0.471
13
Document ClusteringBio (test)
NMI0.557
13
Document ClusteringDBLP (test)
NMI0.254
13
Topic ModelingM10
NPMI0.101
13
Showing 10 of 21 rows

Other info

Code

Follow for update