Search publications

Reset filters Search by keyword

No publications found.

 

Neural topic modeling on hyperspheres: Spherical representation learning with von Mises-Fisher mixtures

Authors: Guo DLuo ZBouguila NFan W


Affiliations

1 Guangdong Provincial/Zhuhai Key Laboratory IRADS and Department of Computer Science, Beijing Normal-Hong Kong Baptist University, Zhuhai, Guangdong, 519087, China. Electronic address: lsdayuguo@bnbu.edu.cn.
2 Concordia Institute for Information Systems Engineering, Concordia University, Montreal, Quebec, H3G 1M8, Canada. Electronic address: zhiwen.luo@mail.concordia.ca.
3 Concordia Institute for Information Systems Engineering, Concordia University, Montreal, Quebec, H3G 1M8, Canada. Electronic address: nizar.bouguila@concordia.ca.
4 Guangdong Provincial/Zhuhai Key Laboratory IRADS and Department of Computer Science, Beijing Normal-Hong Kong Baptist University, Zhuhai, Guangdong, 519087, China. Electronic address: wentaofan@uic.edu.cn.

Description

Neural topic models (NTMs) based on variational autoencoders (VAEs) have emerged as a scalable and flexible alternative to classical probabilistic models for uncovering latent thematic structures in text corpora. However, most existing NTMs either overlook the geometric structure of word embeddings or rely on Euclidean priors that are poorly aligned with L2-normalized semantic spaces. Moreover, VAE-based NTMs often suffer from Kullback-Leibler (KL) divergence collapse and fail to capture directional semantics critical for coherent topic discovery. In this work, we propose the von Mises-Fisher Mixture Neural Variational Topic Model (vMNVTM), a hyperspherical neural topic modeling framework that explicitly models both document representations and topic-word relationship using the von Mises-Fisher (vMF) distribution. Our approach leverages the hyperspherical nature of vMF distribution to capture directional alignment and angular dispersion in the latent space, and enables more coherent and interpretable topic representations in word embedding space. We further introduce a vMF-aware embedding clustering (vEC) loss to align topics with semantically meaningful regions in the word embedding space and a temperature-controlled concentration mechanism to enhance topic separability. Extensive experiments on benchmark datasets demonstrate that vMNVTM outperforms state-of-the-art NTMs in terms of topic coherence, diversity, and interpretability, highlighting the importance of directional modeling in neural topic inference. The source code of our model is publicly accessible at https://github.com/gvbdxgvc/vMNVTM.


Keywords: Hyperspherical word embeddingsNeural topic model (NTM)Variational autoencoder (VAE)von Mises-Fisher distribution


Links

PubMed: https://pubmed.ncbi.nlm.nih.gov/41791177/

DOI: 10.1016/j.neunet.2026.108792