TopicNet: Semantic Graph-Guided Topic Discovery
- URL: http://arxiv.org/abs/2110.14286v1
- Date: Wed, 27 Oct 2021 09:07:14 GMT
- Title: TopicNet: Semantic Graph-Guided Topic Discovery
- Authors: Zhibin Duan, Yishi Xu, Bo Chen, Dongsheng Wang, Chaojie Wang, Mingyuan
Zhou
- Abstract summary: Existing deep hierarchical topic models are able to extract semantically meaningful topics from a text corpus in an unsupervised manner.
We introduce TopicNet as a deep hierarchical topic model that can inject prior structural knowledge as an inductive bias to influence learning.
- Score: 51.71374479354178
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing deep hierarchical topic models are able to extract semantically
meaningful topics from a text corpus in an unsupervised manner and
automatically organize them into a topic hierarchy. However, it is unclear how
to incorporate prior beliefs such as knowledge graph to guide the learning of
the topic hierarchy. To address this issue, we introduce TopicNet as a deep
hierarchical topic model that can inject prior structural knowledge as an
inductive bias to influence learning. TopicNet represents each topic as a
Gaussian-distributed embedding vector, projects the topics of all layers into a
shared embedding space, and explores both the symmetric and asymmetric
similarities between Gaussian embedding vectors to incorporate prior semantic
hierarchies. With an auto-encoding variational inference network, the model
parameters are optimized by minimizing the evidence lower bound and a
regularization term via stochastic gradient descent. Experiments on widely used
benchmarks show that TopicNet outperforms related deep topic models on
discovering deeper interpretable topics and mining better
document~representations.
Related papers
- Self-supervised Topic Taxonomy Discovery in the Box Embedding Space [23.942807248774514]
This paper develops a Box embedding-based Topic Model (BoxTM) that maps words and topics into the box embedding space.
Our BoxTM explicitly infers upper-level topics based on correlation between specific topics.
Extensive experiments validate high-quality of the topic taxonomy learned by BoxTM.
arXiv Detail & Related papers (2024-08-27T13:19:32Z) - KMF: Knowledge-Aware Multi-Faceted Representation Learning for Zero-Shot
Node Classification [75.95647590619929]
Zero-Shot Node Classification (ZNC) has been an emerging and crucial task in graph data analysis.
We propose a Knowledge-Aware Multi-Faceted framework (KMF) that enhances the richness of label semantics.
A novel geometric constraint is developed to alleviate the problem of prototype drift caused by node information aggregation.
arXiv Detail & Related papers (2023-08-15T02:38:08Z) - HyperMiner: Topic Taxonomy Mining with Hyperbolic Embedding [54.52651110749165]
We present a novel framework that introduces hyperbolic embeddings to represent words and topics.
With the tree-likeness property of hyperbolic space, the underlying semantic hierarchy can be better exploited to mine more interpretable topics.
arXiv Detail & Related papers (2022-10-16T02:54:17Z) - Knowledge-Aware Bayesian Deep Topic Model [50.58975785318575]
We propose a Bayesian generative model for incorporating prior domain knowledge into hierarchical topic modeling.
Our proposed model efficiently integrates the prior knowledge and improves both hierarchical topic discovery and document representation.
arXiv Detail & Related papers (2022-09-20T09:16:05Z) - Sawtooth Factorial Topic Embeddings Guided Gamma Belief Network [49.458250193768826]
We propose sawtooth factorial topic embedding guided GBN, a deep generative model of documents.
Both the words and topics are represented as embedding vectors of the same dimension.
Our models outperform other neural topic models on extracting deeper interpretable topics.
arXiv Detail & Related papers (2021-06-30T10:14:57Z) - Hierarchical Topic Mining via Joint Spherical Tree and Text Embedding [37.7780399311715]
Hierarchical Topic Mining aims to mine a set of representative terms for each category from a text corpus to help a user comprehend his/her interested topics.
Our model, named JoSH, mines a high-quality set of hierarchical topics with high efficiency and benefits weakly-supervised hierarchical text classification tasks.
arXiv Detail & Related papers (2020-07-18T23:30:47Z) - Deep Autoencoding Topic Model with Scalable Hybrid Bayesian Inference [55.35176938713946]
We develop deep autoencoding topic model (DATM) that uses a hierarchy of gamma distributions to construct its multi-stochastic-layer generative network.
We propose a Weibull upward-downward variational encoder that deterministically propagates information upward via a deep neural network, followed by a downward generative model.
The efficacy and scalability of our models are demonstrated on both unsupervised and supervised learning tasks on big corpora.
arXiv Detail & Related papers (2020-06-15T22:22:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.