A Gamma-Poisson Mixture Topic Model for Short Text
- URL: http://arxiv.org/abs/2004.11464v1
- Date: Thu, 23 Apr 2020 21:13:53 GMT
- Title: A Gamma-Poisson Mixture Topic Model for Short Text
- Authors: Jocelyn Mazarura, Alta de Waal and Pieter de Villiers
- Abstract summary: Most topic models are constructed under the assumption that documents follow a multinomial distribution.
For topic modelling, the Poisson distribution describes the number of occurrences of a word in documents of fixed length.
The few Poisson topic models in literature are admixture models, making the assumption that a document is generated from a mixture of topics.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most topic models are constructed under the assumption that documents follow
a multinomial distribution. The Poisson distribution is an alternative
distribution to describe the probability of count data. For topic modelling,
the Poisson distribution describes the number of occurrences of a word in
documents of fixed length. The Poisson distribution has been successfully
applied in text classification, but its application to topic modelling is not
well documented, specifically in the context of a generative probabilistic
model. Furthermore, the few Poisson topic models in literature are admixture
models, making the assumption that a document is generated from a mixture of
topics. In this study, we focus on short text. Many studies have shown that the
simpler assumption of a mixture model fits short text better. With mixture
models, as opposed to admixture models, the generative assumption is that a
document is generated from a single topic. One topic model, which makes this
one-topic-per-document assumption, is the Dirichlet-multinomial mixture model.
The main contributions of this work are a new Gamma-Poisson mixture model, as
well as a collapsed Gibbs sampler for the model. The benefit of the collapsed
Gibbs sampler derivation is that the model is able to automatically select the
number of topics contained in the corpus. The results show that the
Gamma-Poisson mixture model performs better than the Dirichlet-multinomial
mixture model at selecting the number of topics in labelled corpora.
Furthermore, the Gamma-Poisson mixture produces better topic coherence scores
than the Dirichlet-multinomial mixture model, thus making it a viable option
for the challenging task of topic modelling of short text.
Related papers
- BayesBlend: Easy Model Blending using Pseudo-Bayesian Model Averaging, Stacking and Hierarchical Stacking in Python [0.0]
We introduce the BayesBlend Python package to estimate weights and blend multiple (Bayesian) models' predictive distributions.
BayesBlend implements pseudo-Bayesian model averaging, stacking and, uniquely, hierarchical Bayesian stacking to estimate model weights.
We demonstrate the usage of BayesBlend with examples of insurance loss modeling.
arXiv Detail & Related papers (2024-04-30T19:15:33Z) - Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot
Text Classification Tasks [75.42002070547267]
We propose a self evolution learning (SE) based mixup approach for data augmentation in text classification.
We introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up.
arXiv Detail & Related papers (2023-05-22T23:43:23Z) - Topic Analysis for Text with Side Data [18.939336393665553]
We introduce a hybrid generative probabilistic model that combines a neural network with a latent topic model.
In the model, each document is modeled as a finite mixture over an underlying set of topics.
Each topic is modeled as an infinite mixture over an underlying set of topic probabilities.
arXiv Detail & Related papers (2022-03-01T22:06:30Z) - Sampling from Arbitrary Functions via PSD Models [55.41644538483948]
We take a two-step approach by first modeling the probability distribution and then sampling from that model.
We show that these models can approximate a large class of densities concisely using few evaluations, and present a simple algorithm to effectively sample from these models.
arXiv Detail & Related papers (2021-10-20T12:25:22Z) - On the Generative Utility of Cyclic Conditionals [103.1624347008042]
We study whether and how can we model a joint distribution $p(x,z)$ using two conditional models $p(x|z)$ that form a cycle.
We propose the CyGen framework for cyclic-conditional generative modeling, including methods to enforce compatibility and use the determined distribution to fit and generate data.
arXiv Detail & Related papers (2021-06-30T10:23:45Z) - Argmax Flows and Multinomial Diffusion: Towards Non-Autoregressive
Language Models [76.22217735434661]
This paper introduces two new classes of generative models for categorical data: Argmax Flows and Multinomial Diffusion.
We demonstrate that our models perform competitively on language modelling and modelling of image segmentation maps.
arXiv Detail & Related papers (2021-02-10T11:04:17Z) - Variational Mixture of Normalizing Flows [0.0]
Deep generative models, such as generative adversarial networks autociteGAN, variational autoencoders autocitevaepaper, and their variants, have seen wide adoption for the task of modelling complex data distributions.
Normalizing flows have overcome this limitation by leveraging the change-of-suchs formula for probability density functions.
The present work overcomes this by using normalizing flows as components in a mixture model and devising an end-to-end training procedure for such a model.
arXiv Detail & Related papers (2020-09-01T17:20:08Z) - Probability Link Models with Symmetric Information Divergence [1.5749416770494706]
Two general classes of link models are proposed.
The first model links two survival functions and is applicable to models such as the proportional odds and change point.
The second model links two cumulative probability distribution functions.
arXiv Detail & Related papers (2020-08-10T19:49:51Z) - Contextuality scenarios arising from networks of stochastic processes [68.8204255655161]
An empirical model is said contextual if its distributions cannot be obtained marginalizing a joint distribution over X.
We present a different and classical source of contextual empirical models: the interaction among many processes.
The statistical behavior of the network in the long run makes the empirical model generically contextual and even strongly contextual.
arXiv Detail & Related papers (2020-06-22T16:57:52Z) - On the Discrepancy between Density Estimation and Sequence Generation [92.70116082182076]
log-likelihood is highly correlated with BLEU when we consider models within the same family.
We observe no correlation between rankings of models across different families.
arXiv Detail & Related papers (2020-02-17T20:13:35Z) - Review of Probability Distributions for Modeling Count Data [0.0]
Generalized linear models enable direct modeling of counts in a regression context.
When counts contain only relative information, multinomial or Dirichlet-multinomial models can be more appropriate.
arXiv Detail & Related papers (2020-01-10T18:28:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.