DenseHMM: Learning Hidden Markov Models by Learning Dense
Representations
- URL: http://arxiv.org/abs/2012.09783v1
- Date: Thu, 17 Dec 2020 17:48:27 GMT
- Title: DenseHMM: Learning Hidden Markov Models by Learning Dense
Representations
- Authors: Joachim Sicking, Maximilian Pintz, Maram Akila, Tim Wirtz
- Abstract summary: We propose a modification of Hidden Markov Models (HMMs) that allows to learn dense representations of both the hidden states and the observables.
Compared to the standard HMM, transition probabilities are not atomic but composed of these representations via kernelization.
The properties of the DenseHMM like learned co-occurrences and log-likelihoods are studied empirically on synthetic and biomedical datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose DenseHMM - a modification of Hidden Markov Models (HMMs) that
allows to learn dense representations of both the hidden states and the
observables. Compared to the standard HMM, transition probabilities are not
atomic but composed of these representations via kernelization. Our approach
enables constraint-free and gradient-based optimization. We propose two
optimization schemes that make use of this: a modification of the Baum-Welch
algorithm and a direct co-occurrence optimization. The latter one is highly
scalable and comes empirically without loss of performance compared to standard
HMMs. We show that the non-linearity of the kernelization is crucial for the
expressiveness of the representations. The properties of the DenseHMM like
learned co-occurrences and log-likelihoods are studied empirically on synthetic
and biomedical datasets.
Related papers
- Recursive Learning of Asymptotic Variational Objectives [49.69399307452126]
General state-space models (SSMs) are widely used in statistical machine learning and are among the most classical generative models for sequential time-series data.
Online sequential IWAE (OSIWAE) allows for online learning of both model parameters and a Markovian recognition model for inferring latent states.
This approach is more theoretically well-founded than recently proposed online variational SMC methods.
arXiv Detail & Related papers (2024-11-04T16:12:37Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Algebraic Reduction of Hidden Markov Models [0.0]
We propose two algorithms that return models that exactly reproduce the single-time distribution of a given output process.
The reduction method exploits not only the structure of the observed output, but also its initial condition.
Optimal algorithms are derived for a class of HMM, namely ones.
arXiv Detail & Related papers (2022-08-11T02:46:05Z) - Learning Hidden Markov Models When the Locations of Missing Observations
are Unknown [54.40592050737724]
We consider the general problem of learning an HMM from data with unknown missing observation locations.
We provide reconstruction algorithms that do not require any assumptions about the structure of the underlying chain.
We show that under proper specifications one can reconstruct the process dynamics as well as if the missing observations positions were known.
arXiv Detail & Related papers (2022-03-12T22:40:43Z) - Towards interpretability of Mixtures of Hidden Markov Models [0.0]
Mixtures of Hidden Markov Models (MHMMs) are frequently used for clustering of sequential data.
An information-theoretic measure (entropy) is proposed for interpretability of MHMMs.
An entropy-regularized Expectation Maximization (EM) algorithm is proposed to improve interpretability.
arXiv Detail & Related papers (2021-03-23T14:25:03Z) - Robust Classification using Hidden Markov Models and Mixtures of
Normalizing Flows [25.543231171094384]
We use a generative model that combines the state transitions of a hidden Markov model (HMM) and the neural network based probability distributions for the hidden states of the HMM.
We verify the improved robustness of NMM-HMM classifiers in an application to speech recognition.
arXiv Detail & Related papers (2021-02-15T00:40:30Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Scaling Hidden Markov Language Models [118.55908381553056]
This work revisits the challenge of scaling HMMs to language modeling datasets.
We propose methods for scaling HMMs to massive state spaces while maintaining efficient exact inference, a compact parameterization, and effective regularization.
arXiv Detail & Related papers (2020-11-09T18:51:55Z) - A Rigorous Link Between Self-Organizing Maps and Gaussian Mixture Models [78.6363825307044]
This work presents a mathematical treatment of the relation between Self-Organizing Maps (SOMs) and Gaussian Mixture Models (GMMs)
We show that energy-based SOM models can be interpreted as performing gradient descent.
This link allows to treat SOMs as generative probabilistic models, giving a formal justification for using SOMs to detect outliers, or for sampling.
arXiv Detail & Related papers (2020-09-24T14:09:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.