Bridging the Usability Gap: Theoretical and Methodological Advances for Spectral Learning of Hidden Markov Models
- URL: http://arxiv.org/abs/2302.07437v3
- Date: Mon, 26 Aug 2024 14:46:08 GMT
- Title: Bridging the Usability Gap: Theoretical and Methodological Advances for Spectral Learning of Hidden Markov Models
- Authors: Xiaoyuan Ma, Jordan Rodu,
- Abstract summary: The Baum-Welch (B-W) algorithm is the most widely accepted method for inferring hidden Markov models (HMM)
It is prone to getting stuck in local optima, and can be too slow for many real-time applications.
We propose a novel algorithm called projected SHMM (PSHMM) that mitigates the problem of error propagation.
- Score: 0.8287206589886879
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Baum-Welch (B-W) algorithm is the most widely accepted method for inferring hidden Markov models (HMM). However, it is prone to getting stuck in local optima, and can be too slow for many real-time applications. Spectral learning of HMMs (SHMM), based on the method of moments (MOM) has been proposed in the literature to overcome these obstacles. Despite its promises, asymptotic theory for SHMM has been elusive, and the long-run performance of SHMM can degrade due to unchecked propagation of error. In this paper, we (1) provide an asymptotic distribution for the approximate error of the likelihood estimated by SHMM, (2) propose a novel algorithm called projected SHMM (PSHMM) that mitigates the problem of error propagation, and (3) develop online learning variants of both SHMM and PSHMM that accommodate potential nonstationarity. We compare the performance of SHMM with PSHMM and estimation through the B-W algorithm on both simulated data and data from real world applications, and find that PSHMM not only retains the computational advantages of SHMM, but also provides more robust estimation and forecasting.
Related papers
- Sample-efficient neural likelihood-free Bayesian inference of implicit HMMs [1.8843687952462742]
We propose a novel, sample-efficient likelihood-free method for estimating the high-dimensional hidden states of an implicit HMM.
Our approach relies on learning directly the intractable posterior distribution of the hidden states, using an autoregressive-flow, by exploiting the Markov property.
arXiv Detail & Related papers (2024-05-02T21:13:34Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Mixture of Coupled HMMs for Robust Modeling of Multivariate Healthcare
Time Series [7.5986411724707095]
We propose a novel class of models, a mixture of coupled hidden Markov models (M-CHMM)
To make the model learning feasible, we derive two algorithms to sample the sequences of the latent variables in the CHMM.
Compared to existing inference methods, our algorithms are computationally tractable, improve mixing, and allow for likelihood estimation.
arXiv Detail & Related papers (2023-11-14T02:55:37Z) - SGMM: Stochastic Approximation to Generalized Method of Moments [8.48870560391056]
We introduce a new class of algorithms, Generalized Method of Moments (SGMM) for estimation and inference on (overidentified) moment restriction models.
Our SGMM is a novel approximation to the popular Hansen (1982) (offline) GMM, and offers fast and scalable implementation with the ability to handle streaming datasets in real time.
arXiv Detail & Related papers (2023-08-25T00:22:45Z) - Efficient Learning and Decoding of the Continuous-Time Hidden Markov
Model for Disease Progression Modeling [119.50438407358862]
We present the first complete characterization of efficient EM-based learning methods for CT-HMM models.
We show that EM-based learning consists of two challenges: the estimation of posterior state probabilities and the computation of end-state conditioned statistics.
We demonstrate the use of CT-HMMs with more than 100 states to visualize and predict disease progression using a glaucoma dataset and an Alzheimer's disease dataset.
arXiv Detail & Related papers (2021-10-26T20:06:05Z) - Towards interpretability of Mixtures of Hidden Markov Models [0.0]
Mixtures of Hidden Markov Models (MHMMs) are frequently used for clustering of sequential data.
An information-theoretic measure (entropy) is proposed for interpretability of MHMMs.
An entropy-regularized Expectation Maximization (EM) algorithm is proposed to improve interpretability.
arXiv Detail & Related papers (2021-03-23T14:25:03Z) - DenseHMM: Learning Hidden Markov Models by Learning Dense
Representations [0.0]
We propose a modification of Hidden Markov Models (HMMs) that allows to learn dense representations of both the hidden states and the observables.
Compared to the standard HMM, transition probabilities are not atomic but composed of these representations via kernelization.
The properties of the DenseHMM like learned co-occurrences and log-likelihoods are studied empirically on synthetic and biomedical datasets.
arXiv Detail & Related papers (2020-12-17T17:48:27Z) - Efficient semidefinite-programming-based inference for binary and
multi-class MRFs [83.09715052229782]
We propose an efficient method for computing the partition function or MAP estimate in a pairwise MRF.
We extend semidefinite relaxations from the typical binary MRF to the full multi-class setting, and develop a compact semidefinite relaxation that can again be solved efficiently using the solver.
arXiv Detail & Related papers (2020-12-04T15:36:29Z) - DS-UI: Dual-Supervised Mixture of Gaussian Mixture Models for
Uncertainty Inference [52.899219617256655]
We propose a dual-supervised uncertainty inference (DS-UI) framework for improving Bayesian estimation-based uncertainty inference (UI) in deep neural network (DNN)-based image recognition.
In the DS-UI, we combine the last fully-connected (FC) layer with a mixture of Gaussian mixture models (MoGMM) to obtain an MoGMM-FC layer.
Experimental results show the DS-UI outperforms the state-of-the-art UI methods in misclassification detection.
arXiv Detail & Related papers (2020-11-17T12:35:02Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - Q-GADMM: Quantized Group ADMM for Communication Efficient Decentralized Machine Learning [66.18202188565922]
We propose a communication-efficient decentralized machine learning (ML) algorithm, coined QGADMM (QGADMM)
We develop a novel quantization method to adaptively adjust modelization levels and their probabilities, while proving the convergence of QGADMM for convex functions.
arXiv Detail & Related papers (2019-10-23T10:47:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.