Learning Hidden Markov Models When the Locations of Missing Observations
are Unknown
- URL: http://arxiv.org/abs/2203.06527v3
- Date: Sun, 2 Jul 2023 11:08:09 GMT
- Title: Learning Hidden Markov Models When the Locations of Missing Observations
are Unknown
- Authors: Binyamin Perets, Mark Kozdoba, Shie Mannor
- Abstract summary: We consider the general problem of learning an HMM from data with unknown missing observation locations.
We provide reconstruction algorithms that do not require any assumptions about the structure of the underlying chain.
We show that under proper specifications one can reconstruct the process dynamics as well as if the missing observations positions were known.
- Score: 54.40592050737724
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: The Hidden Markov Model (HMM) is one of the most widely used statistical
models for sequential data analysis. One of the key reasons for this
versatility is the ability of HMM to deal with missing data. However, standard
HMM learning algorithms rely crucially on the assumption that the positions of
the missing observations \emph{within the observation sequence} are known. In
the natural sciences, where this assumption is often violated, special variants
of HMM, commonly known as Silent-state HMMs (SHMMs), are used. Despite their
widespread use, these algorithms strongly rely on specific structural
assumptions of the underlying chain, such as acyclicity, thus limiting the
applicability of these methods. Moreover, even in the acyclic case, it has been
shown that these methods can lead to poor reconstruction. In this paper we
consider the general problem of learning an HMM from data with unknown missing
observation locations. We provide reconstruction algorithms that do not require
any assumptions about the structure of the underlying chain, and can also be
used with limited prior knowledge, unlike SHMM. We evaluate and compare the
algorithms in a variety of scenarios, measuring their reconstruction precision,
and robustness under model miss-specification. Notably, we show that under
proper specifications one can reconstruct the process dynamics as well as if
the missing observations positions were known.
Related papers
- Unsupervised Learning of Harmonic Analysis Based on Neural HSMM with
Code Quality Templates [0.3233195475347961]
This paper presents a method of unsupervised learning of harmonic analysis based on a hidden semi-Markov model.
We show how to recognize the tonic without prior knowledge, based on the transition probabilities of the Markov model.
arXiv Detail & Related papers (2024-03-07T01:29:48Z) - Learning Hidden Markov Models Using Conditional Samples [72.20944611510198]
This paper is concerned with the computational complexity of learning the Hidden Markov Model (HMM)
In this paper, we consider an interactive access model, in which the algorithm can query for samples from the conditional distributions of the HMMs.
Specifically, we obtain efficient algorithms for learning HMMs in settings where we have query access to the exact conditional probabilities.
arXiv Detail & Related papers (2023-02-28T16:53:41Z) - Learning to Bound Counterfactual Inference in Structural Causal Models
from Observational and Randomised Data [64.96984404868411]
We derive a likelihood characterisation for the overall data that leads us to extend a previous EM-based algorithm.
The new algorithm learns to approximate the (unidentifiability) region of model parameters from such mixed data sources.
It delivers interval approximations to counterfactual results, which collapse to points in the identifiable case.
arXiv Detail & Related papers (2022-12-06T12:42:11Z) - On Uncertainty in Deep State Space Models for Model-Based Reinforcement
Learning [21.63642325390798]
We show that RSSMs use a suboptimal inference scheme and that models trained using this inference overestimate the aleatoric uncertainty of the ground truth system.
We propose an alternative approach building on well-understood components for modeling aleatoric and epistemic uncertainty, dubbed Variational Recurrent Kalman Network (VRKN)
Our experiments show that using the VRKN instead of the RSSM improves performance in tasks where appropriately capturing aleatoric uncertainty is crucial.
arXiv Detail & Related papers (2022-10-17T16:59:48Z) - MissDAG: Causal Discovery in the Presence of Missing Data with
Continuous Additive Noise Models [78.72682320019737]
We develop a general method, which we call MissDAG, to perform causal discovery from data with incomplete observations.
MissDAG maximizes the expected likelihood of the visible part of observations under the expectation-maximization framework.
We demonstrate the flexibility of MissDAG for incorporating various causal discovery algorithms and its efficacy through extensive simulations and real data experiments.
arXiv Detail & Related papers (2022-05-27T09:59:46Z) - Towards interpretability of Mixtures of Hidden Markov Models [0.0]
Mixtures of Hidden Markov Models (MHMMs) are frequently used for clustering of sequential data.
An information-theoretic measure (entropy) is proposed for interpretability of MHMMs.
An entropy-regularized Expectation Maximization (EM) algorithm is proposed to improve interpretability.
arXiv Detail & Related papers (2021-03-23T14:25:03Z) - Robust Classification using Hidden Markov Models and Mixtures of
Normalizing Flows [25.543231171094384]
We use a generative model that combines the state transitions of a hidden Markov model (HMM) and the neural network based probability distributions for the hidden states of the HMM.
We verify the improved robustness of NMM-HMM classifiers in an application to speech recognition.
arXiv Detail & Related papers (2021-02-15T00:40:30Z) - DenseHMM: Learning Hidden Markov Models by Learning Dense
Representations [0.0]
We propose a modification of Hidden Markov Models (HMMs) that allows to learn dense representations of both the hidden states and the observables.
Compared to the standard HMM, transition probabilities are not atomic but composed of these representations via kernelization.
The properties of the DenseHMM like learned co-occurrences and log-likelihoods are studied empirically on synthetic and biomedical datasets.
arXiv Detail & Related papers (2020-12-17T17:48:27Z) - Scaling Hidden Markov Language Models [118.55908381553056]
This work revisits the challenge of scaling HMMs to language modeling datasets.
We propose methods for scaling HMMs to massive state spaces while maintaining efficient exact inference, a compact parameterization, and effective regularization.
arXiv Detail & Related papers (2020-11-09T18:51:55Z) - Robust Compressed Sensing using Generative Models [98.64228459705859]
In this paper we propose an algorithm inspired by the Median-of-Means (MOM)
Our algorithm guarantees recovery for heavy-tailed data, even in the presence of outliers.
arXiv Detail & Related papers (2020-06-16T19:07:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.