Infinity Learning: Learning Markov Chains from Aggregate Steady-State
Observations
- URL: http://arxiv.org/abs/2002.04186v1
- Date: Tue, 11 Feb 2020 03:29:13 GMT
- Title: Infinity Learning: Learning Markov Chains from Aggregate Steady-State
Observations
- Authors: Jianfei Gao, Mohamed A. Zahran, Amit Sheoran, Sonia Fahmy, Bruno
Ribeiro
- Abstract summary: We consider the task of learning a parametric Continuous Time Markov Chain (CTMC) sequence model without examples of sequences.
We propose $infty$-SGD, a gradient descent method that uses randomly-stopped estimators to avoid infinite sums required by the steady state.
We apply $infty$-SGD to a real-world testbed and synthetic experiments showcasing its accuracy, ability to extrapolate the steady state distribution to unobserved states.
- Score: 13.973232545822247
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: We consider the task of learning a parametric Continuous Time Markov Chain
(CTMC) sequence model without examples of sequences, where the training data
consists entirely of aggregate steady-state statistics. Making the problem
harder, we assume that the states we wish to predict are unobserved in the
training data. Specifically, given a parametric model over the transition rates
of a CTMC and some known transition rates, we wish to extrapolate its steady
state distribution to states that are unobserved. A technical roadblock to
learn a CTMC from its steady state has been that the chain rule to compute
gradients will not work over the arbitrarily long sequences necessary to reach
steady state ---from where the aggregate statistics are sampled. To overcome
this optimization challenge, we propose $\infty$-SGD, a principled stochastic
gradient descent method that uses randomly-stopped estimators to avoid infinite
sums required by the steady state computation, while learning even when only a
subset of the CTMC states can be observed. We apply $\infty$-SGD to a
real-world testbed and synthetic experiments showcasing its accuracy, ability
to extrapolate the steady state distribution to unobserved states under
unobserved conditions (heavy loads, when training under light loads), and
succeeding in difficult scenarios where even a tailor-made extension of
existing methods fails.
Related papers
- Amortized Control of Continuous State Space Feynman-Kac Model for Irregular Time Series [14.400596021890863]
Many real-world datasets, such as healthcare, climate, and economics, are often collected as irregular time series.
We propose the Amortized Control of continuous State Space Model (ACSSM) for continuous dynamical modeling of time series.
arXiv Detail & Related papers (2024-10-08T01:27:46Z) - Learning Unstable Continuous-Time Stochastic Linear Control Systems [0.0]
We study the problem of system identification for continuous-time dynamics, based on a single finite-length state trajectory.
We present a method for estimating the possibly unstable open-loop matrix by employing properly randomized control inputs.
We establish theoretical performance guarantees showing that the estimation error decays with trajectory length, a measure of excitability, and the signal-to-noise ratio.
arXiv Detail & Related papers (2024-09-17T16:24:51Z) - Data-driven Bayesian State Estimation with Compressed Measurement of Model-free Process using Semi-supervised Learning [57.04370580292727]
The research topic is: data-driven Bayesian state estimation with compressed measurement (BSCM) of model-free process.
The dimension of the temporal measurement vector is lower than the dimension of the temporal state vector to be estimated.
Two existing unsupervised learning-based data-driven methods fail to address the BSCM problem for model-free process.
We develop a semi-supervised learning-based DANSE method, referred to as SemiDANSE.
arXiv Detail & Related papers (2024-07-10T05:03:48Z) - Online Variational Sequential Monte Carlo [49.97673761305336]
We build upon the variational sequential Monte Carlo (VSMC) method, which provides computationally efficient and accurate model parameter estimation and Bayesian latent-state inference.
Online VSMC is capable of performing efficiently, entirely on-the-fly, both parameter estimation and particle proposal adaptation.
arXiv Detail & Related papers (2023-12-19T21:45:38Z) - Likelihood-Free Inference in State-Space Models with Unknown Dynamics [71.94716503075645]
We introduce a method for inferring and predicting latent states in state-space models where observations can only be simulated, and transition dynamics are unknown.
We propose a way of doing likelihood-free inference (LFI) of states and state prediction with a limited number of simulations.
arXiv Detail & Related papers (2021-11-02T12:33:42Z) - Efficient Learning and Decoding of the Continuous-Time Hidden Markov
Model for Disease Progression Modeling [119.50438407358862]
We present the first complete characterization of efficient EM-based learning methods for CT-HMM models.
We show that EM-based learning consists of two challenges: the estimation of posterior state probabilities and the computation of end-state conditioned statistics.
We demonstrate the use of CT-HMMs with more than 100 states to visualize and predict disease progression using a glaucoma dataset and an Alzheimer's disease dataset.
arXiv Detail & Related papers (2021-10-26T20:06:05Z) - Comparing Probability Distributions with Conditional Transport [63.11403041984197]
We propose conditional transport (CT) as a new divergence and approximate it with the amortized CT (ACT) cost.
ACT amortizes the computation of its conditional transport plans and comes with unbiased sample gradients that are straightforward to compute.
On a wide variety of benchmark datasets generative modeling, substituting the default statistical distance of an existing generative adversarial network with ACT is shown to consistently improve the performance.
arXiv Detail & Related papers (2020-12-28T05:14:22Z) - Deep Learning based Uncertainty Decomposition for Real-time Control [9.067368638784355]
We propose a novel method for detecting the absence of training data using deep learning.
We show its advantages over existing approaches on synthetic and real-world datasets.
We further demonstrate the practicality of this uncertainty estimate in deploying online data-efficient control on a simulated quadcopter.
arXiv Detail & Related papers (2020-10-06T10:46:27Z) - Targeted stochastic gradient Markov chain Monte Carlo for hidden Markov models with rare latent states [48.705095800341944]
Markov chain Monte Carlo (MCMC) algorithms for hidden Markov models often rely on the forward-backward sampler.
This makes them computationally slow as the length of the time series increases, motivating the development of sub-sampling-based approaches.
We propose a targeted sub-sampling approach that over-samples observations corresponding to rare latent states when calculating the gradient of parameters associated with them.
arXiv Detail & Related papers (2018-10-31T17:44:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.