Dynamical Variational Autoencoders: A Comprehensive Review
- URL: http://arxiv.org/abs/2008.12595v4
- Date: Mon, 4 Jul 2022 16:11:08 GMT
- Title: Dynamical Variational Autoencoders: A Comprehensive Review
- Authors: Laurent Girin and Simon Leglaive and Xiaoyu Bie and Julien Diard and
Thomas Hueber and Xavier Alameda-Pineda
- Abstract summary: We introduce and discuss a general class of models, called dynamical variational autoencoders (DVAEs)
We present in detail seven recently proposed DVAE models, with an aim to homogenize the notations and presentation lines.
We have reimplemented those seven DVAE models and present the results of an experimental benchmark conducted on the speech analysis-resynthesis task.
- Score: 23.25573952809074
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational autoencoders (VAEs) are powerful deep generative models widely
used to represent high-dimensional complex data through a low-dimensional
latent space learned in an unsupervised manner. In the original VAE model, the
input data vectors are processed independently. Recently, a series of papers
have presented different extensions of the VAE to process sequential data,
which model not only the latent space but also the temporal dependencies within
a sequence of data vectors and corresponding latent vectors, relying on
recurrent neural networks or state-space models. In this paper, we perform a
literature review of these models. We introduce and discuss a general class of
models, called dynamical variational autoencoders (DVAEs), which encompasses a
large subset of these temporal VAE extensions. Then, we present in detail seven
recently proposed DVAE models, with an aim to homogenize the notations and
presentation lines, as well as to relate these models with existing classical
temporal models. We have reimplemented those seven DVAE models and present the
results of an experimental benchmark conducted on the speech
analysis-resynthesis task (the PyTorch code is made publicly available). The
paper concludes with a discussion on important issues concerning the DVAE class
of models and future research guidelines.
Related papers
- PDETime: Rethinking Long-Term Multivariate Time Series Forecasting from
the perspective of partial differential equations [49.80959046861793]
We present PDETime, a novel LMTF model inspired by the principles of Neural PDE solvers.
Our experimentation across seven diversetemporal real-world LMTF datasets reveals that PDETime adapts effectively to the intrinsic nature of the data.
arXiv Detail & Related papers (2024-02-25T17:39:44Z) - Generative Modeling of Regular and Irregular Time Series Data via Koopman VAEs [50.25683648762602]
We introduce Koopman VAE, a new generative framework that is based on a novel design for the model prior.
Inspired by Koopman theory, we represent the latent conditional prior dynamics using a linear map.
KoVAE outperforms state-of-the-art GAN and VAE methods across several challenging synthetic and real-world time series generation benchmarks.
arXiv Detail & Related papers (2023-10-04T07:14:43Z) - Speech Modeling with a Hierarchical Transformer Dynamical VAE [23.847366888695266]
We propose to model speech signals with the Hierarchical Transformer DVAE (HiT-DVAE)
We show that HiT-DVAE outperforms several other DVAEs for speech spectrogram modeling, while enabling a simpler training procedure.
arXiv Detail & Related papers (2023-03-07T13:35:45Z) - A Variational Autoencoder for Heterogeneous Temporal and Longitudinal
Data [0.3749861135832073]
Recently proposed extensions to VAEs that can handle temporal and longitudinal data have applications in healthcare, behavioural modelling, and predictive maintenance.
We propose the heterogeneous longitudinal VAE (HL-VAE) that extends the existing temporal and longitudinal VAEs to heterogeneous data.
HL-VAE provides efficient inference for high-dimensional datasets and includes likelihood models for continuous, count, categorical, and ordinal data.
arXiv Detail & Related papers (2022-04-20T10:18:39Z) - Dynamically-Scaled Deep Canonical Correlation Analysis [77.34726150561087]
Canonical Correlation Analysis (CCA) is a method for feature extraction of two views by finding maximally correlated linear projections of them.
We introduce a novel dynamic scaling method for training an input-dependent canonical correlation model.
arXiv Detail & Related papers (2022-03-23T12:52:49Z) - Unsupervised Multiple-Object Tracking with a Dynamical Variational
Autoencoder [25.293475313066967]
We present an unsupervised probabilistic model and associated estimation algorithm for multi-object tracking (MOT) based on a dynamical variational autoencoder (DVAE)
DVAE is a latent-variable deep generative model that can be seen as an extension of the variational autoencoder for the modeling of temporal sequences.
It is included in DVAE-UMOT to model the objects' dynamics, after being pre-trained on an unlabeled synthetic dataset single-object trajectories.
arXiv Detail & Related papers (2022-02-18T17:27:27Z) - Multivariate Data Explanation by Jumping Emerging Patterns Visualization [78.6363825307044]
We present VAX (multiVariate dAta eXplanation), a new VA method to support the identification and visual interpretation of patterns in multivariate data sets.
Unlike the existing similar approaches, VAX uses the concept of Jumping Emerging Patterns to identify and aggregate several diversified patterns, producing explanations through logic combinations of data variables.
arXiv Detail & Related papers (2021-06-21T13:49:44Z) - Discrete Auto-regressive Variational Attention Models for Text Modeling [53.38382932162732]
Variational autoencoders (VAEs) have been widely applied for text modeling.
They are troubled by two challenges: information underrepresentation and posterior collapse.
We propose Discrete Auto-regressive Variational Attention Model (DAVAM) to address the challenges.
arXiv Detail & Related papers (2021-06-16T06:36:26Z) - Anomaly Detection of Time Series with Smoothness-Inducing Sequential
Variational Auto-Encoder [59.69303945834122]
We present a Smoothness-Inducing Sequential Variational Auto-Encoder (SISVAE) model for robust estimation and anomaly detection of time series.
Our model parameterizes mean and variance for each time-stamp with flexible neural networks.
We show the effectiveness of our model on both synthetic datasets and public real-world benchmarks.
arXiv Detail & Related papers (2021-02-02T06:15:15Z) - Multivariate Temporal Autoencoder for Predictive Reconstruction of Deep
Sequences [0.0]
Time series sequence prediction and modelling has proven to be a challenging endeavor in real world datasets.
Two key issues are the multi-dimensionality of data and the interaction of independent dimensions forming a latent output signal.
This paper proposes a multi-branch deep neural network approach to tackling the aforementioned problems by modelling a latent state vector representation of data windows.
arXiv Detail & Related papers (2020-10-07T21:25:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.