A VAE Approach to Sample Multivariate Extremes
- URL: http://arxiv.org/abs/2306.10987v1
- Date: Mon, 19 Jun 2023 14:53:40 GMT
- Title: A VAE Approach to Sample Multivariate Extremes
- Authors: Nicolas Lafon, Philippe Naveau, Ronan Fablet
- Abstract summary: This paper describes a variational autoencoder (VAE) approach for sampling heavy-tailed distributions likely to have extremes of particularly large intensities.
We illustrate the relevance of our approach on a synthetic data set and on a real data set of discharge measurements along the Danube river network.
In addition to outperforming the standard VAE for the tested data sets, we also provide a comparison with a competing EVT-based generative approach.
- Score: 6.548734807475054
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generating accurate extremes from an observational data set is crucial when
seeking to estimate risks associated with the occurrence of future extremes
which could be larger than those already observed. Applications range from the
occurrence of natural disasters to financial crashes. Generative approaches
from the machine learning community do not apply to extreme samples without
careful adaptation. Besides, asymptotic results from extreme value theory (EVT)
give a theoretical framework to model multivariate extreme events, especially
through the notion of multivariate regular variation. Bridging these two
fields, this paper details a variational autoencoder (VAE) approach for
sampling multivariate heavy-tailed distributions, i.e., distributions likely to
have extremes of particularly large intensities. We illustrate the relevance of
our approach on a synthetic data set and on a real data set of discharge
measurements along the Danube river network. The latter shows the potential of
our approach for flood risks' assessment. In addition to outperforming the
standard VAE for the tested data sets, we also provide a comparison with a
competing EVT-based generative approach. On the tested cases, our approach
improves the learning of the dependency structure between extremes.
Related papers
- Distributionally Robust Skeleton Learning of Discrete Bayesian Networks [9.46389554092506]
We consider the problem of learning the exact skeleton of general discrete Bayesian networks from potentially corrupted data.
We propose to optimize the most adverse risk over a family of distributions within bounded Wasserstein distance or KL divergence to the empirical distribution.
We present efficient algorithms and show the proposed methods are closely related to the standard regularized regression approach.
arXiv Detail & Related papers (2023-11-10T15:33:19Z) - Generalized Mixture Model for Extreme Events Forecasting in Time Series
Data [10.542258423966492]
Time Series Forecasting (TSF) is a widely researched topic with broad applications in weather forecasting, traffic control, and stock price prediction.
Extreme values in time series often significantly impact human and natural systems, but predicting them is challenging due to their rare occurrence.
We propose a novel framework to enhance the focus on extreme events. Specifically, we propose a Deep Extreme Mixture Model with Autoencoder (DEMMA) for time series prediction.
arXiv Detail & Related papers (2023-10-11T12:36:42Z) - Deep Anti-Regularized Ensembles provide reliable out-of-distribution
uncertainty quantification [4.750521042508541]
Deep ensemble often return overconfident estimates outside the training domain.
We show that an ensemble of networks with large weights fitting the training data are likely to meet these two objectives.
We derive a theoretical framework for this approach and show that the proposed optimization can be seen as a "water-filling" problem.
arXiv Detail & Related papers (2023-04-08T15:25:12Z) - Tailoring Language Generation Models under Total Variation Distance [55.89964205594829]
The standard paradigm of neural language generation adopts maximum likelihood estimation (MLE) as the optimizing method.
We develop practical bounds to apply it to language generation.
We introduce the TaiLr objective that balances the tradeoff of estimating TVD.
arXiv Detail & Related papers (2023-02-26T16:32:52Z) - Pseudo value-based Deep Neural Networks for Multi-state Survival
Analysis [9.659041001051415]
We propose a new class of pseudo-value-based deep learning models for multi-state survival analysis.
Our proposed models achieve state-of-the-art results under various censoring settings.
arXiv Detail & Related papers (2022-07-12T03:58:05Z) - Trustworthy Multimodal Regression with Mixture of Normal-inverse Gamma
Distributions [91.63716984911278]
We introduce a novel Mixture of Normal-Inverse Gamma distributions (MoNIG) algorithm, which efficiently estimates uncertainty in principle for adaptive integration of different modalities and produces a trustworthy regression result.
Experimental results on both synthetic and different real-world data demonstrate the effectiveness and trustworthiness of our method on various multimodal regression tasks.
arXiv Detail & Related papers (2021-11-11T14:28:12Z) - Regularizing Variational Autoencoder with Diversity and Uncertainty
Awareness [61.827054365139645]
Variational Autoencoder (VAE) approximates the posterior of latent variables based on amortized variational inference.
We propose an alternative model, DU-VAE, for learning a more Diverse and less Uncertain latent space.
arXiv Detail & Related papers (2021-10-24T07:58:13Z) - GANs with Variational Entropy Regularizers: Applications in Mitigating
the Mode-Collapse Issue [95.23775347605923]
Building on the success of deep learning, Generative Adversarial Networks (GANs) provide a modern approach to learn a probability distribution from observed samples.
GANs often suffer from the mode collapse issue where the generator fails to capture all existing modes of the input distribution.
We take an information-theoretic approach and maximize a variational lower bound on the entropy of the generated samples to increase their diversity.
arXiv Detail & Related papers (2020-09-24T19:34:37Z) - Improving Maximum Likelihood Training for Text Generation with Density
Ratio Estimation [51.091890311312085]
We propose a new training scheme for auto-regressive sequence generative models, which is effective and stable when operating at large sample space encountered in text generation.
Our method stably outperforms Maximum Likelihood Estimation and other state-of-the-art sequence generative models in terms of both quality and diversity.
arXiv Detail & Related papers (2020-07-12T15:31:24Z) - Decision-Making with Auto-Encoding Variational Bayes [71.44735417472043]
We show that a posterior approximation distinct from the variational distribution should be used for making decisions.
Motivated by these theoretical results, we propose learning several approximate proposals for the best model.
In addition to toy examples, we present a full-fledged case study of single-cell RNA sequencing.
arXiv Detail & Related papers (2020-02-17T19:23:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.