A domain-decomposed VAE method for Bayesian inverse problems
- URL: http://arxiv.org/abs/2301.05708v1
- Date: Mon, 9 Jan 2023 07:35:43 GMT
- Title: A domain-decomposed VAE method for Bayesian inverse problems
- Authors: Xu Zhihang, Xia Yingzhi, Liao Qifeng
- Abstract summary: This paper proposes a domain-decomposed variational auto-encoder Markov chain Monte Carlo (DD-VAE-MCMC) method to tackle these challenges simultaneously.
The proposed method first constructs local deterministic generative models based on local historical data, which provide efficient local prior representations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Bayesian inverse problems are often computationally challenging when the
forward model is governed by complex partial differential equations (PDEs).
This is typically caused by expensive forward model evaluations and
high-dimensional parameterization of priors. This paper proposes a
domain-decomposed variational auto-encoder Markov chain Monte Carlo
(DD-VAE-MCMC) method to tackle these challenges simultaneously. Through
partitioning the global physical domain into small subdomains, the proposed
method first constructs local deterministic generative models based on local
historical data, which provide efficient local prior representations. Gaussian
process models with active learning address the domain decomposition interface
conditions. Then inversions are conducted on each subdomain independently in
parallel and in low-dimensional latent parameter spaces. The local inference
solutions are post-processed through the Poisson image blending procedure to
result in an efficient global inference result. Numerical examples are provided
to demonstrate the performance of the proposed method.
Related papers
- VI-DGP: A variational inference method with deep generative prior for
solving high-dimensional inverse problems [0.7734726150561089]
We propose a novel approximation method for estimating the high-dimensional posterior distribution.
This approach leverages a deep generative model to learn a prior model capable of generating spatially-varying parameters.
The proposed method can be fully implemented in an automatic differentiation manner.
arXiv Detail & Related papers (2023-02-22T06:48:10Z) - Score-based Diffusion Models in Function Space [140.792362459734]
Diffusion models have recently emerged as a powerful framework for generative modeling.
We introduce a mathematically rigorous framework called Denoising Diffusion Operators (DDOs) for training diffusion models in function space.
We show that the corresponding discretized algorithm generates accurate samples at a fixed cost independent of the data resolution.
arXiv Detail & Related papers (2023-02-14T23:50:53Z) - Inverse Models for Estimating the Initial Condition of Spatio-Temporal
Advection-Diffusion Processes [5.814371485767541]
Inverse problems involve making inference about unknown parameters of a physical process using observational data.
This paper investigates the estimation of the initial condition of a-temporal advection-diffusion process using spatially sparse data streams.
arXiv Detail & Related papers (2023-02-08T15:30:16Z) - Variational Laplace Autoencoders [53.08170674326728]
Variational autoencoders employ an amortized inference model to approximate the posterior of latent variables.
We present a novel approach that addresses the limited posterior expressiveness of fully-factorized Gaussian assumption.
We also present a general framework named Variational Laplace Autoencoders (VLAEs) for training deep generative models.
arXiv Detail & Related papers (2022-11-30T18:59:27Z) - Towards a machine learning pipeline in reduced order modelling for
inverse problems: neural networks for boundary parametrization,
dimensionality reduction and solution manifold approximation [0.0]
Inverse problems, especially in a partial differential equation context, require a huge computational load.
We apply a numerical pipeline that involves artificial neural networks to parametrize the boundary conditions of the problem in hand.
It derives a general framework capable to provide an ad-hoc parametrization of the inlet boundary and quickly converges to the optimal solution.
arXiv Detail & Related papers (2022-10-26T14:53:07Z) - Super-model ecosystem: A domain-adaptation perspective [101.76769818069072]
This paper attempts to establish the theoretical foundation for the emerging super-model paradigm via domain adaptation.
Super-model paradigms help reduce computational and data cost and carbon emission, which is critical to AI industry.
arXiv Detail & Related papers (2022-08-30T09:09:43Z) - An application of the splitting-up method for the computation of a
neural network representation for the solution for the filtering equations [68.8204255655161]
Filtering equations play a central role in many real-life applications, including numerical weather prediction, finance and engineering.
One of the classical approaches to approximate the solution of the filtering equations is to use a PDE inspired method, called the splitting-up method.
We combine this method with a neural network representation to produce an approximation of the unnormalised conditional distribution of the signal process.
arXiv Detail & Related papers (2022-01-10T11:01:36Z) - Model-Based Domain Generalization [96.84818110323518]
We propose a novel approach for the domain generalization problem called Model-Based Domain Generalization.
Our algorithms beat the current state-of-the-art methods on the very-recently-proposed WILDS benchmark by up to 20 percentage points.
arXiv Detail & Related papers (2021-02-23T00:59:02Z) - Distributed Variational Bayesian Algorithms Over Sensor Networks [6.572330981878818]
We propose two novel distributed VB algorithms for general Bayesian inference problem.
The proposed algorithms have excellent performance, which are almost as good as the corresponding centralized VB algorithm relying on all data available in a fusion center.
arXiv Detail & Related papers (2020-11-27T08:12:18Z) - On the implementation of a global optimization method for mixed-variable
problems [0.30458514384586394]
The algorithm is based on the radial basis function of Gutmann and the metric response surface method of Regis and Shoemaker.
We propose several modifications aimed at generalizing and improving these two algorithms.
arXiv Detail & Related papers (2020-09-04T13:36:56Z) - Model Fusion with Kullback--Leibler Divergence [58.20269014662046]
We propose a method to fuse posterior distributions learned from heterogeneous datasets.
Our algorithm relies on a mean field assumption for both the fused model and the individual dataset posteriors.
arXiv Detail & Related papers (2020-07-13T03:27:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.