Scaling Up Bayesian Uncertainty Quantification for Inverse Problems
using Deep Neural Networks
- URL: http://arxiv.org/abs/2101.03906v1
- Date: Mon, 11 Jan 2021 14:18:38 GMT
- Title: Scaling Up Bayesian Uncertainty Quantification for Inverse Problems
using Deep Neural Networks
- Authors: Shiwei Lan, Shuyi Li, Babak Shahbaba
- Abstract summary: We propose a novel CES approach for Bayesian inference based on deep neural network (DNN) models for the emulation phase.
The resulting algorithm is not only computationally more efficient, but also less sensitive to the training set.
Overall, our method, henceforth called emphReduced- Dimension Emulative Autoencoder Monte Carlo (DREAM) algorithm, is able to scale Bayesian UQ up to thousands of dimensions in physics-constrained inverse problems.
- Score: 2.455468619225742
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Due to the importance of uncertainty quantification (UQ), Bayesian approach
to inverse problems has recently gained popularity in applied mathematics,
physics, and engineering. However, traditional Bayesian inference methods based
on Markov Chain Monte Carlo (MCMC) tend to be computationally intensive and
inefficient for such high dimensional problems. To address this issue, several
methods based on surrogate models have been proposed to speed up the inference
process. More specifically, the calibration-emulation-sampling (CES) scheme has
been proven to be successful in large dimensional UQ problems. In this work, we
propose a novel CES approach for Bayesian inference based on deep neural
network (DNN) models for the emulation phase. The resulting algorithm is not
only computationally more efficient, but also less sensitive to the training
set. Further, by using an Autoencoder (AE) for dimension reduction, we have
been able to speed up our Bayesian inference method up to three orders of
magnitude. Overall, our method, henceforth called \emph{Dimension-Reduced
Emulative Autoencoder Monte Carlo (DREAM)} algorithm, is able to scale Bayesian
UQ up to thousands of dimensions in physics-constrained inverse problems. Using
two low-dimensional (linear and nonlinear) inverse problems we illustrate the
validity this approach. Next, we apply our method to two high-dimensional
numerical examples (elliptic and advection-diffussion) to demonstrate its
computational advantage over existing algorithms.
Related papers
- Pushing the Limits of Large Language Model Quantization via the Linearity Theorem [71.3332971315821]
We present a "line theoremarity" establishing a direct relationship between the layer-wise $ell$ reconstruction error and the model perplexity increase due to quantization.
This insight enables two novel applications: (1) a simple data-free LLM quantization method using Hadamard rotations and MSE-optimal grids, dubbed HIGGS, and (2) an optimal solution to the problem of finding non-uniform per-layer quantization levels.
arXiv Detail & Related papers (2024-11-26T15:35:44Z) - On Uncertainty Quantification for Near-Bayes Optimal Algorithms [2.622066970118316]
We show that it is possible to recover the Bayesian posterior defined by the task distribution, which is unknown but optimal in this setting, by building a martingale posterior using the algorithm.
Experiments based on a variety of non-NN and NN algorithms demonstrate the efficacy of our method.
arXiv Detail & Related papers (2024-03-28T12:42:25Z) - An Efficient Algorithm for Clustered Multi-Task Compressive Sensing [60.70532293880842]
Clustered multi-task compressive sensing is a hierarchical model that solves multiple compressive sensing tasks.
The existing inference algorithm for this model is computationally expensive and does not scale well in high dimensions.
We propose a new algorithm that substantially accelerates model inference by avoiding the need to explicitly compute these covariance matrices.
arXiv Detail & Related papers (2023-09-30T15:57:14Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Efficient Bayesian Physics Informed Neural Networks for Inverse Problems
via Ensemble Kalman Inversion [0.0]
We present a new efficient inference algorithm for B-PINNs that uses Ensemble Kalman Inversion (EKI) for high-dimensional inference tasks.
We find that our proposed method can achieve inference results with informative uncertainty estimates comparable to Hamiltonian Monte Carlo (HMC)-based B-PINNs with a much reduced computational cost.
arXiv Detail & Related papers (2023-03-13T18:15:26Z) - Entropic Neural Optimal Transport via Diffusion Processes [105.34822201378763]
We propose a novel neural algorithm for the fundamental problem of computing the entropic optimal transport (EOT) plan between continuous probability distributions.
Our algorithm is based on the saddle point reformulation of the dynamic version of EOT which is known as the Schr"odinger Bridge problem.
In contrast to the prior methods for large-scale EOT, our algorithm is end-to-end and consists of a single learning step.
arXiv Detail & Related papers (2022-11-02T14:35:13Z) - Recent advances in Bayesian optimization with applications to parameter
reconstruction in optical nano-metrology [0.0]
reconstruction is a common problem in optical nano metrology.
We present a Bayesian Target Vector Optimization scheme which combines two approaches.
We find that the presented method generally uses fewer calls of the model function than any of the competing schemes to achieve similar reconstruction performance.
arXiv Detail & Related papers (2021-07-12T15:32:15Z) - SHINE: SHaring the INverse Estimate from the forward pass for bi-level
optimization and implicit models [15.541264326378366]
In recent years, implicit deep learning has emerged as a method to increase the depth of deep neural networks.
The training is performed as a bi-level problem, and its computational complexity is partially driven by the iterative inversion of a huge Jacobian matrix.
We propose a novel strategy to tackle this computational bottleneck from which many bi-level problems suffer.
arXiv Detail & Related papers (2021-06-01T15:07:34Z) - Novel Deep neural networks for solving Bayesian statistical inverse [0.0]
We introduce a fractional deep neural network based approach for the forward solves within an MCMC routine.
We discuss some approximation error estimates and illustrate the efficiency of our approach via several numerical examples.
arXiv Detail & Related papers (2021-02-08T02:54:46Z) - Deep Magnification-Flexible Upsampling over 3D Point Clouds [103.09504572409449]
We propose a novel end-to-end learning-based framework to generate dense point clouds.
We first formulate the problem explicitly, which boils down to determining the weights and high-order approximation errors.
Then, we design a lightweight neural network to adaptively learn unified and sorted weights as well as the high-order refinements.
arXiv Detail & Related papers (2020-11-25T14:00:18Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.