Gaussian process regression and conditional Karhunen-Lo\'{e}ve models
for data assimilation in inverse problems
- URL: http://arxiv.org/abs/2301.11279v1
- Date: Thu, 26 Jan 2023 18:14:12 GMT
- Title: Gaussian process regression and conditional Karhunen-Lo\'{e}ve models
for data assimilation in inverse problems
- Authors: Yu-Hong Yeung and David A. Barajas-Solano and Alexandre M. Tartakovsky
- Abstract summary: We present a model inversion algorithm, CKLEMAP, for data assimilation and parameter estimation in partial differential equation models.
The CKLEMAP method provides better scalability compared to the standard MAP method.
- Score: 68.8204255655161
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a model inversion algorithm, CKLEMAP, for data assimilation and
parameter estimation in partial differential equation models of physical
systems with spatially heterogeneous parameter fields. These fields are
approximated using low-dimensional conditional Karhunen-Lo\'{e}ve expansions,
which are constructed using Gaussian process regression models of these fields
trained on the parameters' measurements. We then assimilate measurements of the
state of the system and compute the maximum a posteriori estimate of the CKLE
coefficients by solving a nonlinear least-squares problem. When solving this
optimization problem, we efficiently compute the Jacobian of the vector
objective by exploiting the sparsity structure of the linear system of
equations associated with the forward solution of the physics problem. The
CKLEMAP method provides better scalability compared to the standard MAP method.
In the MAP method, the number of unknowns to be estimated is equal to the
number of elements in the numerical forward model. On the other hand, in
CKLEMAP, the number of unknowns (CKLE coefficients) is controlled by the
smoothness of the parameter field and the number of measurements, and is in
general much smaller than the number of discretization nodes, which leads to a
significant reduction of computational cost with respect to the standard MAP
method. To show its advantage in scalability, we apply CKLEMAP to estimate the
transmissivity field in a two-dimensional steady-state subsurface flow model of
the Hanford Site by assimilating synthetic measurements of transmissivity and
hydraulic head. We find that the execution time of CKLEMAP scales nearly
linearly as $N^{1.33}$, where $N$ is the number of discretization nodes, while
the execution time of standard MAP scales as $N^{2.91}$. The CKLEMAP method
improved execution time without sacrificing accuracy when compared to the
standard MAP.
Related papers
- Maximum a Posteriori Estimation for Linear Structural Dynamics Models Using Bayesian Optimization with Rational Polynomial Chaos Expansions [0.01578888899297715]
We propose an extension to an existing sparse Bayesian learning approach for MAP estimation.
We introduce a Bayesian optimization approach, which allows to adaptively enrich the experimental design.
By combining the sparsity-inducing learning procedure with the experimental design, we effectively reduce the number of model evaluations.
arXiv Detail & Related papers (2024-08-07T06:11:37Z) - Randomized Physics-Informed Machine Learning for Uncertainty
Quantification in High-Dimensional Inverse Problems [49.1574468325115]
We propose a physics-informed machine learning method for uncertainty quantification in high-dimensional inverse problems.
We show analytically and through comparison with Hamiltonian Monte Carlo that the rPICKLE posterior converges to the true posterior given by the Bayes rule.
arXiv Detail & Related papers (2023-12-11T07:33:16Z) - Adaptive LASSO estimation for functional hidden dynamic geostatistical
model [69.10717733870575]
We propose a novel model selection algorithm based on a penalized maximum likelihood estimator (PMLE) for functional hiddenstatistical models (f-HD)
The algorithm is based on iterative optimisation and uses an adaptive least absolute shrinkage and selector operator (GMSOLAS) penalty function, wherein the weights are obtained by the unpenalised f-HD maximum-likelihood estimators.
arXiv Detail & Related papers (2022-08-10T19:17:45Z) - Beyond EM Algorithm on Over-specified Two-Component Location-Scale
Gaussian Mixtures [29.26015093627193]
We develop the Exponential Location Update (ELU) algorithm to efficiently explore the curvature of the negative log-likelihood functions.
We demonstrate that the ELU algorithm converges to the final statistical radius of the models after a logarithmic number of iterations.
arXiv Detail & Related papers (2022-05-23T06:49:55Z) - Near-optimal estimation of smooth transport maps with kernel
sums-of-squares [81.02564078640275]
Under smoothness conditions, the squared Wasserstein distance between two distributions could be efficiently computed with appealing statistical error upper bounds.
The object of interest for applications such as generative modeling is the underlying optimal transport map.
We propose the first tractable algorithm for which the statistical $L2$ error on the maps nearly matches the existing minimax lower-bounds for smooth map estimation.
arXiv Detail & Related papers (2021-12-03T13:45:36Z) - Physics-Informed Machine Learning Method for Large-Scale Data
Assimilation Problems [48.7576911714538]
We extend the physics-informed conditional Karhunen-Lo'eve expansion (PICKLE) method for modeling subsurface flow with unknown flux (Neumann) and varying head (Dirichlet) boundary conditions.
We demonstrate that the PICKLE method is comparable in accuracy with the standard maximum a posteriori (MAP) method, but is significantly faster than MAP for large-scale problems.
arXiv Detail & Related papers (2021-07-30T18:43:14Z) - Recent advances in Bayesian optimization with applications to parameter
reconstruction in optical nano-metrology [0.0]
reconstruction is a common problem in optical nano metrology.
We present a Bayesian Target Vector Optimization scheme which combines two approaches.
We find that the presented method generally uses fewer calls of the model function than any of the competing schemes to achieve similar reconstruction performance.
arXiv Detail & Related papers (2021-07-12T15:32:15Z) - A Near-Optimal Gradient Flow for Learning Neural Energy-Based Models [93.24030378630175]
We propose a novel numerical scheme to optimize the gradient flows for learning energy-based models (EBMs)
We derive a second-order Wasserstein gradient flow of the global relative entropy from Fokker-Planck equation.
Compared with existing schemes, Wasserstein gradient flow is a smoother and near-optimal numerical scheme to approximate real data densities.
arXiv Detail & Related papers (2019-10-31T02:26:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.