Recent advances in Bayesian optimization with applications to parameter
reconstruction in optical nano-metrology
- URL: http://arxiv.org/abs/2107.05499v1
- Date: Mon, 12 Jul 2021 15:32:15 GMT
- Title: Recent advances in Bayesian optimization with applications to parameter
reconstruction in optical nano-metrology
- Authors: Matthias Plock, Sven Burger, Philipp-Immanuel Schneider
- Abstract summary: reconstruction is a common problem in optical nano metrology.
We present a Bayesian Target Vector Optimization scheme which combines two approaches.
We find that the presented method generally uses fewer calls of the model function than any of the competing schemes to achieve similar reconstruction performance.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Parameter reconstruction is a common problem in optical nano metrology. It
generally involves a set of measurements, to which one attempts to fit a
numerical model of the measurement process. The model evaluation typically
involves to solve Maxwell's equations and is thus time consuming. This makes
the reconstruction computationally demanding. Several methods exist for fitting
the model to the measurements. On the one hand, Bayesian optimization methods
for expensive black-box optimization enable an efficient reconstruction by
training a machine learning model of the squared sum of deviations. On the
other hand, curve fitting algorithms, such as the Levenberg-Marquardt method,
take the deviations between all model outputs and corresponding measurement
values into account which enables a fast local convergence. In this paper we
present a Bayesian Target Vector Optimization scheme which combines these two
approaches. We compare the performance of the presented method against a
standard Levenberg-Marquardt-like algorithm, a conventional Bayesian
optimization scheme, and the L-BFGS-B and Nelder-Mead simplex algorithms. As a
stand-in for problems from nano metrology, we employ a non-linear least-square
problem from the NIST Standard Reference Database. We find that the presented
method generally uses fewer calls of the model function than any of the
competing schemes to achieve similar reconstruction performance.
Related papers
- Symmetric Rank-One Quasi-Newton Methods for Deep Learning Using Cubic Regularization [0.5120567378386615]
First-order descent and other first-order variants, such as Adam and AdaGrad, are commonly used in the field of deep learning.
However, these methods do not exploit curvature information.
Quasi-Newton methods re-use previously computed low Hessian approximations.
arXiv Detail & Related papers (2025-02-17T20:20:11Z) - Training Deep Learning Models with Norm-Constrained LMOs [56.00317694850397]
We study optimization methods that leverage the linear minimization oracle (LMO) over a norm-ball.
We propose a new family of algorithms that uses the LMO to adapt to the geometry of the problem and, perhaps surprisingly, show that they can be applied to unconstrained problems.
arXiv Detail & Related papers (2025-02-11T13:10:34Z) - Go With the Flow: Fast Diffusion for Gaussian Mixture Models [13.03355083378673]
Schr"odinger Bridges (SB) are diffusion processes that steer, in finite time, a given initial distribution to another final one while minimizing a suitable cost functional.
We propose latentmetrization of a set of SB policies for steering a system from one distribution to another.
We showcase the potential this approach in low-to-dimensional problems such as image-to-image translation in the space of an autoencoder.
arXiv Detail & Related papers (2024-12-12T08:40:22Z) - Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - An Optimization-based Deep Equilibrium Model for Hyperspectral Image
Deconvolution with Convergence Guarantees [71.57324258813675]
We propose a novel methodology for addressing the hyperspectral image deconvolution problem.
A new optimization problem is formulated, leveraging a learnable regularizer in the form of a neural network.
The derived iterative solver is then expressed as a fixed-point calculation problem within the Deep Equilibrium framework.
arXiv Detail & Related papers (2023-06-10T08:25:16Z) - Direct Estimation of Parameters in ODE Models Using WENDy: Weak-form
Estimation of Nonlinear Dynamics [0.0]
We introduce the Weak-form Estimation of Dynamics (WENDy) method for estimating model parameters for non-linear systems of ODEs.
WENDy computes accurate estimates and is robust to large (biologically relevant) levels of measurement noise.
We demonstrate the high robustness and computational efficiency by applying WENDy to estimate parameters in some common models from population biology, neuroscience, and biochemistry.
arXiv Detail & Related papers (2023-02-26T08:49:34Z) - Gaussian process regression and conditional Karhunen-Lo\'{e}ve models
for data assimilation in inverse problems [68.8204255655161]
We present a model inversion algorithm, CKLEMAP, for data assimilation and parameter estimation in partial differential equation models.
The CKLEMAP method provides better scalability compared to the standard MAP method.
arXiv Detail & Related papers (2023-01-26T18:14:12Z) - Sparse high-dimensional linear regression with a partitioned empirical
Bayes ECM algorithm [62.997667081978825]
We propose a computationally efficient and powerful Bayesian approach for sparse high-dimensional linear regression.
Minimal prior assumptions on the parameters are used through the use of plug-in empirical Bayes estimates.
The proposed approach is implemented in the R package probe.
arXiv Detail & Related papers (2022-09-16T19:15:50Z) - Bayesian Target-Vector Optimization for Efficient Parameter
Reconstruction [0.0]
We introduce a target-vector optimization scheme that considers all $K$ contributions of the model function and that is specifically suited for parameter reconstruction problems.
It also enables to determine accurate uncertainty estimates with very few observations of the actual model function.
arXiv Detail & Related papers (2022-02-23T15:13:32Z) - Effective Dimension Adaptive Sketching Methods for Faster Regularized
Least-Squares Optimization [56.05635751529922]
We propose a new randomized algorithm for solving L2-regularized least-squares problems based on sketching.
We consider two of the most popular random embeddings, namely, Gaussian embeddings and the Subsampled Randomized Hadamard Transform (SRHT)
arXiv Detail & Related papers (2020-06-10T15:00:09Z) - Bayesian System ID: Optimal management of parameter, model, and
measurement uncertainty [0.0]
We evaluate the robustness of a probabilistic formulation of system identification (ID) to sparse, noisy, and indirect data.
We show that the log posterior has improved geometric properties compared with the objective function surfaces of traditional methods.
arXiv Detail & Related papers (2020-03-04T22:48:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.