Bayesian Target-Vector Optimization for Efficient Parameter
Reconstruction
- URL: http://arxiv.org/abs/2202.11559v1
- Date: Wed, 23 Feb 2022 15:13:32 GMT
- Title: Bayesian Target-Vector Optimization for Efficient Parameter
Reconstruction
- Authors: Matthias Plock, Anna Andrle, Sven Burger, Philipp-Immanuel Schneider
- Abstract summary: We introduce a target-vector optimization scheme that considers all $K$ contributions of the model function and that is specifically suited for parameter reconstruction problems.
It also enables to determine accurate uncertainty estimates with very few observations of the actual model function.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Parameter reconstructions are indispensable in metrology. Here, on wants to
explain $K$ experimental measurements by fitting to them a parameterized model
of the measurement process. The model parameters are regularly determined by
least-square methods, i.e., by minimizing the sum of the squared residuals
between the $K$ model predictions and the $K$ experimental observations,
$\chi^2$. The model functions often involve computationally demanding numerical
simulations. Bayesian optimization methods are specifically suited for
minimizing expensive model functions. However, in contrast to least-square
methods such as the Levenberg-Marquardt algorithm, they only take the value of
$\chi^2$ into account, and neglect the $K$ individual model outputs. We
introduce a Bayesian target-vector optimization scheme that considers all $K$
contributions of the model function and that is specifically suited for
parameter reconstruction problems which are often based on hundreds of
observations. Its performance is compared to established methods for an optical
metrology reconstruction problem and two synthetic least-squares problems. The
proposed method outperforms established optimization methods. It also enables
to determine accurate uncertainty estimates with very few observations of the
actual model function by using Markov chain Monte Carlo sampling on a trained
surrogate model.
Related papers
- Accelerated zero-order SGD under high-order smoothness and overparameterized regime [79.85163929026146]
We present a novel gradient-free algorithm to solve convex optimization problems.
Such problems are encountered in medicine, physics, and machine learning.
We provide convergence guarantees for the proposed algorithm under both types of noise.
arXiv Detail & Related papers (2024-11-21T10:26:17Z) - Multivariate root-n-consistent smoothing parameter free matching estimators and estimators of inverse density weighted expectations [51.000851088730684]
We develop novel modifications of nearest-neighbor and matching estimators which converge at the parametric $sqrt n $-rate.
We stress that our estimators do not involve nonparametric function estimators and in particular do not rely on sample-size dependent parameters smoothing.
arXiv Detail & Related papers (2024-07-11T13:28:34Z) - Variational Bayesian surrogate modelling with application to robust design optimisation [0.9626666671366836]
Surrogate models provide a quick-to-evaluate approximation to complex computational models.
We consider Bayesian inference for constructing statistical surrogates with input uncertainties and dimensionality reduction.
We demonstrate intrinsic and robust structural optimisation problems where cost functions depend on a weighted sum of the mean and standard deviation of model outputs.
arXiv Detail & Related papers (2024-04-23T09:22:35Z) - Improving Sample Efficiency of Model-Free Algorithms for Zero-Sum Markov Games [66.2085181793014]
We show that a model-free stage-based Q-learning algorithm can enjoy the same optimality in the $H$ dependence as model-based algorithms.
Our algorithm features a key novel design of updating the reference value functions as the pair of optimistic and pessimistic value functions.
arXiv Detail & Related papers (2023-08-17T08:34:58Z) - On the Effectiveness of Parameter-Efficient Fine-Tuning [79.6302606855302]
Currently, many research works propose to only fine-tune a small portion of the parameters while keeping most of the parameters shared across different tasks.
We show that all of the methods are actually sparse fine-tuned models and conduct a novel theoretical analysis of them.
Despite the effectiveness of sparsity grounded by our theory, it still remains an open problem of how to choose the tunable parameters.
arXiv Detail & Related papers (2022-11-28T17:41:48Z) - KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal [70.15267479220691]
We consider and analyze the sample complexity of model reinforcement learning with a generative variance-free model.
Our analysis shows that it is nearly minimax-optimal for finding an $varepsilon$-optimal policy when $varepsilon$ is sufficiently small.
arXiv Detail & Related papers (2022-05-27T19:39:24Z) - Inverting brain grey matter models with likelihood-free inference: a
tool for trustable cytoarchitecture measurements [62.997667081978825]
characterisation of the brain grey matter cytoarchitecture with quantitative sensitivity to soma density and volume remains an unsolved challenge in dMRI.
We propose a new forward model, specifically a new system of equations, requiring a few relatively sparse b-shells.
We then apply modern tools from Bayesian analysis known as likelihood-free inference (LFI) to invert our proposed model.
arXiv Detail & Related papers (2021-11-15T09:08:27Z) - Variational Inference with NoFAS: Normalizing Flow with Adaptive
Surrogate for Computationally Expensive Models [7.217783736464403]
Use of sampling-based approaches such as Markov chain Monte Carlo may become intractable when each likelihood evaluation is computationally expensive.
New approaches combining variational inference with normalizing flow are characterized by a computational cost that grows only linearly with the dimensionality of the latent variable space.
We propose Normalizing Flow with Adaptive Surrogate (NoFAS), an optimization strategy that alternatively updates the normalizing flow parameters and the weights of a neural network surrogate model.
arXiv Detail & Related papers (2021-08-28T14:31:45Z) - Compressed particle methods for expensive models with application in
Astronomy and Remote Sensing [15.874578163779047]
We introduce a novel approach where the expensive model is evaluated only in some well-chosen samples.
We provide theoretical results supporting the novel algorithms and give empirical evidence of the performance of the proposed method in several numerical experiments.
Two of them are real-world applications in astronomy and satellite remote sensing.
arXiv Detail & Related papers (2021-07-18T14:45:23Z) - Recent advances in Bayesian optimization with applications to parameter
reconstruction in optical nano-metrology [0.0]
reconstruction is a common problem in optical nano metrology.
We present a Bayesian Target Vector Optimization scheme which combines two approaches.
We find that the presented method generally uses fewer calls of the model function than any of the competing schemes to achieve similar reconstruction performance.
arXiv Detail & Related papers (2021-07-12T15:32:15Z) - Learning the Stein Discrepancy for Training and Evaluating Energy-Based
Models without Sampling [30.406623987492726]
We present a new method for evaluating and training unnormalized density models.
We estimate the Stein discrepancy between the data density $p(x)$ and the model density $q(x)$ defined by a vector function of the data.
This yields a novel goodness-of-fit test which outperforms existing methods on high dimensional data.
arXiv Detail & Related papers (2020-02-13T16:39:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.