Frequentist Parameter Estimation with Supervised Learning
- URL: http://arxiv.org/abs/2105.12302v1
- Date: Wed, 26 May 2021 02:24:25 GMT
- Title: Frequentist Parameter Estimation with Supervised Learning
- Authors: Samuel P. Nolan and Luca Pezz\`e and Augusto Smerzi
- Abstract summary: We use regression to infer a machine-learned point estimate of an unknown parameter.
When the number of training measurements are large, this is identical to the well-known maximum-likelihood estimator (MLE)
We show that the machine-learned estimator inherits the desirable properties of the MLE, up to a limit imposed by the resolution of the training grid.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently there has been a great deal of interest surrounding the calibration
of quantum sensors using machine learning techniques. In this work, we explore
the use of regression to infer a machine-learned point estimate of an unknown
parameter. Although the analysis is neccessarily frequentist - relying on
repeated esitmates to build up statistics - we clarify that this
machine-learned estimator converges to the Bayesian maximum a-posterori
estimator (subject to some regularity conditions). When the number of training
measurements are large, this is identical to the well-known maximum-likelihood
estimator (MLE), and using this fact, we argue that the Cram{\'e}r-Rao
sensitivity bound applies to the mean-square error cost function and can
therefore be used to select optimal model and training parameters. We show that
the machine-learned estimator inherits the desirable asymptotic properties of
the MLE, up to a limit imposed by the resolution of the training grid.
Furthermore, we investigate the role of quantum noise the training process, and
show that this noise imposes a fundamental limit on number of grid points. This
manuscript paves the way for machine-learning to assist the calibration of
quantum sensors, thereby allowing maximum-likelihood inference to play a more
prominent role in the design and operation of the next generation of
ultra-precise sensors.
Related papers
- Relaxed Quantile Regression: Prediction Intervals for Asymmetric Noise [51.87307904567702]
Quantile regression is a leading approach for obtaining such intervals via the empirical estimation of quantiles in the distribution of outputs.
We propose Relaxed Quantile Regression (RQR), a direct alternative to quantile regression based interval construction that removes this arbitrary constraint.
We demonstrate that this added flexibility results in intervals with an improvement in desirable qualities.
arXiv Detail & Related papers (2024-06-05T13:36:38Z) - Applications of model-aware reinforcement learning in Bayesian quantum
metrology [0.6138671548064355]
In Belliardo et al., arXiv:2312.16985 (2023), we introduced a procedure capable of optimizing a wide range of tasks in quantum metrology and estimation.
We take a model-based approach to the optimisation where the physics describing the system is explicitly taken into account in the training.
This exposition showcases the broad applicability of the method, which has been implemented in the qsensoropt library released on PyPI.
arXiv Detail & Related papers (2024-03-08T22:48:07Z) - In Defense of Softmax Parametrization for Calibrated and Consistent
Learning to Defer [27.025808709031864]
It has been theoretically shown that popular estimators for learning to defer parameterized with softmax provide unbounded estimates for the likelihood of deferring.
We show that the cause of the miscalibrated and unbounded estimator in prior literature is due to the symmetric nature of the surrogate losses used and not due to softmax.
We propose a novel statistically consistent asymmetric softmax-based surrogate loss that can produce valid estimates without the issue of unboundedness.
arXiv Detail & Related papers (2023-11-02T09:15:52Z) - Stochastic Marginal Likelihood Gradients using Neural Tangent Kernels [78.6096486885658]
We introduce lower bounds to the linearized Laplace approximation of the marginal likelihood.
These bounds are amenable togradient-based optimization and allow to trade off estimation accuracy against computational complexity.
arXiv Detail & Related papers (2023-06-06T19:02:57Z) - PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language
Models [52.09865918265002]
We propose a novel quantize before fine-tuning'' framework, PreQuant.
PreQuant is compatible with various quantization strategies, with outlier-aware fine-tuning incorporated to correct the induced quantization error.
We demonstrate the effectiveness of PreQuant on the GLUE benchmark using BERT, RoBERTa, and T5.
arXiv Detail & Related papers (2023-05-30T08:41:33Z) - Sharp Calibrated Gaussian Processes [58.94710279601622]
State-of-the-art approaches for designing calibrated models rely on inflating the Gaussian process posterior variance.
We present a calibration approach that generates predictive quantiles using a computation inspired by the vanilla Gaussian process posterior variance.
Our approach is shown to yield a calibrated model under reasonable assumptions.
arXiv Detail & Related papers (2023-02-23T12:17:36Z) - Scalable Gaussian Process Hyperparameter Optimization via Coverage
Regularization [0.0]
We present a novel algorithm which estimates the smoothness and length-scale parameters in the Matern kernel in order to improve robustness of the resulting prediction uncertainties.
We achieve improved UQ over leave-one-out likelihood while maintaining a high degree of scalability as demonstrated in numerical experiments.
arXiv Detail & Related papers (2022-09-22T19:23:37Z) - Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled
Markov Chains [34.77971292478243]
The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture.
We develop a training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient.
We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.
arXiv Detail & Related papers (2020-10-05T08:11:55Z) - A machine learning approach to Bayesian parameter estimation [0.0]
We formulate parameter estimation as a classification task and use artificial neural networks to efficiently perform Bayesian estimation.
We show that the network's posterior distribution is centered at the true (unknown) value of the parameter within an uncertainty given by the inverse Fisher information.
arXiv Detail & Related papers (2020-06-03T16:33:21Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z) - Localized Debiased Machine Learning: Efficient Inference on Quantile
Treatment Effects and Beyond [69.83813153444115]
We consider an efficient estimating equation for the (local) quantile treatment effect ((L)QTE) in causal inference.
Debiased machine learning (DML) is a data-splitting approach to estimating high-dimensional nuisances.
We propose localized debiased machine learning (LDML), which avoids this burdensome step.
arXiv Detail & Related papers (2019-12-30T14:42:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.